From ee2783fdf403ca08faeb6870f2fe752602f7bc8a Mon Sep 17 00:00:00 2001 From: Richard Theis Date: Tue, 27 Oct 2020 16:17:13 -0500 Subject: [PATCH] Update conformance results for v1.17/ibm-openshift (#1179) Red Hat OpenShift on IBM Cloud conformance results updated for version 4.4.27. --- v1.17/ibm-openshift/PRODUCT.yaml | 2 +- v1.17/ibm-openshift/e2e.log | 21657 ++++++++++++++--------------- v1.17/ibm-openshift/junit_01.xml | 9688 ++++++------- 3 files changed, 15398 insertions(+), 15949 deletions(-) diff --git a/v1.17/ibm-openshift/PRODUCT.yaml b/v1.17/ibm-openshift/PRODUCT.yaml index f2b231cd8a..5ca4f5d7a4 100644 --- a/v1.17/ibm-openshift/PRODUCT.yaml +++ b/v1.17/ibm-openshift/PRODUCT.yaml @@ -1,6 +1,6 @@ vendor: IBM name: Red Hat OpenShift on IBM Cloud -version: 4.4.20 +version: 4.4.27 website_url: https://www.ibm.com/cloud/openshift documentation_url: https://cloud.ibm.com/docs/openshift product_logo_url: https://github.com/ibm-cloud-docs/containers/blob/master/images-ui-only/logo-red-hat-openshift-on-ibm-cloud-light.svg diff --git a/v1.17/ibm-openshift/e2e.log b/v1.17/ibm-openshift/e2e.log index 33666f6191..a87d8ab849 100644 --- a/v1.17/ibm-openshift/e2e.log +++ b/v1.17/ibm-openshift/e2e.log @@ -1,887 +1,855 @@ -I0921 15:40:18.823289 24 test_context.go:406] Using a temporary kubeconfig file from in-cluster config : /tmp/kubeconfig-468631458 -I0921 15:40:18.823522 24 test_context.go:419] Tolerating taints "node-role.kubernetes.io/master" when considering if nodes are ready -I0921 15:40:18.823706 24 e2e.go:109] Starting e2e run "ada7e14e-2eab-4f79-8006-553689f5547c" on Ginkgo node 1 +I1026 16:00:38.225947 26 test_context.go:406] Using a temporary kubeconfig file from in-cluster config : /tmp/kubeconfig-871182394 +I1026 16:00:38.226120 26 test_context.go:419] Tolerating taints "node-role.kubernetes.io/master" when considering if nodes are ready +I1026 16:00:38.226383 26 e2e.go:109] Starting e2e run "2d1c55d1-9357-4e53-970b-33b80690768d" on Ginkgo node 1 {"msg":"Test Suite starting","total":280,"completed":0,"skipped":0,"failed":0} Running Suite: Kubernetes e2e suite =================================== -Random Seed: 1600702817 - Will randomize all specs +Random Seed: 1603728036 - Will randomize all specs Will run 280 of 4843 specs -Sep 21 15:40:18.835: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -Sep 21 15:40:18.839: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable -Sep 21 15:40:18.898: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready -Sep 21 15:40:18.995: INFO: 13 / 13 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) -Sep 21 15:40:18.995: INFO: expected 4 pod replicas in namespace 'kube-system', 4 are Running and Ready. -Sep 21 15:40:18.995: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start -Sep 21 15:40:19.028: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'ibm-keepalived-watcher' (0 seconds elapsed) -Sep 21 15:40:19.029: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'ibmcloud-block-storage-driver' (0 seconds elapsed) -Sep 21 15:40:19.029: INFO: e2e test version: v1.17.1 -Sep 21 15:40:19.036: INFO: kube-apiserver version: v1.17.1+6af3663 -Sep 21 15:40:19.036: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -Sep 21 15:40:19.101: INFO: Cluster IP family: ipv4 -SSSSSSSSSSSSSS +Oct 26 16:00:38.240: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +Oct 26 16:00:38.243: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable +Oct 26 16:00:38.302: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready +Oct 26 16:00:38.385: INFO: 13 / 13 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) +Oct 26 16:00:38.385: INFO: expected 4 pod replicas in namespace 'kube-system', 4 are Running and Ready. +Oct 26 16:00:38.385: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start +Oct 26 16:00:38.412: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'ibm-keepalived-watcher' (0 seconds elapsed) +Oct 26 16:00:38.412: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'ibmcloud-block-storage-driver' (0 seconds elapsed) +Oct 26 16:00:38.412: INFO: e2e test version: v1.17.1 +Oct 26 16:00:38.418: INFO: kube-apiserver version: v1.17.1+45f8ddb +Oct 26 16:00:38.418: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +Oct 26 16:00:38.436: INFO: Cluster IP family: ipv4 +SSSSSSS ------------------------------ -[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] - should include custom resource definition resources in discovery documents [Conformance] +[sig-storage] Projected downwardAPI + should provide podname only [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +[BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:40:19.103: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename custom-resource-definition -Sep 21 15:40:19.314: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled. +Oct 26 16:00:38.438: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename projected +Oct 26 16:00:38.624: INFO: No PodSecurityPolicies found; assuming PodSecurityPolicy is disabled. STEP: Waiting for a default service account to be provisioned in namespace -[It] should include custom resource definition resources in discovery documents [Conformance] +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40 +[It] should provide podname only [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: fetching the /apis discovery document -STEP: finding the apiextensions.k8s.io API group in the /apis discovery document -STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis discovery document -STEP: fetching the /apis/apiextensions.k8s.io discovery document -STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis/apiextensions.k8s.io discovery document -STEP: fetching the /apis/apiextensions.k8s.io/v1 discovery document -STEP: finding customresourcedefinitions resources in the /apis/apiextensions.k8s.io/v1 discovery document -[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +STEP: Creating a pod to test downward API volume plugin +Oct 26 16:00:38.678: INFO: Waiting up to 5m0s for pod "downwardapi-volume-ad397ed2-7a2a-4157-b6e5-bcd3404a4916" in namespace "projected-3847" to be "success or failure" +Oct 26 16:00:38.698: INFO: Pod "downwardapi-volume-ad397ed2-7a2a-4157-b6e5-bcd3404a4916": Phase="Pending", Reason="", readiness=false. Elapsed: 20.335813ms +Oct 26 16:00:40.714: INFO: Pod "downwardapi-volume-ad397ed2-7a2a-4157-b6e5-bcd3404a4916": Phase="Pending", Reason="", readiness=false. Elapsed: 2.036443917s +Oct 26 16:00:42.729: INFO: Pod "downwardapi-volume-ad397ed2-7a2a-4157-b6e5-bcd3404a4916": Phase="Pending", Reason="", readiness=false. Elapsed: 4.050832911s +Oct 26 16:00:44.746: INFO: Pod "downwardapi-volume-ad397ed2-7a2a-4157-b6e5-bcd3404a4916": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.067975732s +STEP: Saw pod success +Oct 26 16:00:44.746: INFO: Pod "downwardapi-volume-ad397ed2-7a2a-4157-b6e5-bcd3404a4916" satisfied condition "success or failure" +Oct 26 16:00:44.761: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-ad397ed2-7a2a-4157-b6e5-bcd3404a4916 container client-container: +STEP: delete the pod +Oct 26 16:00:44.877: INFO: Waiting for pod downwardapi-volume-ad397ed2-7a2a-4157-b6e5-bcd3404a4916 to disappear +Oct 26 16:00:44.891: INFO: Pod downwardapi-volume-ad397ed2-7a2a-4157-b6e5-bcd3404a4916 no longer exists +[AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:40:19.337: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "custom-resource-definition-7949" for this suite. -•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance]","total":280,"completed":1,"skipped":14,"failed":0} +Oct 26 16:00:44.891: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-3847" for this suite. +• [SLOW TEST:6.498 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:34 + should provide podname only [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -[sig-cli] Kubectl client Kubectl describe - should check if kubectl describe prints relevant information for rc and pods [Conformance] +{"msg":"PASSED [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance]","total":280,"completed":1,"skipped":7,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[k8s.io] Pods + should get a host IP [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-cli] Kubectl client +[BeforeEach] [k8s.io] Pods /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:40:19.376: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename kubectl +Oct 26 16:00:44.936: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278 -[It] should check if kubectl describe prints relevant information for rc and pods [Conformance] +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177 +[It] should get a host IP [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 15:40:19.582: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-4105' -Sep 21 15:40:20.319: INFO: stderr: "" -Sep 21 15:40:20.319: INFO: stdout: "replicationcontroller/agnhost-master created\n" -Sep 21 15:40:20.319: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-4105' -Sep 21 15:40:20.951: INFO: stderr: "" -Sep 21 15:40:20.951: INFO: stdout: "service/agnhost-master created\n" -STEP: Waiting for Agnhost master to start. -Sep 21 15:40:21.966: INFO: Selector matched 1 pods for map[app:agnhost] -Sep 21 15:40:21.966: INFO: Found 0 / 1 -Sep 21 15:40:22.983: INFO: Selector matched 1 pods for map[app:agnhost] -Sep 21 15:40:22.983: INFO: Found 0 / 1 -Sep 21 15:40:23.969: INFO: Selector matched 1 pods for map[app:agnhost] -Sep 21 15:40:23.969: INFO: Found 0 / 1 -Sep 21 15:40:24.966: INFO: Selector matched 1 pods for map[app:agnhost] -Sep 21 15:40:24.966: INFO: Found 0 / 1 -Sep 21 15:40:25.968: INFO: Selector matched 1 pods for map[app:agnhost] -Sep 21 15:40:25.968: INFO: Found 0 / 1 -Sep 21 15:40:26.976: INFO: Selector matched 1 pods for map[app:agnhost] -Sep 21 15:40:26.976: INFO: Found 1 / 1 -Sep 21 15:40:26.976: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 -Sep 21 15:40:27.026: INFO: Selector matched 1 pods for map[app:agnhost] -Sep 21 15:40:27.026: INFO: ForEach: Found 1 pods from the filter. Now looping through them. -Sep 21 15:40:27.026: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 describe pod agnhost-master-747jr --namespace=kubectl-4105' -Sep 21 15:40:27.347: INFO: stderr: "" -Sep 21 15:40:27.347: INFO: stdout: "Name: agnhost-master-747jr\nNamespace: kubectl-4105\nPriority: 0\nNode: 10.241.51.150/10.241.51.150\nStart Time: Mon, 21 Sep 2020 15:40:20 +0000\nLabels: app=agnhost\n role=master\nAnnotations: cni.projectcalico.org/podIP: 172.30.81.47/32\n cni.projectcalico.org/podIPs: 172.30.81.47/32\n k8s.v1.cni.cncf.io/networks-status:\n [{\n \"name\": \"k8s-pod-network\",\n \"ips\": [\n \"172.30.81.47\"\n ],\n \"dns\": {}\n }]\n openshift.io/scc: privileged\nStatus: Running\nIP: 172.30.81.47\nIPs:\n IP: 172.30.81.47\nControlled By: ReplicationController/agnhost-master\nContainers:\n agnhost-master:\n Container ID: cri-o://7437aeeee7f18e615aa35061bc1463df13b69f19ec506a272a530a9227271731\n Image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8\n Image ID: gcr.io/kubernetes-e2e-test-images/agnhost@sha256:02b9e10ea1f7439a03d9bc3ee16d984172ca7eac1818792ad4d721c5c8f72ff4\n Port: 6379/TCP\n Host Port: 0/TCP\n State: Running\n Started: Mon, 21 Sep 2020 15:40:25 +0000\n Ready: True\n Restart Count: 0\n Environment: \n Mounts:\n /var/run/secrets/kubernetes.io/serviceaccount from default-token-6psqq (ro)\nConditions:\n Type Status\n Initialized True \n Ready True \n ContainersReady True \n PodScheduled True \nVolumes:\n default-token-6psqq:\n Type: Secret (a volume populated by a Secret)\n SecretName: default-token-6psqq\n Optional: false\nQoS Class: BestEffort\nNode-Selectors: \nTolerations: node.kubernetes.io/not-ready:NoExecute for 300s\n node.kubernetes.io/unreachable:NoExecute for 300s\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Scheduled 7s default-scheduler Successfully assigned kubectl-4105/agnhost-master-747jr to 10.241.51.150\n Normal Pulling 6s kubelet, 10.241.51.150 Pulling image \"gcr.io/kubernetes-e2e-test-images/agnhost:2.8\"\n Normal Pulled 3s kubelet, 10.241.51.150 Successfully pulled image \"gcr.io/kubernetes-e2e-test-images/agnhost:2.8\"\n Normal Created 2s kubelet, 10.241.51.150 Created container agnhost-master\n Normal Started 2s kubelet, 10.241.51.150 Started container agnhost-master\n" -Sep 21 15:40:27.347: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 describe rc agnhost-master --namespace=kubectl-4105' -Sep 21 15:40:27.532: INFO: stderr: "" -Sep 21 15:40:27.532: INFO: stdout: "Name: agnhost-master\nNamespace: kubectl-4105\nSelector: app=agnhost,role=master\nLabels: app=agnhost\n role=master\nAnnotations: \nReplicas: 1 current / 1 desired\nPods Status: 1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n Labels: app=agnhost\n role=master\n Containers:\n agnhost-master:\n Image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8\n Port: 6379/TCP\n Host Port: 0/TCP\n Environment: \n Mounts: \n Volumes: \nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal SuccessfulCreate 7s replication-controller Created pod: agnhost-master-747jr\n" -Sep 21 15:40:27.532: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 describe service agnhost-master --namespace=kubectl-4105' -Sep 21 15:40:27.964: INFO: stderr: "" -Sep 21 15:40:27.964: INFO: stdout: "Name: agnhost-master\nNamespace: kubectl-4105\nLabels: app=agnhost\n role=master\nAnnotations: \nSelector: app=agnhost,role=master\nType: ClusterIP\nIP: 172.21.244.53\nPort: 6379/TCP\nTargetPort: agnhost-server/TCP\nEndpoints: 172.30.81.47:6379\nSession Affinity: None\nEvents: \n" -Sep 21 15:40:27.984: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 describe node 10.241.51.147' -Sep 21 15:40:28.309: INFO: stderr: "" -Sep 21 15:40:28.310: INFO: stdout: "Name: 10.241.51.147\nRoles: master,worker\nLabels: arch=amd64\n beta.kubernetes.io/arch=amd64\n beta.kubernetes.io/instance-type=b3c.4x16.encrypted\n beta.kubernetes.io/os=linux\n failure-domain.beta.kubernetes.io/region=us-south\n failure-domain.beta.kubernetes.io/zone=dal12\n ibm-cloud.kubernetes.io/encrypted-docker-data=true\n ibm-cloud.kubernetes.io/external-ip=169.48.196.44\n ibm-cloud.kubernetes.io/iaas-provider=softlayer\n ibm-cloud.kubernetes.io/internal-ip=10.241.51.147\n ibm-cloud.kubernetes.io/machine-type=b3c.4x16.encrypted\n ibm-cloud.kubernetes.io/os=REDHAT_7_64\n ibm-cloud.kubernetes.io/region=us-south\n ibm-cloud.kubernetes.io/sgx-enabled=false\n ibm-cloud.kubernetes.io/worker-id=kube-btkasf5d0sb6t34t3ij0-kubee2epvgc-default-000002eb\n ibm-cloud.kubernetes.io/worker-pool-id=btkasf5d0sb6t34t3ij0-951d290\n ibm-cloud.kubernetes.io/worker-pool-name=default\n ibm-cloud.kubernetes.io/worker-version=4.4.20_1517_openshift\n ibm-cloud.kubernetes.io/zone=dal12\n kubernetes.io/arch=amd64\n kubernetes.io/hostname=10.241.51.147\n kubernetes.io/os=linux\n node-role.kubernetes.io/master=\n node-role.kubernetes.io/worker=\n node.kubernetes.io/instance-type=b3c.4x16.encrypted\n node.openshift.io/os_id=rhel\n privateVLAN=2722938\n publicVLAN=2722936\n topology.kubernetes.io/region=us-south\n topology.kubernetes.io/zone=dal12\nAnnotations: projectcalico.org/IPv4Address: 10.241.51.147/26\n projectcalico.org/IPv4IPIPTunnelAddr: 172.30.237.64\nCreationTimestamp: Mon, 21 Sep 2020 14:11:48 +0000\nTaints: \nUnschedulable: false\nLease:\n HolderIdentity: 10.241.51.147\n AcquireTime: \n RenewTime: Mon, 21 Sep 2020 15:40:18 +0000\nConditions:\n Type Status LastHeartbeatTime LastTransitionTime Reason Message\n ---- ------ ----------------- ------------------ ------ -------\n NetworkUnavailable False Mon, 21 Sep 2020 14:12:33 +0000 Mon, 21 Sep 2020 14:12:33 +0000 CalicoIsUp Calico is running on this node\n MemoryPressure False Mon, 21 Sep 2020 15:40:26 +0000 Mon, 21 Sep 2020 14:11:48 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available\n DiskPressure False Mon, 21 Sep 2020 15:40:26 +0000 Mon, 21 Sep 2020 14:11:48 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure\n PIDPressure False Mon, 21 Sep 2020 15:40:26 +0000 Mon, 21 Sep 2020 14:11:48 +0000 KubeletHasSufficientPID kubelet has sufficient PID available\n Ready True Mon, 21 Sep 2020 15:40:26 +0000 Mon, 21 Sep 2020 14:12:48 +0000 KubeletReady kubelet is posting ready status\nAddresses:\n InternalIP: 10.241.51.147\n ExternalIP: 169.48.196.44\n Hostname: 10.241.51.147\nCapacity:\n cpu: 4\n ephemeral-storage: 103078840Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 16260880Ki\n pods: 110\nAllocatable:\n cpu: 3910m\n ephemeral-storage: 100275095474\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 13484816Ki\n pods: 110\nSystem Info:\n Machine ID: dfc7c0e6154b462290a45a74c4e47563\n System UUID: 922BF5F4-B667-3BC2-BE04-414781A3712F\n Boot ID: ebac6163-8b1a-4f78-9105-3c9591f56cfe\n Kernel Version: 3.10.0-1127.19.1.el7.x86_64\n OS Image: Red Hat\n Operating System: linux\n Architecture: amd64\n Container Runtime Version: cri-o://1.17.5-4.rhaos4.4.git7f0085b.el7\n Kubelet Version: v1.17.1+6af3663\n Kube-Proxy Version: v1.17.1+6af3663\nProviderID: ibm://fee034388aa6435883a1f720010ab3a2///btkasf5d0sb6t34t3ij0/kube-btkasf5d0sb6t34t3ij0-kubee2epvgc-default-000002eb\nNon-terminated Pods: (25 in total)\n Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits AGE\n --------- ---- ------------ ---------- --------------- ------------- ---\n calico-system calico-node-t2n87 0 (0%) 0 (0%) 0 (0%) 0 (0%) 88m\n calico-system calico-typha-6fc9ff8b5f-9pmb4 0 (0%) 0 (0%) 0 (0%) 0 (0%) 87m\n default test-k8s-e2e-pvg-master-verification 0 (0%) 0 (0%) 0 (0%) 0 (0%) 83m\n ibm-system ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-t6sqd 5m (0%) 0 (0%) 10Mi (0%) 0 (0%) 87m\n kube-system ibm-keepalived-watcher-h4dbn 5m (0%) 0 (0%) 10Mi (0%) 0 (0%) 88m\n kube-system ibm-master-proxy-static-10.241.51.147 25m (0%) 300m (7%) 32M (0%) 512M (3%) 88m\n kube-system ibmcloud-block-storage-driver-8r6lc 50m (1%) 300m (7%) 100Mi (0%) 300Mi (2%) 88m\n kube-system vpn-64d8d4987c-c89mq 5m (0%) 0 (0%) 5Mi (0%) 0 (0%) 80m\n openshift-cluster-node-tuning-operator tuned-6qvw2 10m (0%) 0 (0%) 50Mi (0%) 0 (0%) 88m\n openshift-console console-8549dd6c4f-dkvld 10m (0%) 0 (0%) 100Mi (0%) 0 (0%) 82m\n openshift-dns dns-default-f6vtl 110m (2%) 0 (0%) 70Mi (0%) 512Mi (3%) 88m\n openshift-image-registry node-ca-6r9nd 10m (0%) 0 (0%) 10Mi (0%) 0 (0%) 88m\n openshift-ingress router-default-84dfcdc9b5-pt6zq 100m (2%) 0 (0%) 256Mi (1%) 0 (0%) 91m\n openshift-kube-proxy openshift-kube-proxy-tvbvw 100m (2%) 0 (0%) 200Mi (1%) 0 (0%) 88m\n openshift-monitoring alertmanager-main-2 6m (0%) 0 (0%) 220Mi (1%) 0 (0%) 86m\n openshift-monitoring node-exporter-46mwx 9m (0%) 0 (0%) 210Mi (1%) 0 (0%) 88m\n openshift-monitoring prometheus-k8s-1 76m (1%) 0 (0%) 1184Mi (8%) 0 (0%) 85m\n openshift-monitoring prometheus-operator-8d96bcc68-b2bz9 5m (0%) 0 (0%) 60Mi (0%) 0 (0%) 82m\n openshift-monitoring thanos-querier-8f4c5c746-nh7wt 8m (0%) 0 (0%) 72Mi (0%) 0 (0%) 85m\n openshift-multus multus-7xt8r 10m (0%) 0 (0%) 150Mi (1%) 0 (0%) 88m\n openshift-multus multus-admission-controller-l9hmr 20m (0%) 0 (0%) 20Mi (0%) 0 (0%) 87m\n openshift-operator-lifecycle-manager packageserver-5ffd66994-7kjtm 10m (0%) 0 (0%) 50Mi (0%) 0 (0%) 81m\n sonobuoy sonobuoy 0 (0%) 0 (0%) 0 (0%) 0 (0%) 44s\n sonobuoy sonobuoy-e2e-job-36e1517c33ed4f81 0 (0%) 0 (0%) 0 (0%) 0 (0%) 36s\n sonobuoy sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-shgq6 0 (0%) 0 (0%) 0 (0%) 0 (0%) 36s\nAllocated resources:\n (Total limits may be over 100 percent, i.e., overcommitted.)\n Resource Requests Limits\n -------- -------- ------\n cpu 574m (14%) 600m (15%)\n memory 2874898Ki (21%) 1363443712 (9%)\n ephemeral-storage 0 (0%) 0 (0%)\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Starting 88m kubelet, 10.241.51.147 Starting kubelet.\n Normal NodeAllocatableEnforced 88m kubelet, 10.241.51.147 Updated Node Allocatable limit across pods\n Normal NodeHasSufficientMemory 88m (x7 over 88m) kubelet, 10.241.51.147 Node 10.241.51.147 status is now: NodeHasSufficientMemory\n Normal NodeHasNoDiskPressure 88m (x7 over 88m) kubelet, 10.241.51.147 Node 10.241.51.147 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 88m (x7 over 88m) kubelet, 10.241.51.147 Node 10.241.51.147 status is now: NodeHasSufficientPID\n Normal Starting 88m kube-proxy, 10.241.51.147 Starting kube-proxy.\n Normal NodeReady 87m kubelet, 10.241.51.147 Node 10.241.51.147 status is now: NodeReady\n" -Sep 21 15:40:28.310: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 describe namespace kubectl-4105' -Sep 21 15:40:28.479: INFO: stderr: "" -Sep 21 15:40:28.479: INFO: stdout: "Name: kubectl-4105\nLabels: e2e-framework=kubectl\n e2e-run=ada7e14e-2eab-4f79-8006-553689f5547c\nAnnotations: openshift.io/sa.scc.mcs: s0:c42,c24\n openshift.io/sa.scc.supplemental-groups: 1001770000/10000\n openshift.io/sa.scc.uid-range: 1001770000/10000\nStatus: Active\n\nNo resource quota.\n\nNo LimitRange resource.\n" -[AfterEach] [sig-cli] Kubectl client +STEP: creating pod +Oct 26 16:00:50.194: INFO: Pod pod-hostip-8ddaec84-eadb-4c35-a682-51a48428f2bb has hostIP: 10.72.119.74 +[AfterEach] [k8s.io] Pods /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:40:28.479: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-4105" for this suite. +Oct 26 16:00:50.194: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-4223" for this suite. -• [SLOW TEST:9.153 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 - Kubectl describe - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1154 - should check if kubectl describe prints relevant information for rc and pods [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +• [SLOW TEST:5.297 seconds] +[k8s.io] Pods +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 + should get a host IP [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance]","total":280,"completed":2,"skipped":14,"failed":0} -SSSSSSS +{"msg":"PASSED [k8s.io] Pods should get a host IP [NodeConformance] [Conformance]","total":280,"completed":2,"skipped":32,"failed":0} +SSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Secrets - should fail to create secret due to empty secret key [Conformance] +[k8s.io] Security Context When creating a pod with privileged + should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] Secrets +[BeforeEach] [k8s.io] Security Context /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:40:28.529: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename secrets +Oct 26 16:00:50.235: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace -[It] should fail to create secret due to empty secret key [Conformance] +[BeforeEach] [k8s.io] Security Context + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39 +[It] should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating projection with secret that has name secret-emptykey-test-7335b724-cd17-42a4-84c0-4272061884a6 -[AfterEach] [sig-api-machinery] Secrets +Oct 26 16:00:50.455: INFO: Waiting up to 5m0s for pod "busybox-privileged-false-66c92399-907f-48fb-93cc-511166c07f53" in namespace "security-context-test-2592" to be "success or failure" +Oct 26 16:00:50.473: INFO: Pod "busybox-privileged-false-66c92399-907f-48fb-93cc-511166c07f53": Phase="Pending", Reason="", readiness=false. Elapsed: 18.104827ms +Oct 26 16:00:52.502: INFO: Pod "busybox-privileged-false-66c92399-907f-48fb-93cc-511166c07f53": Phase="Pending", Reason="", readiness=false. Elapsed: 2.047525854s +Oct 26 16:00:54.518: INFO: Pod "busybox-privileged-false-66c92399-907f-48fb-93cc-511166c07f53": Phase="Pending", Reason="", readiness=false. Elapsed: 4.063469881s +Oct 26 16:00:56.532: INFO: Pod "busybox-privileged-false-66c92399-907f-48fb-93cc-511166c07f53": Phase="Pending", Reason="", readiness=false. Elapsed: 6.076984363s +Oct 26 16:00:58.546: INFO: Pod "busybox-privileged-false-66c92399-907f-48fb-93cc-511166c07f53": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.091330661s +Oct 26 16:00:58.546: INFO: Pod "busybox-privileged-false-66c92399-907f-48fb-93cc-511166c07f53" satisfied condition "success or failure" +Oct 26 16:00:58.584: INFO: Got logs for pod "busybox-privileged-false-66c92399-907f-48fb-93cc-511166c07f53": "ip: RTNETLINK answers: Operation not permitted\n" +[AfterEach] [k8s.io] Security Context /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:40:28.772: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-7166" for this suite. -•{"msg":"PASSED [sig-api-machinery] Secrets should fail to create secret due to empty secret key [Conformance]","total":280,"completed":3,"skipped":21,"failed":0} -SSSSSSSSSSSSSSSS +Oct 26 16:00:58.584: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "security-context-test-2592" for this suite. + +• [SLOW TEST:8.398 seconds] +[k8s.io] Security Context +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 + When creating a pod with privileged + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:225 + should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - should perform canary updates and phased rolling updates of template modifications [Conformance] +{"msg":"PASSED [k8s.io] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":3,"skipped":44,"failed":0} +S +------------------------------ +[sig-storage] Downward API volume + should provide container's memory request [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-apps] StatefulSet +[BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:40:28.822: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename statefulset +Oct 26 16:00:58.634: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] StatefulSet - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64 -[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79 -STEP: Creating service test in namespace statefulset-7671 -[It] should perform canary updates and phased rolling updates of template modifications [Conformance] +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40 +[It] should provide container's memory request [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a new StatefulSet -Sep 21 15:40:29.164: INFO: Found 0 stateful pods, waiting for 3 -Sep 21 15:40:39.181: INFO: Found 2 stateful pods, waiting for 3 -Sep 21 15:40:49.209: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true -Sep 21 15:40:49.209: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true -Sep 21 15:40:49.209: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Pending - Ready=false -Sep 21 15:40:59.254: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true -Sep 21 15:40:59.255: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true -Sep 21 15:40:59.255: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true -STEP: Updating stateful set template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine -Sep 21 15:40:59.354: INFO: Updating stateful set ss2 -STEP: Creating a new revision -STEP: Not applying an update when the partition is greater than the number of replicas -STEP: Performing a canary update -Sep 21 15:41:09.471: INFO: Updating stateful set ss2 -Sep 21 15:41:09.501: INFO: Waiting for Pod statefulset-7671/ss2-2 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 -STEP: Restoring Pods to the correct revision when they are deleted -Sep 21 15:41:19.660: INFO: Found 2 stateful pods, waiting for 3 -Sep 21 15:41:29.679: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true -Sep 21 15:41:29.679: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true -Sep 21 15:41:29.679: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true -STEP: Performing a phased rolling update -Sep 21 15:41:29.760: INFO: Updating stateful set ss2 -Sep 21 15:41:29.797: INFO: Waiting for Pod statefulset-7671/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 -Sep 21 15:41:39.871: INFO: Updating stateful set ss2 -Sep 21 15:41:39.899: INFO: Waiting for StatefulSet statefulset-7671/ss2 to complete update -Sep 21 15:41:39.899: INFO: Waiting for Pod statefulset-7671/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 -Sep 21 15:42:18.731: INFO: Waiting for StatefulSet statefulset-7671/ss2 to complete update -Sep 21 15:42:18.731: INFO: Waiting for Pod statefulset-7671/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 -Sep 21 15:42:19.948: INFO: Waiting for StatefulSet statefulset-7671/ss2 to complete update -Sep 21 15:42:19.948: INFO: Waiting for Pod statefulset-7671/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 -Sep 21 15:42:29.933: INFO: Waiting for StatefulSet statefulset-7671/ss2 to complete update -[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 -Sep 21 15:42:39.942: INFO: Deleting all statefulset in ns statefulset-7671 -Sep 21 15:42:39.969: INFO: Scaling statefulset ss2 to 0 -Sep 21 15:43:00.049: INFO: Waiting for statefulset status.replicas updated to 0 -Sep 21 15:43:00.062: INFO: Deleting statefulset ss2 -[AfterEach] [sig-apps] StatefulSet +STEP: Creating a pod to test downward API volume plugin +Oct 26 16:00:58.839: INFO: Waiting up to 5m0s for pod "downwardapi-volume-4a383249-a52b-4f5b-a643-5bdf68ca884d" in namespace "downward-api-8720" to be "success or failure" +Oct 26 16:00:58.865: INFO: Pod "downwardapi-volume-4a383249-a52b-4f5b-a643-5bdf68ca884d": Phase="Pending", Reason="", readiness=false. Elapsed: 26.114277ms +Oct 26 16:01:00.884: INFO: Pod "downwardapi-volume-4a383249-a52b-4f5b-a643-5bdf68ca884d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.045107656s +Oct 26 16:01:02.898: INFO: Pod "downwardapi-volume-4a383249-a52b-4f5b-a643-5bdf68ca884d": Phase="Pending", Reason="", readiness=false. Elapsed: 4.059267766s +Oct 26 16:01:04.913: INFO: Pod "downwardapi-volume-4a383249-a52b-4f5b-a643-5bdf68ca884d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.07389313s +STEP: Saw pod success +Oct 26 16:01:04.913: INFO: Pod "downwardapi-volume-4a383249-a52b-4f5b-a643-5bdf68ca884d" satisfied condition "success or failure" +Oct 26 16:01:04.933: INFO: Trying to get logs from node 10.72.119.82 pod downwardapi-volume-4a383249-a52b-4f5b-a643-5bdf68ca884d container client-container: +STEP: delete the pod +Oct 26 16:01:05.042: INFO: Waiting for pod downwardapi-volume-4a383249-a52b-4f5b-a643-5bdf68ca884d to disappear +Oct 26 16:01:05.061: INFO: Pod downwardapi-volume-4a383249-a52b-4f5b-a643-5bdf68ca884d no longer exists +[AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:43:00.125: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "statefulset-7671" for this suite. +Oct 26 16:01:05.061: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-8720" for this suite. -• [SLOW TEST:151.348 seconds] -[sig-apps] StatefulSet -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 - [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - should perform canary updates and phased rolling updates of template modifications [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +• [SLOW TEST:6.470 seconds] +[sig-storage] Downward API volume +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:35 + should provide container's memory request [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +------------------------------ +{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance]","total":280,"completed":4,"skipped":45,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance]","total":280,"completed":4,"skipped":37,"failed":0} [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - works for CRD without validation schema [Conformance] + works for CRD with validation schema [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:43:00.171: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:01:05.104: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename crd-publish-openapi STEP: Waiting for a default service account to be provisioned in namespace -[It] works for CRD without validation schema [Conformance] +[It] works for CRD with validation schema [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 15:43:00.419: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: client-side validation (kubectl create and apply) allows request with any unknown properties -Sep 21 15:43:09.432: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-3927 create -f -' -Sep 21 15:43:10.320: INFO: stderr: "" -Sep 21 15:43:10.320: INFO: stdout: "e2e-test-crd-publish-openapi-606-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" -Sep 21 15:43:10.320: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-3927 delete e2e-test-crd-publish-openapi-606-crds test-cr' -Sep 21 15:43:10.497: INFO: stderr: "" -Sep 21 15:43:10.497: INFO: stdout: "e2e-test-crd-publish-openapi-606-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" -Sep 21 15:43:10.497: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-3927 apply -f -' -Sep 21 15:43:11.079: INFO: stderr: "" -Sep 21 15:43:11.080: INFO: stdout: "e2e-test-crd-publish-openapi-606-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" -Sep 21 15:43:11.080: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-3927 delete e2e-test-crd-publish-openapi-606-crds test-cr' -Sep 21 15:43:11.232: INFO: stderr: "" -Sep 21 15:43:11.232: INFO: stdout: "e2e-test-crd-publish-openapi-606-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" -STEP: kubectl explain works to explain CR without validation schema -Sep 21 15:43:11.232: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 explain e2e-test-crd-publish-openapi-606-crds' -Sep 21 15:43:11.777: INFO: stderr: "" -Sep 21 15:43:11.777: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-606-crd\nVERSION: crd-publish-openapi-test-empty.example.com/v1\n\nDESCRIPTION:\n \n" +Oct 26 16:01:05.247: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: client-side validation (kubectl create and apply) allows request with known and required properties +Oct 26 16:01:13.748: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-111 create -f -' +Oct 26 16:01:14.527: INFO: stderr: "" +Oct 26 16:01:14.527: INFO: stdout: "e2e-test-crd-publish-openapi-252-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" +Oct 26 16:01:14.527: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-111 delete e2e-test-crd-publish-openapi-252-crds test-foo' +Oct 26 16:01:14.781: INFO: stderr: "" +Oct 26 16:01:14.782: INFO: stdout: "e2e-test-crd-publish-openapi-252-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" +Oct 26 16:01:14.782: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-111 apply -f -' +Oct 26 16:01:15.379: INFO: stderr: "" +Oct 26 16:01:15.380: INFO: stdout: "e2e-test-crd-publish-openapi-252-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" +Oct 26 16:01:15.380: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-111 delete e2e-test-crd-publish-openapi-252-crds test-foo' +Oct 26 16:01:15.577: INFO: stderr: "" +Oct 26 16:01:15.577: INFO: stdout: "e2e-test-crd-publish-openapi-252-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" +STEP: client-side validation (kubectl create and apply) rejects request with unknown properties when disallowed by the schema +Oct 26 16:01:15.577: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-111 create -f -' +Oct 26 16:01:16.074: INFO: rc: 1 +Oct 26 16:01:16.078: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-111 apply -f -' +Oct 26 16:01:16.388: INFO: rc: 1 +STEP: client-side validation (kubectl create and apply) rejects request without required properties +Oct 26 16:01:16.388: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-111 create -f -' +Oct 26 16:01:16.686: INFO: rc: 1 +Oct 26 16:01:16.686: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-111 apply -f -' +Oct 26 16:01:17.216: INFO: rc: 1 +STEP: kubectl explain works to explain CR properties +Oct 26 16:01:17.216: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 explain e2e-test-crd-publish-openapi-252-crds' +Oct 26 16:01:17.734: INFO: stderr: "" +Oct 26 16:01:17.734: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-252-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nDESCRIPTION:\n Foo CRD for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t\n Specification of Foo\n\n status\t\n Status of Foo\n\n" +STEP: kubectl explain works to explain CR properties recursively +Oct 26 16:01:17.735: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 explain e2e-test-crd-publish-openapi-252-crds.metadata' +Oct 26 16:01:18.227: INFO: stderr: "" +Oct 26 16:01:18.227: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-252-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: metadata \n\nDESCRIPTION:\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n ObjectMeta is metadata that all persisted resources must have, which\n includes all objects users must create.\n\nFIELDS:\n annotations\t\n Annotations is an unstructured key value map stored with a resource that\n may be set by external tools to store and retrieve arbitrary metadata. They\n are not queryable and should be preserved when modifying objects. More\n info: http://kubernetes.io/docs/user-guide/annotations\n\n clusterName\t\n The name of the cluster which the object belongs to. This is used to\n distinguish resources with same name and namespace in different clusters.\n This field is not set anywhere right now and apiserver is going to ignore\n it if set in create or update request.\n\n creationTimestamp\t\n CreationTimestamp is a timestamp representing the server time when this\n object was created. It is not guaranteed to be set in happens-before order\n across separate operations. Clients may not set this value. It is\n represented in RFC3339 form and is in UTC. Populated by the system.\n Read-only. Null for lists. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n deletionGracePeriodSeconds\t\n Number of seconds allowed for this object to gracefully terminate before it\n will be removed from the system. Only set when deletionTimestamp is also\n set. May only be shortened. Read-only.\n\n deletionTimestamp\t\n DeletionTimestamp is RFC 3339 date and time at which this resource will be\n deleted. This field is set by the server when a graceful deletion is\n requested by the user, and is not directly settable by a client. The\n resource is expected to be deleted (no longer visible from resource lists,\n and not reachable by name) after the time in this field, once the\n finalizers list is empty. As long as the finalizers list contains items,\n deletion is blocked. Once the deletionTimestamp is set, this value may not\n be unset or be set further into the future, although it may be shortened or\n the resource may be deleted prior to this time. For example, a user may\n request that a pod is deleted in 30 seconds. The Kubelet will react by\n sending a graceful termination signal to the containers in the pod. After\n that 30 seconds, the Kubelet will send a hard termination signal (SIGKILL)\n to the container and after cleanup, remove the pod from the API. In the\n presence of network partitions, this object may still exist after this\n timestamp, until an administrator or automated process can determine the\n resource is fully terminated. If not set, graceful deletion of the object\n has not been requested. Populated by the system when a graceful deletion is\n requested. Read-only. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n finalizers\t<[]string>\n Must be empty before the object is deleted from the registry. Each entry is\n an identifier for the responsible component that will remove the entry from\n the list. If the deletionTimestamp of the object is non-nil, entries in\n this list can only be removed. Finalizers may be processed and removed in\n any order. Order is NOT enforced because it introduces significant risk of\n stuck finalizers. finalizers is a shared field, any actor with permission\n can reorder it. If the finalizer list is processed in order, then this can\n lead to a situation in which the component responsible for the first\n finalizer in the list is waiting for a signal (field value, external\n system, or other) produced by a component responsible for a finalizer later\n in the list, resulting in a deadlock. Without enforced ordering finalizers\n are free to order amongst themselves and are not vulnerable to ordering\n changes in the list.\n\n generateName\t\n GenerateName is an optional prefix, used by the server, to generate a\n unique name ONLY IF the Name field has not been provided. If this field is\n used, the name returned to the client will be different than the name\n passed. This value will also be combined with a unique suffix. The provided\n value has the same validation rules as the Name field, and may be truncated\n by the length of the suffix required to make the value unique on the\n server. If this field is specified and the generated name exists, the\n server will NOT return a 409 - instead, it will either return 201 Created\n or 500 with Reason ServerTimeout indicating a unique name could not be\n found in the time allotted, and the client should retry (optionally after\n the time indicated in the Retry-After header). Applied only if Name is not\n specified. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency\n\n generation\t\n A sequence number representing a specific generation of the desired state.\n Populated by the system. Read-only.\n\n labels\t\n Map of string keys and values that can be used to organize and categorize\n (scope and select) objects. May match selectors of replication controllers\n and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n managedFields\t<[]Object>\n ManagedFields maps workflow-id and version to the set of fields that are\n managed by that workflow. This is mostly for internal housekeeping, and\n users typically shouldn't need to set or understand this field. A workflow\n can be the user's name, a controller's name, or the name of a specific\n apply path like \"ci-cd\". The set of fields is always in the version that\n the workflow used when modifying the object.\n\n name\t\n Name must be unique within a namespace. Is required when creating\n resources, although some resources may allow a client to request the\n generation of an appropriate name automatically. Name is primarily intended\n for creation idempotence and configuration definition. Cannot be updated.\n More info: http://kubernetes.io/docs/user-guide/identifiers#names\n\n namespace\t\n Namespace defines the space within each name must be unique. An empty\n namespace is equivalent to the \"default\" namespace, but \"default\" is the\n canonical representation. Not all objects are required to be scoped to a\n namespace - the value of this field for those objects will be empty. Must\n be a DNS_LABEL. Cannot be updated. More info:\n http://kubernetes.io/docs/user-guide/namespaces\n\n ownerReferences\t<[]Object>\n List of objects depended by this object. If ALL objects in the list have\n been deleted, this object will be garbage collected. If this object is\n managed by a controller, then an entry in this list will point to this\n controller, with the controller field set to true. There cannot be more\n than one managing controller.\n\n resourceVersion\t\n An opaque value that represents the internal version of this object that\n can be used by clients to determine when objects have changed. May be used\n for optimistic concurrency, change detection, and the watch operation on a\n resource or set of resources. Clients must treat these values as opaque and\n passed unmodified back to the server. They may only be valid for a\n particular resource or set of resources. Populated by the system.\n Read-only. Value must be treated as opaque by clients and . More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency\n\n selfLink\t\n SelfLink is a URL representing this object. Populated by the system.\n Read-only. DEPRECATED Kubernetes will stop propagating this field in 1.20\n release and the field is planned to be removed in 1.21 release.\n\n uid\t\n UID is the unique in time and space value for this object. It is typically\n generated by the server on successful creation of a resource and is not\n allowed to change on PUT operations. Populated by the system. Read-only.\n More info: http://kubernetes.io/docs/user-guide/identifiers#uids\n\n" +Oct 26 16:01:18.227: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 explain e2e-test-crd-publish-openapi-252-crds.spec' +Oct 26 16:01:18.791: INFO: stderr: "" +Oct 26 16:01:18.791: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-252-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: spec \n\nDESCRIPTION:\n Specification of Foo\n\nFIELDS:\n bars\t<[]Object>\n List of Bars and their specs.\n\n" +Oct 26 16:01:18.791: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 explain e2e-test-crd-publish-openapi-252-crds.spec.bars' +Oct 26 16:01:19.281: INFO: stderr: "" +Oct 26 16:01:19.281: INFO: stdout: "KIND: E2e-test-crd-publish-openapi-252-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: bars <[]Object>\n\nDESCRIPTION:\n List of Bars and their specs.\n\nFIELDS:\n age\t\n Age of Bar.\n\n bazs\t<[]string>\n List of Bazs.\n\n name\t -required-\n Name of Bar.\n\n" +STEP: kubectl explain works to return error when explain is called on property that doesn't exist +Oct 26 16:01:19.282: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 explain e2e-test-crd-publish-openapi-252-crds.spec.bars2' +Oct 26 16:01:19.858: INFO: rc: 1 [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:43:23.321: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-publish-openapi-3927" for this suite. +Oct 26 16:01:27.279: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-publish-openapi-111" for this suite. -• [SLOW TEST:23.197 seconds] +• [SLOW TEST:22.210 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - works for CRD without validation schema [Conformance] + works for CRD with validation schema [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance]","total":280,"completed":5,"skipped":37,"failed":0} -SSSSSSSS +{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance]","total":280,"completed":5,"skipped":95,"failed":0} +SSS ------------------------------ -[sig-network] DNS - should provide DNS for pods for Hostname [LinuxOnly] [Conformance] +[sig-storage] EmptyDir volumes + should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-network] DNS +[BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:43:23.369: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename dns +Oct 26 16:01:27.314: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace -[It] should provide DNS for pods for Hostname [LinuxOnly] [Conformance] +[It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a test headless service -STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-2657.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-2.dns-test-service-2.dns-2657.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/wheezy_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-2657.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done - -STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-2657.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-2.dns-test-service-2.dns-2657.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/jessie_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-2657.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done - -STEP: creating a pod to probe DNS -STEP: submitting the pod to kubernetes -STEP: retrieving the pod -STEP: looking for the results for each expected name from probers -Sep 21 15:43:38.288: INFO: DNS probes using dns-2657/dns-test-be3dd1c7-1933-4d49-aa94-22140b90ad5b succeeded - -STEP: deleting the pod -STEP: deleting the test headless service -[AfterEach] [sig-network] DNS +STEP: Creating a pod to test emptydir 0666 on node default medium +Oct 26 16:01:28.523: INFO: Waiting up to 5m0s for pod "pod-a6d16237-21dc-43eb-bf1f-34af7c141eb8" in namespace "emptydir-7265" to be "success or failure" +Oct 26 16:01:28.538: INFO: Pod "pod-a6d16237-21dc-43eb-bf1f-34af7c141eb8": Phase="Pending", Reason="", readiness=false. Elapsed: 14.586985ms +Oct 26 16:01:30.553: INFO: Pod "pod-a6d16237-21dc-43eb-bf1f-34af7c141eb8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03029141s +Oct 26 16:01:32.572: INFO: Pod "pod-a6d16237-21dc-43eb-bf1f-34af7c141eb8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.04858644s +STEP: Saw pod success +Oct 26 16:01:32.572: INFO: Pod "pod-a6d16237-21dc-43eb-bf1f-34af7c141eb8" satisfied condition "success or failure" +Oct 26 16:01:32.593: INFO: Trying to get logs from node 10.72.119.74 pod pod-a6d16237-21dc-43eb-bf1f-34af7c141eb8 container test-container: +STEP: delete the pod +Oct 26 16:01:32.700: INFO: Waiting for pod pod-a6d16237-21dc-43eb-bf1f-34af7c141eb8 to disappear +Oct 26 16:01:32.717: INFO: Pod pod-a6d16237-21dc-43eb-bf1f-34af7c141eb8 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:43:38.493: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "dns-2657" for this suite. +Oct 26 16:01:32.717: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-7265" for this suite. -• [SLOW TEST:15.180 seconds] -[sig-network] DNS -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 - should provide DNS for pods for Hostname [LinuxOnly] [Conformance] +• [SLOW TEST:5.434 seconds] +[sig-storage] EmptyDir volumes +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 + should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Hostname [LinuxOnly] [Conformance]","total":280,"completed":6,"skipped":45,"failed":0} +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":6,"skipped":98,"failed":0} SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] ResourceQuota - should create a ResourceQuota and capture the life of a pod. [Conformance] +[sig-apps] Job + should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] ResourceQuota +[BeforeEach] [sig-apps] Job /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:43:38.551: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename resourcequota +Oct 26 16:01:32.749: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename job STEP: Waiting for a default service account to be provisioned in namespace -[It] should create a ResourceQuota and capture the life of a pod. [Conformance] +[It] should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Counting existing ResourceQuota -STEP: Creating a ResourceQuota -STEP: Ensuring resource quota status is calculated -STEP: Creating a Pod that fits quota -STEP: Ensuring ResourceQuota status captures the pod usage -STEP: Not allowing a pod to be created that exceeds remaining quota -STEP: Not allowing a pod to be created that exceeds remaining quota(validation on extended resources) -STEP: Ensuring a pod cannot update its resource requirements -STEP: Ensuring attempts to update pod resource requirements did not change quota usage -STEP: Deleting the pod -STEP: Ensuring resource quota status released the pod usage -[AfterEach] [sig-api-machinery] ResourceQuota +STEP: Creating a job +STEP: Ensuring job reaches completions +[AfterEach] [sig-apps] Job /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:43:52.109: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-1309" for this suite. +Oct 26 16:01:40.977: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "job-1248" for this suite. -• [SLOW TEST:13.618 seconds] -[sig-api-machinery] ResourceQuota -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - should create a ResourceQuota and capture the life of a pod. [Conformance] +• [SLOW TEST:8.282 seconds] +[sig-apps] Job +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 + should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance]","total":280,"completed":7,"skipped":74,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]","total":280,"completed":7,"skipped":127,"failed":0} +SSS ------------------------------ -[sig-storage] ConfigMap - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] +[k8s.io] Docker Containers + should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] ConfigMap +[BeforeEach] [k8s.io] Docker Containers /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:43:52.169: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename configmap +Oct 26 16:01:41.032: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename containers STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] +[It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating configMap with name configmap-test-volume-e69fb91d-9d7a-400b-a90c-fdf6e46a968d -STEP: Creating a pod to test consume configMaps -Sep 21 15:43:52.479: INFO: Waiting up to 5m0s for pod "pod-configmaps-f28db1f3-79f2-48fa-9323-1b0ac7649fb6" in namespace "configmap-5592" to be "success or failure" -Sep 21 15:43:52.496: INFO: Pod "pod-configmaps-f28db1f3-79f2-48fa-9323-1b0ac7649fb6": Phase="Pending", Reason="", readiness=false. Elapsed: 16.156817ms -Sep 21 15:43:54.514: INFO: Pod "pod-configmaps-f28db1f3-79f2-48fa-9323-1b0ac7649fb6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035014208s -Sep 21 15:43:56.534: INFO: Pod "pod-configmaps-f28db1f3-79f2-48fa-9323-1b0ac7649fb6": Phase="Pending", Reason="", readiness=false. Elapsed: 4.054910953s -Sep 21 15:43:58.550: INFO: Pod "pod-configmaps-f28db1f3-79f2-48fa-9323-1b0ac7649fb6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.070505615s +STEP: Creating a pod to test override command +Oct 26 16:01:41.228: INFO: Waiting up to 5m0s for pod "client-containers-873b6213-1865-475a-91ed-8903142eb001" in namespace "containers-3653" to be "success or failure" +Oct 26 16:01:41.242: INFO: Pod "client-containers-873b6213-1865-475a-91ed-8903142eb001": Phase="Pending", Reason="", readiness=false. Elapsed: 13.897742ms +Oct 26 16:01:43.258: INFO: Pod "client-containers-873b6213-1865-475a-91ed-8903142eb001": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030292715s +Oct 26 16:01:45.289: INFO: Pod "client-containers-873b6213-1865-475a-91ed-8903142eb001": Phase="Pending", Reason="", readiness=false. Elapsed: 4.061027555s +Oct 26 16:01:47.305: INFO: Pod "client-containers-873b6213-1865-475a-91ed-8903142eb001": Phase="Pending", Reason="", readiness=false. Elapsed: 6.076646492s +Oct 26 16:01:49.321: INFO: Pod "client-containers-873b6213-1865-475a-91ed-8903142eb001": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.092971603s STEP: Saw pod success -Sep 21 15:43:58.550: INFO: Pod "pod-configmaps-f28db1f3-79f2-48fa-9323-1b0ac7649fb6" satisfied condition "success or failure" -Sep 21 15:43:58.567: INFO: Trying to get logs from node 10.241.51.150 pod pod-configmaps-f28db1f3-79f2-48fa-9323-1b0ac7649fb6 container configmap-volume-test: +Oct 26 16:01:49.321: INFO: Pod "client-containers-873b6213-1865-475a-91ed-8903142eb001" satisfied condition "success or failure" +Oct 26 16:01:49.336: INFO: Trying to get logs from node 10.72.119.74 pod client-containers-873b6213-1865-475a-91ed-8903142eb001 container test-container: STEP: delete the pod -Sep 21 15:43:58.699: INFO: Waiting for pod pod-configmaps-f28db1f3-79f2-48fa-9323-1b0ac7649fb6 to disappear -Sep 21 15:43:58.720: INFO: Pod pod-configmaps-f28db1f3-79f2-48fa-9323-1b0ac7649fb6 no longer exists -[AfterEach] [sig-storage] ConfigMap +Oct 26 16:01:49.420: INFO: Waiting for pod client-containers-873b6213-1865-475a-91ed-8903142eb001 to disappear +Oct 26 16:01:49.436: INFO: Pod client-containers-873b6213-1865-475a-91ed-8903142eb001 no longer exists +[AfterEach] [k8s.io] Docker Containers /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:43:58.720: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-5592" for this suite. +Oct 26 16:01:49.436: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "containers-3653" for this suite. -• [SLOW TEST:6.598 seconds] -[sig-storage] ConfigMap -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] +• [SLOW TEST:8.440 seconds] +[k8s.io] Docker Containers +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 + should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":8,"skipped":108,"failed":0} -[sig-storage] Projected downwardAPI - should update labels on modification [NodeConformance] [Conformance] +{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]","total":280,"completed":8,"skipped":130,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] Secrets + should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] Projected downwardAPI +[BeforeEach] [sig-storage] Secrets /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:43:58.768: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename projected +Oct 26 16:01:49.476: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40 -[It] should update labels on modification [NodeConformance] [Conformance] +[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating the pod -Sep 21 15:44:03.718: INFO: Successfully updated pod "labelsupdateaeff9803-afa2-4749-954b-fb66b32a06a9" -[AfterEach] [sig-storage] Projected downwardAPI +STEP: Creating secret with name secret-test-d58dde4c-031d-4339-affe-db272440ed03 +STEP: Creating a pod to test consume secrets +Oct 26 16:01:49.714: INFO: Waiting up to 5m0s for pod "pod-secrets-cd86f716-3d3f-47a7-bbaa-56d961238127" in namespace "secrets-5035" to be "success or failure" +Oct 26 16:01:49.730: INFO: Pod "pod-secrets-cd86f716-3d3f-47a7-bbaa-56d961238127": Phase="Pending", Reason="", readiness=false. Elapsed: 15.274417ms +Oct 26 16:01:51.746: INFO: Pod "pod-secrets-cd86f716-3d3f-47a7-bbaa-56d961238127": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031261796s +STEP: Saw pod success +Oct 26 16:01:51.746: INFO: Pod "pod-secrets-cd86f716-3d3f-47a7-bbaa-56d961238127" satisfied condition "success or failure" +Oct 26 16:01:51.760: INFO: Trying to get logs from node 10.72.119.74 pod pod-secrets-cd86f716-3d3f-47a7-bbaa-56d961238127 container secret-volume-test: +STEP: delete the pod +Oct 26 16:01:51.852: INFO: Waiting for pod pod-secrets-cd86f716-3d3f-47a7-bbaa-56d961238127 to disappear +Oct 26 16:01:51.868: INFO: Pod pod-secrets-cd86f716-3d3f-47a7-bbaa-56d961238127 no longer exists +[AfterEach] [sig-storage] Secrets /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:44:05.805: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-4940" for this suite. - -• [SLOW TEST:7.099 seconds] -[sig-storage] Projected downwardAPI -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:34 - should update labels on modification [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance]","total":280,"completed":9,"skipped":108,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSS +Oct 26 16:01:51.868: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-5035" for this suite. +•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":9,"skipped":156,"failed":0} +SS ------------------------------ -[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - should be able to convert from CR v1 to CR v2 [Conformance] +[sig-cli] Kubectl client Kubectl run job + should create a job from an image when restart is OnFailure [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] +[BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:44:05.868: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename crd-webhook +Oct 26 16:01:51.906: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:125 -STEP: Setting up server cert -STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication -STEP: Deploying the custom resource conversion webhook pod -STEP: Wait for the deployment to be ready -Sep 21 15:44:06.460: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set -Sep 21 15:44:08.507: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736299846, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736299846, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736299846, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736299846, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-78dcf5dd84\" is progressing."}}, CollisionCount:(*int32)(nil)} -Sep 21 15:44:10.523: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736299846, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736299846, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736299846, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736299846, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-78dcf5dd84\" is progressing."}}, CollisionCount:(*int32)(nil)} -STEP: Deploying the webhook service -STEP: Verifying the service has paired with the endpoint -Sep 21 15:44:13.571: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 -[It] should be able to convert from CR v1 to CR v2 [Conformance] +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278 +[BeforeEach] Kubectl run job + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1788 +[It] should create a job from an image when restart is OnFailure [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 15:44:13.590: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Creating a v1 custom resource -STEP: v2 custom resource should be converted -[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] +STEP: running the image docker.io/library/httpd:2.4.38-alpine +Oct 26 16:01:52.049: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 run e2e-test-httpd-job --restart=OnFailure --generator=job/v1 --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-1498' +Oct 26 16:01:52.201: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" +Oct 26 16:01:52.201: INFO: stdout: "job.batch/e2e-test-httpd-job created\n" +STEP: verifying the job e2e-test-httpd-job was created +[AfterEach] Kubectl run job + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1793 +Oct 26 16:01:52.218: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete jobs e2e-test-httpd-job --namespace=kubectl-1498' +Oct 26 16:01:52.387: INFO: stderr: "" +Oct 26 16:01:52.387: INFO: stdout: "job.batch \"e2e-test-httpd-job\" deleted\n" +[AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:44:16.013: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-webhook-9180" for this suite. -[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:136 +Oct 26 16:01:52.387: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-1498" for this suite. +•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run job should create a job from an image when restart is OnFailure [Conformance]","total":280,"completed":10,"skipped":158,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-api-machinery] ResourceQuota + should verify ResourceQuota with best effort scope. [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-api-machinery] ResourceQuota + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:01:52.425: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename resourcequota +STEP: Waiting for a default service account to be provisioned in namespace +[It] should verify ResourceQuota with best effort scope. [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating a ResourceQuota with best effort scope +STEP: Ensuring ResourceQuota status is calculated +STEP: Creating a ResourceQuota with not best effort scope +STEP: Ensuring ResourceQuota status is calculated +STEP: Creating a best-effort pod +STEP: Ensuring resource quota with best effort scope captures the pod usage +STEP: Ensuring resource quota with not best effort ignored the pod usage +STEP: Deleting the pod +STEP: Ensuring resource quota status released the pod usage +STEP: Creating a not best-effort pod +STEP: Ensuring resource quota with not best effort scope captures the pod usage +STEP: Ensuring resource quota with best effort scope ignored the pod usage +STEP: Deleting the pod +STEP: Ensuring resource quota status released the pod usage +[AfterEach] [sig-api-machinery] ResourceQuota + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:02:08.943: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-4529" for this suite. -• [SLOW TEST:11.245 seconds] -[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] +• [SLOW TEST:16.552 seconds] +[sig-api-machinery] ResourceQuota /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - should be able to convert from CR v1 to CR v2 [Conformance] + should verify ResourceQuota with best effort scope. [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance]","total":280,"completed":10,"skipped":132,"failed":0} -SSSSSSSSS +{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance]","total":280,"completed":11,"skipped":206,"failed":0} +SSSSSS ------------------------------ -[sig-storage] Projected configMap - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] +[sig-storage] Projected downwardAPI + should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] Projected configMap +[BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:44:17.113: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:02:08.979: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40 +[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating configMap with name projected-configmap-test-volume-map-eebd38c9-046b-4a6b-be30-71b3e2ef4647 -STEP: Creating a pod to test consume configMaps -Sep 21 15:44:17.391: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-c7737fd9-230e-4339-beca-8385cfac6720" in namespace "projected-3870" to be "success or failure" -Sep 21 15:44:17.408: INFO: Pod "pod-projected-configmaps-c7737fd9-230e-4339-beca-8385cfac6720": Phase="Pending", Reason="", readiness=false. Elapsed: 16.655515ms -Sep 21 15:44:19.422: INFO: Pod "pod-projected-configmaps-c7737fd9-230e-4339-beca-8385cfac6720": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030861137s -Sep 21 15:44:21.441: INFO: Pod "pod-projected-configmaps-c7737fd9-230e-4339-beca-8385cfac6720": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.049500805s +STEP: Creating a pod to test downward API volume plugin +Oct 26 16:02:09.174: INFO: Waiting up to 5m0s for pod "downwardapi-volume-5b50cc9b-5ae3-41ec-8c8d-94fa0208f685" in namespace "projected-1584" to be "success or failure" +Oct 26 16:02:09.197: INFO: Pod "downwardapi-volume-5b50cc9b-5ae3-41ec-8c8d-94fa0208f685": Phase="Pending", Reason="", readiness=false. Elapsed: 23.560971ms +Oct 26 16:02:11.213: INFO: Pod "downwardapi-volume-5b50cc9b-5ae3-41ec-8c8d-94fa0208f685": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.039227133s STEP: Saw pod success -Sep 21 15:44:21.441: INFO: Pod "pod-projected-configmaps-c7737fd9-230e-4339-beca-8385cfac6720" satisfied condition "success or failure" -Sep 21 15:44:21.453: INFO: Trying to get logs from node 10.241.51.147 pod pod-projected-configmaps-c7737fd9-230e-4339-beca-8385cfac6720 container projected-configmap-volume-test: +Oct 26 16:02:11.213: INFO: Pod "downwardapi-volume-5b50cc9b-5ae3-41ec-8c8d-94fa0208f685" satisfied condition "success or failure" +Oct 26 16:02:11.231: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-5b50cc9b-5ae3-41ec-8c8d-94fa0208f685 container client-container: STEP: delete the pod -Sep 21 15:44:21.544: INFO: Waiting for pod pod-projected-configmaps-c7737fd9-230e-4339-beca-8385cfac6720 to disappear -Sep 21 15:44:21.557: INFO: Pod pod-projected-configmaps-c7737fd9-230e-4339-beca-8385cfac6720 no longer exists -[AfterEach] [sig-storage] Projected configMap +Oct 26 16:02:11.310: INFO: Waiting for pod downwardapi-volume-5b50cc9b-5ae3-41ec-8c8d-94fa0208f685 to disappear +Oct 26 16:02:11.324: INFO: Pod downwardapi-volume-5b50cc9b-5ae3-41ec-8c8d-94fa0208f685 no longer exists +[AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:44:21.557: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-3870" for this suite. -•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":11,"skipped":141,"failed":0} -SSSSS +Oct 26 16:02:11.324: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-1584" for this suite. +•{"msg":"PASSED [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":12,"skipped":212,"failed":0} +SSSSSSS ------------------------------ -[sig-cli] Kubectl client Kubectl run pod - should create a pod from an image when restart is Never [Conformance] +[sig-network] DNS + should provide DNS for the cluster [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-cli] Kubectl client +[BeforeEach] [sig-network] DNS /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:44:21.602: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename kubectl +Oct 26 16:02:11.357: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename dns STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278 -[BeforeEach] Kubectl run pod - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1861 -[It] should create a pod from an image when restart is Never [Conformance] +[It] should provide DNS for the cluster [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: running the image docker.io/library/httpd:2.4.38-alpine -Sep 21 15:44:21.804: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 run e2e-test-httpd-pod --restart=Never --generator=run-pod/v1 --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-3658' -Sep 21 15:44:22.145: INFO: stderr: "" -Sep 21 15:44:22.145: INFO: stdout: "pod/e2e-test-httpd-pod created\n" -STEP: verifying the pod e2e-test-httpd-pod was created -[AfterEach] Kubectl run pod - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1866 -Sep 21 15:44:22.158: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete pods e2e-test-httpd-pod --namespace=kubectl-3658' -Sep 21 15:44:39.593: INFO: stderr: "" -Sep 21 15:44:39.593: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" -[AfterEach] [sig-cli] Kubectl client +STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1341.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done + +STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1341.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done + +STEP: creating a pod to probe DNS +STEP: submitting the pod to kubernetes +STEP: retrieving the pod +STEP: looking for the results for each expected name from probers +Oct 26 16:02:29.867: INFO: DNS probes using dns-1341/dns-test-4c088134-5952-41bc-a60e-855eb0c07cd7 succeeded + +STEP: deleting the pod +[AfterEach] [sig-network] DNS /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:44:39.593: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-3658" for this suite. +Oct 26 16:02:29.926: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "dns-1341" for this suite. -• [SLOW TEST:18.043 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 - Kubectl run pod - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1857 - should create a pod from an image when restart is Never [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +• [SLOW TEST:18.607 seconds] +[sig-network] DNS +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 + should provide DNS for the cluster [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never [Conformance]","total":280,"completed":12,"skipped":146,"failed":0} -SSSSSSSSSSSSS +{"msg":"PASSED [sig-network] DNS should provide DNS for the cluster [Conformance]","total":280,"completed":13,"skipped":219,"failed":0} +SSS ------------------------------ -[sig-cli] Kubectl client Kubectl version - should check is all data is printed [Conformance] +[sig-api-machinery] Secrets + should fail to create secret due to empty secret key [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-cli] Kubectl client +[BeforeEach] [sig-api-machinery] Secrets /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:44:39.645: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename kubectl +Oct 26 16:02:29.967: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278 -[It] should check is all data is printed [Conformance] +[It] should fail to create secret due to empty secret key [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 15:44:39.909: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 version' -Sep 21 15:44:40.041: INFO: stderr: "" -Sep 21 15:44:40.041: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"17\", GitVersion:\"v1.17.1\", GitCommit:\"d224476cd0730baca2b6e357d144171ed74192d6\", GitTreeState:\"clean\", BuildDate:\"2020-01-14T21:04:32Z\", GoVersion:\"go1.13.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"17+\", GitVersion:\"v1.17.1+6af3663\", GitCommit:\"6af3663\", GitTreeState:\"clean\", BuildDate:\"2020-08-25T04:54:06Z\", GoVersion:\"go1.13.4\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" -[AfterEach] [sig-cli] Kubectl client +STEP: Creating projection with secret that has name secret-emptykey-test-c7ac6a76-9d0f-43ae-a7fb-da3f1ee420e8 +[AfterEach] [sig-api-machinery] Secrets /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:44:40.041: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-7010" for this suite. -•{"msg":"PASSED [sig-cli] Kubectl client Kubectl version should check is all data is printed [Conformance]","total":280,"completed":13,"skipped":159,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +Oct 26 16:02:30.136: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-9902" for this suite. +•{"msg":"PASSED [sig-api-machinery] Secrets should fail to create secret due to empty secret key [Conformance]","total":280,"completed":14,"skipped":222,"failed":0} +SSSS ------------------------------ -[k8s.io] Pods - should get a host IP [NodeConformance] [Conformance] +[sig-network] Services + should provide secure master service [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Pods +[BeforeEach] [sig-network] Services /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:44:40.102: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename pods +Oct 26 16:02:30.171: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177 -[It] should get a host IP [NodeConformance] [Conformance] +[BeforeEach] [sig-network] Services + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139 +[It] should provide secure master service [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: creating pod -Sep 21 15:44:44.438: INFO: Pod pod-hostip-4a032136-678c-4815-b3e3-bb6ea709a2c8 has hostIP: 10.241.51.147 -[AfterEach] [k8s.io] Pods +[AfterEach] [sig-network] Services /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:44:44.438: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-5448" for this suite. -•{"msg":"PASSED [k8s.io] Pods should get a host IP [NodeConformance] [Conformance]","total":280,"completed":14,"skipped":192,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +Oct 26 16:02:30.340: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-5560" for this suite. +[AfterEach] [sig-network] Services + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143 +•{"msg":"PASSED [sig-network] Services should provide secure master service [Conformance]","total":280,"completed":15,"skipped":226,"failed":0} +SSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] +[k8s.io] Probing container + should have monotonically increasing restart count [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] EmptyDir volumes +[BeforeEach] [k8s.io] Probing container /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:44:44.485: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename emptydir +Oct 26 16:02:30.394: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 +[It] should have monotonically increasing restart count [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test emptydir 0644 on node default medium -Sep 21 15:44:44.736: INFO: Waiting up to 5m0s for pod "pod-8164a162-6125-4cb7-9712-caf87c18eaf6" in namespace "emptydir-6667" to be "success or failure" -Sep 21 15:44:44.759: INFO: Pod "pod-8164a162-6125-4cb7-9712-caf87c18eaf6": Phase="Pending", Reason="", readiness=false. Elapsed: 23.010937ms -Sep 21 15:44:46.780: INFO: Pod "pod-8164a162-6125-4cb7-9712-caf87c18eaf6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.044517813s -Sep 21 15:44:48.796: INFO: Pod "pod-8164a162-6125-4cb7-9712-caf87c18eaf6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.060798434s +STEP: Creating pod liveness-f00d6068-7e61-412b-b204-3fe28658ddb0 in namespace container-probe-6927 +Oct 26 16:02:35.666: INFO: Started pod liveness-f00d6068-7e61-412b-b204-3fe28658ddb0 in namespace container-probe-6927 +STEP: checking the pod's current state and verifying that restartCount is present +Oct 26 16:02:35.685: INFO: Initial restart count of pod liveness-f00d6068-7e61-412b-b204-3fe28658ddb0 is 0 +Oct 26 16:02:47.798: INFO: Restart count of pod container-probe-6927/liveness-f00d6068-7e61-412b-b204-3fe28658ddb0 is now 1 (12.112982154s elapsed) +Oct 26 16:03:07.983: INFO: Restart count of pod container-probe-6927/liveness-f00d6068-7e61-412b-b204-3fe28658ddb0 is now 2 (32.298102422s elapsed) +Oct 26 16:03:28.171: INFO: Restart count of pod container-probe-6927/liveness-f00d6068-7e61-412b-b204-3fe28658ddb0 is now 3 (52.485292275s elapsed) +Oct 26 16:03:48.363: INFO: Restart count of pod container-probe-6927/liveness-f00d6068-7e61-412b-b204-3fe28658ddb0 is now 4 (1m12.677315351s elapsed) +Oct 26 16:04:52.910: INFO: Restart count of pod container-probe-6927/liveness-f00d6068-7e61-412b-b204-3fe28658ddb0 is now 5 (2m17.224269189s elapsed) +STEP: deleting the pod +[AfterEach] [k8s.io] Probing container + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:04:52.967: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-6927" for this suite. + +• [SLOW TEST:142.606 seconds] +[k8s.io] Probing container +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 + should have monotonically increasing restart count [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +------------------------------ +{"msg":"PASSED [k8s.io] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance]","total":280,"completed":16,"skipped":232,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] Secrets + should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-storage] Secrets + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:04:53.001: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename secrets +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating secret with name secret-test-4806b11b-df47-49a5-8f2d-2ea306162446 +STEP: Creating a pod to test consume secrets +Oct 26 16:04:53.414: INFO: Waiting up to 5m0s for pod "pod-secrets-35319af5-de1a-4bb0-b8ff-c6122b8e3a2d" in namespace "secrets-3196" to be "success or failure" +Oct 26 16:04:53.429: INFO: Pod "pod-secrets-35319af5-de1a-4bb0-b8ff-c6122b8e3a2d": Phase="Pending", Reason="", readiness=false. Elapsed: 14.669086ms +Oct 26 16:04:55.445: INFO: Pod "pod-secrets-35319af5-de1a-4bb0-b8ff-c6122b8e3a2d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031012295s STEP: Saw pod success -Sep 21 15:44:48.797: INFO: Pod "pod-8164a162-6125-4cb7-9712-caf87c18eaf6" satisfied condition "success or failure" -Sep 21 15:44:48.811: INFO: Trying to get logs from node 10.241.51.150 pod pod-8164a162-6125-4cb7-9712-caf87c18eaf6 container test-container: +Oct 26 16:04:55.445: INFO: Pod "pod-secrets-35319af5-de1a-4bb0-b8ff-c6122b8e3a2d" satisfied condition "success or failure" +Oct 26 16:04:55.463: INFO: Trying to get logs from node 10.72.119.74 pod pod-secrets-35319af5-de1a-4bb0-b8ff-c6122b8e3a2d container secret-volume-test: STEP: delete the pod -Sep 21 15:44:48.896: INFO: Waiting for pod pod-8164a162-6125-4cb7-9712-caf87c18eaf6 to disappear -Sep 21 15:44:48.910: INFO: Pod pod-8164a162-6125-4cb7-9712-caf87c18eaf6 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +Oct 26 16:04:55.575: INFO: Waiting for pod pod-secrets-35319af5-de1a-4bb0-b8ff-c6122b8e3a2d to disappear +Oct 26 16:04:55.592: INFO: Pod pod-secrets-35319af5-de1a-4bb0-b8ff-c6122b8e3a2d no longer exists +[AfterEach] [sig-storage] Secrets /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:44:48.910: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-6667" for this suite. -•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":15,"skipped":239,"failed":0} -SSSSSSSSSSSSSSSSSSSSSS +Oct 26 16:04:55.592: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-3196" for this suite. +STEP: Destroying namespace "secret-namespace-6437" for this suite. +•{"msg":"PASSED [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]","total":280,"completed":17,"skipped":272,"failed":0} +SSSSSS ------------------------------ -[sig-cli] Kubectl client Kubectl run rc - should create an rc from an image [Conformance] +[sig-api-machinery] Garbage collector + should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-cli] Kubectl client +[BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:44:48.961: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename kubectl +Oct 26 16:04:55.646: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278 -[BeforeEach] Kubectl run rc - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1632 -[It] should create an rc from an image [Conformance] +[It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: running the image docker.io/library/httpd:2.4.38-alpine -Sep 21 15:44:49.142: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 run e2e-test-httpd-rc --image=docker.io/library/httpd:2.4.38-alpine --generator=run/v1 --namespace=kubectl-9712' -Sep 21 15:44:49.393: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" -Sep 21 15:44:49.393: INFO: stdout: "replicationcontroller/e2e-test-httpd-rc created\n" -STEP: verifying the rc e2e-test-httpd-rc was created -STEP: verifying the pod controlled by rc e2e-test-httpd-rc was created -STEP: confirm that you can get logs from an rc -Sep 21 15:44:51.439: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [e2e-test-httpd-rc-9pjxx] -Sep 21 15:44:51.439: INFO: Waiting up to 5m0s for pod "e2e-test-httpd-rc-9pjxx" in namespace "kubectl-9712" to be "running and ready" -Sep 21 15:44:51.454: INFO: Pod "e2e-test-httpd-rc-9pjxx": Phase="Pending", Reason="", readiness=false. Elapsed: 14.907137ms -Sep 21 15:44:53.470: INFO: Pod "e2e-test-httpd-rc-9pjxx": Phase="Running", Reason="", readiness=true. Elapsed: 2.03057323s -Sep 21 15:44:53.470: INFO: Pod "e2e-test-httpd-rc-9pjxx" satisfied condition "running and ready" -Sep 21 15:44:53.470: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [e2e-test-httpd-rc-9pjxx] -Sep 21 15:44:53.470: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 logs rc/e2e-test-httpd-rc --namespace=kubectl-9712' -Sep 21 15:44:53.663: INFO: stderr: "" -Sep 21 15:44:53.663: INFO: stdout: "AH00558: httpd: Could not reliably determine the server's fully qualified domain name, using 172.30.81.55. Set the 'ServerName' directive globally to suppress this message\nAH00558: httpd: Could not reliably determine the server's fully qualified domain name, using 172.30.81.55. Set the 'ServerName' directive globally to suppress this message\n[Mon Sep 21 15:44:51.021711 2020] [mpm_event:notice] [pid 1:tid 140480747846504] AH00489: Apache/2.4.38 (Unix) configured -- resuming normal operations\n[Mon Sep 21 15:44:51.021783 2020] [core:notice] [pid 1:tid 140480747846504] AH00094: Command line: 'httpd -D FOREGROUND'\n" -[AfterEach] Kubectl run rc - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1637 -Sep 21 15:44:53.663: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete rc e2e-test-httpd-rc --namespace=kubectl-9712' -Sep 21 15:44:53.856: INFO: stderr: "" -Sep 21 15:44:53.856: INFO: stdout: "replicationcontroller \"e2e-test-httpd-rc\" deleted\n" -[AfterEach] [sig-cli] Kubectl client +STEP: create the rc +STEP: delete the rc +STEP: wait for the rc to be deleted +STEP: Gathering metrics +W1026 16:05:01.912828 26 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +Oct 26 16:05:01.913: INFO: For apiserver_request_total: +For apiserver_request_latency_seconds: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: + +[AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:44:53.856: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-9712" for this suite. -•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run rc should create an rc from an image [Conformance]","total":280,"completed":16,"skipped":261,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSS +Oct 26 16:05:01.913: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-7314" for this suite. + +• [SLOW TEST:6.308 seconds] +[sig-api-machinery] Garbage collector +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 + should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +------------------------------ +{"msg":"PASSED [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]","total":280,"completed":18,"skipped":278,"failed":0} +SSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap - should be consumable from pods in volume [NodeConformance] [Conformance] + should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 [BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:44:53.910: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:05:01.956: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume [NodeConformance] [Conformance] +[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating configMap with name configmap-test-volume-5035fb7d-a03e-4636-a04f-d24da67cb379 +STEP: Creating configMap with name configmap-test-volume-map-ffc5d6ae-153f-4528-95e2-a4dfd648fc40 STEP: Creating a pod to test consume configMaps -Sep 21 15:44:54.247: INFO: Waiting up to 5m0s for pod "pod-configmaps-7e7661a0-46db-4c02-b121-a68d430dc331" in namespace "configmap-8218" to be "success or failure" -Sep 21 15:44:54.266: INFO: Pod "pod-configmaps-7e7661a0-46db-4c02-b121-a68d430dc331": Phase="Pending", Reason="", readiness=false. Elapsed: 18.338954ms -Sep 21 15:44:56.289: INFO: Pod "pod-configmaps-7e7661a0-46db-4c02-b121-a68d430dc331": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.041652141s +Oct 26 16:05:02.206: INFO: Waiting up to 5m0s for pod "pod-configmaps-3e84c370-f7d7-4d32-b98f-b6d9eaf8ebfb" in namespace "configmap-3499" to be "success or failure" +Oct 26 16:05:02.231: INFO: Pod "pod-configmaps-3e84c370-f7d7-4d32-b98f-b6d9eaf8ebfb": Phase="Pending", Reason="", readiness=false. Elapsed: 23.294862ms +Oct 26 16:05:04.248: INFO: Pod "pod-configmaps-3e84c370-f7d7-4d32-b98f-b6d9eaf8ebfb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.040105481s +Oct 26 16:05:06.264: INFO: Pod "pod-configmaps-3e84c370-f7d7-4d32-b98f-b6d9eaf8ebfb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.056409829s STEP: Saw pod success -Sep 21 15:44:56.289: INFO: Pod "pod-configmaps-7e7661a0-46db-4c02-b121-a68d430dc331" satisfied condition "success or failure" -Sep 21 15:44:56.303: INFO: Trying to get logs from node 10.241.51.150 pod pod-configmaps-7e7661a0-46db-4c02-b121-a68d430dc331 container configmap-volume-test: +Oct 26 16:05:06.264: INFO: Pod "pod-configmaps-3e84c370-f7d7-4d32-b98f-b6d9eaf8ebfb" satisfied condition "success or failure" +Oct 26 16:05:06.279: INFO: Trying to get logs from node 10.72.119.74 pod pod-configmaps-3e84c370-f7d7-4d32-b98f-b6d9eaf8ebfb container configmap-volume-test: STEP: delete the pod -Sep 21 15:44:56.387: INFO: Waiting for pod pod-configmaps-7e7661a0-46db-4c02-b121-a68d430dc331 to disappear -Sep 21 15:44:56.399: INFO: Pod pod-configmaps-7e7661a0-46db-4c02-b121-a68d430dc331 no longer exists +Oct 26 16:05:06.354: INFO: Waiting for pod pod-configmaps-3e84c370-f7d7-4d32-b98f-b6d9eaf8ebfb to disappear +Oct 26 16:05:06.368: INFO: Pod pod-configmaps-3e84c370-f7d7-4d32-b98f-b6d9eaf8ebfb no longer exists [AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:44:56.400: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-8218" for this suite. -•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":17,"skipped":288,"failed":0} -SSSSSSSSS +Oct 26 16:05:06.368: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-3499" for this suite. +•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":19,"skipped":293,"failed":0} + ------------------------------ -[sig-storage] EmptyDir volumes - should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] +[sig-auth] ServiceAccounts + should mount an API token into pods [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] EmptyDir volumes +[BeforeEach] [sig-auth] ServiceAccounts /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:44:56.447: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename emptydir +Oct 26 16:05:06.402: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] +[It] should mount an API token into pods [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test emptydir 0777 on node default medium -Sep 21 15:44:56.688: INFO: Waiting up to 5m0s for pod "pod-c6139c3a-6076-431a-856a-50a906af8e16" in namespace "emptydir-2416" to be "success or failure" -Sep 21 15:44:56.707: INFO: Pod "pod-c6139c3a-6076-431a-856a-50a906af8e16": Phase="Pending", Reason="", readiness=false. Elapsed: 18.947573ms -Sep 21 15:44:58.721: INFO: Pod "pod-c6139c3a-6076-431a-856a-50a906af8e16": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03284844s -Sep 21 15:45:00.736: INFO: Pod "pod-c6139c3a-6076-431a-856a-50a906af8e16": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.04752128s -STEP: Saw pod success -Sep 21 15:45:00.736: INFO: Pod "pod-c6139c3a-6076-431a-856a-50a906af8e16" satisfied condition "success or failure" -Sep 21 15:45:00.752: INFO: Trying to get logs from node 10.241.51.150 pod pod-c6139c3a-6076-431a-856a-50a906af8e16 container test-container: -STEP: delete the pod -Sep 21 15:45:00.843: INFO: Waiting for pod pod-c6139c3a-6076-431a-856a-50a906af8e16 to disappear -Sep 21 15:45:00.858: INFO: Pod pod-c6139c3a-6076-431a-856a-50a906af8e16 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: getting the auto-created API token +STEP: reading a file in the container +Oct 26 16:05:11.177: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-2423 pod-service-account-eb35384f-2d80-4af8-bff7-04d1e959bfa4 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token' +STEP: reading a file in the container +Oct 26 16:05:11.579: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-2423 pod-service-account-eb35384f-2d80-4af8-bff7-04d1e959bfa4 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt' +STEP: reading a file in the container +Oct 26 16:05:11.930: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-2423 pod-service-account-eb35384f-2d80-4af8-bff7-04d1e959bfa4 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace' +[AfterEach] [sig-auth] ServiceAccounts /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:45:00.858: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-2416" for this suite. -•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":18,"skipped":297,"failed":0} -SSSSSSSSSSSSS ------------------------------- -[sig-storage] Projected downwardAPI - should provide container's cpu limit [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 15:45:00.909: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename projected -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40 -[It] should provide container's cpu limit [NodeConformance] [Conformance] +Oct 26 16:05:12.315: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svcaccounts-2423" for this suite. + +• [SLOW TEST:5.948 seconds] +[sig-auth] ServiceAccounts +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23 + should mount an API token into pods [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test downward API volume plugin -Sep 21 15:45:01.166: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d3bcccb6-d3a6-405c-8731-f6525ffee194" in namespace "projected-8539" to be "success or failure" -Sep 21 15:45:01.180: INFO: Pod "downwardapi-volume-d3bcccb6-d3a6-405c-8731-f6525ffee194": Phase="Pending", Reason="", readiness=false. Elapsed: 13.93752ms -Sep 21 15:45:03.196: INFO: Pod "downwardapi-volume-d3bcccb6-d3a6-405c-8731-f6525ffee194": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029864812s -Sep 21 15:45:05.211: INFO: Pod "downwardapi-volume-d3bcccb6-d3a6-405c-8731-f6525ffee194": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044545502s -STEP: Saw pod success -Sep 21 15:45:05.211: INFO: Pod "downwardapi-volume-d3bcccb6-d3a6-405c-8731-f6525ffee194" satisfied condition "success or failure" -Sep 21 15:45:05.226: INFO: Trying to get logs from node 10.241.51.150 pod downwardapi-volume-d3bcccb6-d3a6-405c-8731-f6525ffee194 container client-container: -STEP: delete the pod -Sep 21 15:45:05.312: INFO: Waiting for pod downwardapi-volume-d3bcccb6-d3a6-405c-8731-f6525ffee194 to disappear -Sep 21 15:45:05.331: INFO: Pod downwardapi-volume-d3bcccb6-d3a6-405c-8731-f6525ffee194 no longer exists -[AfterEach] [sig-storage] Projected downwardAPI - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:45:05.331: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-8539" for this suite. -•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance]","total":280,"completed":19,"skipped":310,"failed":0} -SSSSSSSSS ------------------------------ -[k8s.io] Variable Expansion - should allow substituting values in a container's command [NodeConformance] [Conformance] +{"msg":"PASSED [sig-auth] ServiceAccounts should mount an API token into pods [Conformance]","total":280,"completed":20,"skipped":293,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] Subpath Atomic writer volumes + should support subpaths with configmap pod [LinuxOnly] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Variable Expansion +[BeforeEach] [sig-storage] Subpath /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:45:05.376: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename var-expansion +Oct 26 16:05:12.351: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename subpath STEP: Waiting for a default service account to be provisioned in namespace -[It] should allow substituting values in a container's command [NodeConformance] [Conformance] +[BeforeEach] Atomic writer volumes + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37 +STEP: Setting up data +[It] should support subpaths with configmap pod [LinuxOnly] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test substitution in container's command -Sep 21 15:45:05.645: INFO: Waiting up to 5m0s for pod "var-expansion-822a37a9-ef81-490d-96d9-2188c6b0c664" in namespace "var-expansion-1279" to be "success or failure" -Sep 21 15:45:05.690: INFO: Pod "var-expansion-822a37a9-ef81-490d-96d9-2188c6b0c664": Phase="Pending", Reason="", readiness=false. Elapsed: 44.859834ms -Sep 21 15:45:07.711: INFO: Pod "var-expansion-822a37a9-ef81-490d-96d9-2188c6b0c664": Phase="Pending", Reason="", readiness=false. Elapsed: 2.065955578s -Sep 21 15:45:09.731: INFO: Pod "var-expansion-822a37a9-ef81-490d-96d9-2188c6b0c664": Phase="Pending", Reason="", readiness=false. Elapsed: 4.085705392s -Sep 21 15:45:11.749: INFO: Pod "var-expansion-822a37a9-ef81-490d-96d9-2188c6b0c664": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.10376852s +STEP: Creating pod pod-subpath-test-configmap-gvgw +STEP: Creating a pod to test atomic-volume-subpath +Oct 26 16:05:12.588: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-gvgw" in namespace "subpath-2538" to be "success or failure" +Oct 26 16:05:12.609: INFO: Pod "pod-subpath-test-configmap-gvgw": Phase="Pending", Reason="", readiness=false. Elapsed: 20.955605ms +Oct 26 16:05:14.627: INFO: Pod "pod-subpath-test-configmap-gvgw": Phase="Running", Reason="", readiness=true. Elapsed: 2.038392956s +Oct 26 16:05:16.658: INFO: Pod "pod-subpath-test-configmap-gvgw": Phase="Running", Reason="", readiness=true. Elapsed: 4.069440419s +Oct 26 16:05:18.685: INFO: Pod "pod-subpath-test-configmap-gvgw": Phase="Running", Reason="", readiness=true. Elapsed: 6.096896842s +Oct 26 16:05:20.700: INFO: Pod "pod-subpath-test-configmap-gvgw": Phase="Running", Reason="", readiness=true. Elapsed: 8.111404067s +Oct 26 16:05:22.716: INFO: Pod "pod-subpath-test-configmap-gvgw": Phase="Running", Reason="", readiness=true. Elapsed: 10.127428008s +Oct 26 16:05:24.733: INFO: Pod "pod-subpath-test-configmap-gvgw": Phase="Running", Reason="", readiness=true. Elapsed: 12.14427763s +Oct 26 16:05:26.752: INFO: Pod "pod-subpath-test-configmap-gvgw": Phase="Running", Reason="", readiness=true. Elapsed: 14.163391108s +Oct 26 16:05:28.767: INFO: Pod "pod-subpath-test-configmap-gvgw": Phase="Running", Reason="", readiness=true. Elapsed: 16.17870694s +Oct 26 16:05:30.783: INFO: Pod "pod-subpath-test-configmap-gvgw": Phase="Running", Reason="", readiness=true. Elapsed: 18.194588584s +Oct 26 16:05:32.801: INFO: Pod "pod-subpath-test-configmap-gvgw": Phase="Running", Reason="", readiness=true. Elapsed: 20.212247122s +Oct 26 16:05:34.816: INFO: Pod "pod-subpath-test-configmap-gvgw": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.228088887s STEP: Saw pod success -Sep 21 15:45:11.749: INFO: Pod "var-expansion-822a37a9-ef81-490d-96d9-2188c6b0c664" satisfied condition "success or failure" -Sep 21 15:45:11.766: INFO: Trying to get logs from node 10.241.51.147 pod var-expansion-822a37a9-ef81-490d-96d9-2188c6b0c664 container dapi-container: +Oct 26 16:05:34.816: INFO: Pod "pod-subpath-test-configmap-gvgw" satisfied condition "success or failure" +Oct 26 16:05:34.830: INFO: Trying to get logs from node 10.72.119.74 pod pod-subpath-test-configmap-gvgw container test-container-subpath-configmap-gvgw: STEP: delete the pod -Sep 21 15:45:11.878: INFO: Waiting for pod var-expansion-822a37a9-ef81-490d-96d9-2188c6b0c664 to disappear -Sep 21 15:45:11.911: INFO: Pod var-expansion-822a37a9-ef81-490d-96d9-2188c6b0c664 no longer exists -[AfterEach] [k8s.io] Variable Expansion +Oct 26 16:05:34.915: INFO: Waiting for pod pod-subpath-test-configmap-gvgw to disappear +Oct 26 16:05:34.935: INFO: Pod pod-subpath-test-configmap-gvgw no longer exists +STEP: Deleting pod pod-subpath-test-configmap-gvgw +Oct 26 16:05:34.935: INFO: Deleting pod "pod-subpath-test-configmap-gvgw" in namespace "subpath-2538" +[AfterEach] [sig-storage] Subpath /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:45:11.912: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-1279" for this suite. +Oct 26 16:05:34.950: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "subpath-2538" for this suite. -• [SLOW TEST:6.590 seconds] -[k8s.io] Variable Expansion -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - should allow substituting values in a container's command [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance]","total":280,"completed":20,"skipped":319,"failed":0} -SSSSSSSSSSSSSS +• [SLOW TEST:22.631 seconds] +[sig-storage] Subpath +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 + Atomic writer volumes + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33 + should support subpaths with configmap pod [LinuxOnly] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -[sig-api-machinery] Servers with support for Table transformation - should return a 406 for a backend which does not implement metadata [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] Servers with support for Table transformation - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 15:45:11.967: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename tables -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-api-machinery] Servers with support for Table transformation - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/table_conversion.go:46 -[It] should return a 406 for a backend which does not implement metadata [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[AfterEach] [sig-api-machinery] Servers with support for Table transformation - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:45:12.191: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "tables-9776" for this suite. -•{"msg":"PASSED [sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance]","total":280,"completed":21,"skipped":333,"failed":0} -SSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [LinuxOnly] [Conformance]","total":280,"completed":21,"skipped":326,"failed":0} +SSS ------------------------------ -[sig-cli] Kubectl client Kubectl rolling-update - should support rolling-update to same image [Conformance] +[sig-cli] Kubectl client Kubectl cluster-info + should check if Kubernetes master services is included in cluster-info [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:45:12.249: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:05:34.982: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278 -[BeforeEach] Kubectl rolling-update - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1692 -[It] should support rolling-update to same image [Conformance] +[It] should check if Kubernetes master services is included in cluster-info [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: running the image docker.io/library/httpd:2.4.38-alpine -Sep 21 15:45:12.386: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 run e2e-test-httpd-rc --image=docker.io/library/httpd:2.4.38-alpine --generator=run/v1 --namespace=kubectl-8928' -Sep 21 15:45:12.559: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" -Sep 21 15:45:12.559: INFO: stdout: "replicationcontroller/e2e-test-httpd-rc created\n" -STEP: verifying the rc e2e-test-httpd-rc was created -Sep 21 15:45:12.626: INFO: Waiting for rc e2e-test-httpd-rc to stabilize, generation 1 observed generation 0 spec.replicas 1 status.replicas 0 -Sep 21 15:45:12.666: INFO: Waiting for rc e2e-test-httpd-rc to stabilize, generation 1 observed generation 1 spec.replicas 1 status.replicas 0 -STEP: rolling-update to same image controller -Sep 21 15:45:12.727: INFO: scanned /root for discovery docs: -Sep 21 15:45:12.728: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 rolling-update e2e-test-httpd-rc --update-period=1s --image=docker.io/library/httpd:2.4.38-alpine --image-pull-policy=IfNotPresent --namespace=kubectl-8928' -Sep 21 15:45:28.952: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n" -Sep 21 15:45:28.952: INFO: stdout: "Created e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f\nScaling up e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f from 0 to 1, scaling down e2e-test-httpd-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f up to 1\nScaling e2e-test-httpd-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-httpd-rc\nRenaming e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f to e2e-test-httpd-rc\nreplicationcontroller/e2e-test-httpd-rc rolling updated\n" -Sep 21 15:45:28.952: INFO: stdout: "Created e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f\nScaling up e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f from 0 to 1, scaling down e2e-test-httpd-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f up to 1\nScaling e2e-test-httpd-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-httpd-rc\nRenaming e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f to e2e-test-httpd-rc\nreplicationcontroller/e2e-test-httpd-rc rolling updated\n" -STEP: waiting for all containers in run=e2e-test-httpd-rc pods to come up. -Sep 21 15:45:28.952: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-httpd-rc --namespace=kubectl-8928' -Sep 21 15:45:29.093: INFO: stderr: "" -Sep 21 15:45:29.093: INFO: stdout: "e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f-dgmk9 e2e-test-httpd-rc-tz9mv " -STEP: Replicas for run=e2e-test-httpd-rc: expected=1 actual=2 -Sep 21 15:45:34.093: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-httpd-rc --namespace=kubectl-8928' -Sep 21 15:45:34.243: INFO: stderr: "" -Sep 21 15:45:34.243: INFO: stdout: "e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f-dgmk9 " -Sep 21 15:45:34.244: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f-dgmk9 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "e2e-test-httpd-rc") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-8928' -Sep 21 15:45:34.372: INFO: stderr: "" -Sep 21 15:45:34.372: INFO: stdout: "true" -Sep 21 15:45:34.372: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f-dgmk9 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "e2e-test-httpd-rc"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-8928' -Sep 21 15:45:34.951: INFO: stderr: "" -Sep 21 15:45:34.951: INFO: stdout: "docker.io/library/httpd:2.4.38-alpine" -Sep 21 15:45:34.952: INFO: e2e-test-httpd-rc-fc1d00fbb9ef26351f7785fe26937f9f-dgmk9 is verified up and running -[AfterEach] Kubectl rolling-update - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1698 -Sep 21 15:45:34.952: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete rc e2e-test-httpd-rc --namespace=kubectl-8928' -Sep 21 15:45:35.145: INFO: stderr: "" -Sep 21 15:45:35.145: INFO: stdout: "replicationcontroller \"e2e-test-httpd-rc\" deleted\n" +STEP: validating cluster-info +Oct 26 16:05:35.126: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 cluster-info' +Oct 26 16:05:35.265: INFO: stderr: "" +Oct 26 16:05:35.265: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:45:35.145: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-8928" for this suite. - -• [SLOW TEST:22.944 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 - Kubectl rolling-update - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1687 - should support rolling-update to same image [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-cli] Kubectl client Kubectl rolling-update should support rolling-update to same image [Conformance]","total":280,"completed":22,"skipped":355,"failed":0} -SSSSSS +Oct 26 16:05:35.265: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-3582" for this suite. +•{"msg":"PASSED [sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes master services is included in cluster-info [Conformance]","total":280,"completed":22,"skipped":329,"failed":0} +SS ------------------------------ -[sig-network] Services - should be able to change the type from ExternalName to ClusterIP [Conformance] +[k8s.io] Probing container + should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-network] Services +[BeforeEach] [k8s.io] Probing container /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:45:35.193: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename services +Oct 26 16:05:35.301: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-network] Services - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139 -[It] should be able to change the type from ExternalName to ClusterIP [Conformance] +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 +[It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: creating a service externalname-service with the type=ExternalName in namespace services-6734 -STEP: changing the ExternalName service to type=ClusterIP -STEP: creating replication controller externalname-service in namespace services-6734 -I0921 15:45:36.018923 24 runners.go:189] Created replication controller with name: externalname-service, namespace: services-6734, replica count: 2 -I0921 15:45:39.069688 24 runners.go:189] externalname-service Pods: 2 out of 2 created, 1 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Sep 21 15:45:42.070: INFO: Creating new exec pod -I0921 15:45:42.069957 24 runners.go:189] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Sep 21 15:45:47.146: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-6734 execpodlkccf -- /bin/sh -x -c nc -zv -t -w 2 externalname-service 80' -Sep 21 15:45:48.635: INFO: stderr: "+ nc -zv -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" -Sep 21 15:45:48.635: INFO: stdout: "" -Sep 21 15:45:48.636: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-6734 execpodlkccf -- /bin/sh -x -c nc -zv -t -w 2 172.21.135.101 80' -Sep 21 15:45:49.005: INFO: stderr: "+ nc -zv -t -w 2 172.21.135.101 80\nConnection to 172.21.135.101 80 port [tcp/http] succeeded!\n" -Sep 21 15:45:49.005: INFO: stdout: "" -Sep 21 15:45:49.005: INFO: Cleaning up the ExternalName to ClusterIP test service -[AfterEach] [sig-network] Services +STEP: Creating pod busybox-6248fa34-643c-46ec-a13b-1c09aabe222e in namespace container-probe-8395 +Oct 26 16:05:37.559: INFO: Started pod busybox-6248fa34-643c-46ec-a13b-1c09aabe222e in namespace container-probe-8395 +STEP: checking the pod's current state and verifying that restartCount is present +Oct 26 16:05:37.575: INFO: Initial restart count of pod busybox-6248fa34-643c-46ec-a13b-1c09aabe222e is 0 +Oct 26 16:06:28.031: INFO: Restart count of pod container-probe-8395/busybox-6248fa34-643c-46ec-a13b-1c09aabe222e is now 1 (50.456053971s elapsed) +STEP: deleting the pod +[AfterEach] [k8s.io] Probing container /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:45:49.886: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-6734" for this suite. -[AfterEach] [sig-network] Services - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143 +Oct 26 16:06:28.083: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-8395" for this suite. -• [SLOW TEST:14.737 seconds] -[sig-network] Services -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 - should be able to change the type from ExternalName to ClusterIP [Conformance] +• [SLOW TEST:52.816 seconds] +[k8s.io] Probing container +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 + should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance]","total":280,"completed":23,"skipped":361,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [k8s.io] Probing container should be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":280,"completed":23,"skipped":331,"failed":0} +SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] Garbage collector - should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] + should orphan pods created by rc if delete options say so [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 [BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:45:49.930: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:06:28.121: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace -[It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] +[It] should orphan pods created by rc if delete options say so [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 STEP: create the rc STEP: delete the rc STEP: wait for the rc to be deleted +STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods STEP: Gathering metrics -W0921 15:45:56.292175 24 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. -Sep 21 15:45:56.292: INFO: For apiserver_request_total: +W1026 16:07:08.412717 26 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +Oct 26 16:07:08.412: INFO: For apiserver_request_total: For apiserver_request_latency_seconds: For apiserver_init_events_total: For garbage_collector_attempt_to_delete_queue_latency: @@ -906,2521 +874,2368 @@ For evicted_pods_total: [AfterEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:45:56.292: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "gc-3553" for this suite. +Oct 26 16:07:08.413: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-6913" for this suite. -• [SLOW TEST:6.412 seconds] +• [SLOW TEST:40.325 seconds] [sig-api-machinery] Garbage collector /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] + should orphan pods created by rc if delete options say so [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]","total":280,"completed":24,"skipped":388,"failed":0} -SSSSSSSSS +{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance]","total":280,"completed":24,"skipped":352,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Downward API volume - should provide container's memory request [NodeConformance] [Conformance] +[sig-node] Downward API + should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] Downward API volume +[BeforeEach] [sig-node] Downward API /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:45:56.343: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:07:08.449: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40 -[It] should provide container's memory request [NodeConformance] [Conformance] +[It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test downward API volume plugin -Sep 21 15:45:56.623: INFO: Waiting up to 5m0s for pod "downwardapi-volume-90260f6d-632b-4a81-9608-6b15debf247c" in namespace "downward-api-563" to be "success or failure" -Sep 21 15:45:56.645: INFO: Pod "downwardapi-volume-90260f6d-632b-4a81-9608-6b15debf247c": Phase="Pending", Reason="", readiness=false. Elapsed: 21.313583ms -Sep 21 15:45:58.660: INFO: Pod "downwardapi-volume-90260f6d-632b-4a81-9608-6b15debf247c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03676439s -Sep 21 15:46:00.678: INFO: Pod "downwardapi-volume-90260f6d-632b-4a81-9608-6b15debf247c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.055073178s +STEP: Creating a pod to test downward api env vars +Oct 26 16:07:08.713: INFO: Waiting up to 5m0s for pod "downward-api-4ffa5a15-f408-4e6e-a188-e7b436ba2539" in namespace "downward-api-3833" to be "success or failure" +Oct 26 16:07:08.734: INFO: Pod "downward-api-4ffa5a15-f408-4e6e-a188-e7b436ba2539": Phase="Pending", Reason="", readiness=false. Elapsed: 20.922248ms +Oct 26 16:07:10.762: INFO: Pod "downward-api-4ffa5a15-f408-4e6e-a188-e7b436ba2539": Phase="Pending", Reason="", readiness=false. Elapsed: 2.04942372s +Oct 26 16:07:12.778: INFO: Pod "downward-api-4ffa5a15-f408-4e6e-a188-e7b436ba2539": Phase="Pending", Reason="", readiness=false. Elapsed: 4.065274478s +Oct 26 16:07:14.801: INFO: Pod "downward-api-4ffa5a15-f408-4e6e-a188-e7b436ba2539": Phase="Pending", Reason="", readiness=false. Elapsed: 6.0881009s +Oct 26 16:07:16.824: INFO: Pod "downward-api-4ffa5a15-f408-4e6e-a188-e7b436ba2539": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.110741631s STEP: Saw pod success -Sep 21 15:46:00.678: INFO: Pod "downwardapi-volume-90260f6d-632b-4a81-9608-6b15debf247c" satisfied condition "success or failure" -Sep 21 15:46:00.693: INFO: Trying to get logs from node 10.241.51.150 pod downwardapi-volume-90260f6d-632b-4a81-9608-6b15debf247c container client-container: +Oct 26 16:07:16.824: INFO: Pod "downward-api-4ffa5a15-f408-4e6e-a188-e7b436ba2539" satisfied condition "success or failure" +Oct 26 16:07:16.839: INFO: Trying to get logs from node 10.72.119.82 pod downward-api-4ffa5a15-f408-4e6e-a188-e7b436ba2539 container dapi-container: STEP: delete the pod -Sep 21 15:46:00.778: INFO: Waiting for pod downwardapi-volume-90260f6d-632b-4a81-9608-6b15debf247c to disappear -Sep 21 15:46:00.792: INFO: Pod downwardapi-volume-90260f6d-632b-4a81-9608-6b15debf247c no longer exists -[AfterEach] [sig-storage] Downward API volume - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:46:00.792: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-563" for this suite. -•{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance]","total":280,"completed":25,"skipped":397,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - removes definition from spec when one version gets changed to not be served [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 15:46:00.840: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename crd-publish-openapi -STEP: Waiting for a default service account to be provisioned in namespace -[It] removes definition from spec when one version gets changed to not be served [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: set up a multi version CRD -Sep 21 15:46:01.050: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: mark a version not serverd -STEP: check the unserved version gets removed -STEP: check the other version is not changed -[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +Oct 26 16:07:16.971: INFO: Waiting for pod downward-api-4ffa5a15-f408-4e6e-a188-e7b436ba2539 to disappear +Oct 26 16:07:16.988: INFO: Pod downward-api-4ffa5a15-f408-4e6e-a188-e7b436ba2539 no longer exists +[AfterEach] [sig-node] Downward API /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:46:42.595: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-publish-openapi-9700" for this suite. +Oct 26 16:07:16.989: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-3833" for this suite. -• [SLOW TEST:41.804 seconds] -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - removes definition from spec when one version gets changed to not be served [Conformance] +• [SLOW TEST:8.575 seconds] +[sig-node] Downward API +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:33 + should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance]","total":280,"completed":26,"skipped":428,"failed":0} -SSSSSS +{"msg":"PASSED [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]","total":280,"completed":25,"skipped":392,"failed":0} +SSSSSSSSSSSSSSSSSSS ------------------------------ -[k8s.io] Probing container - should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] +[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Probing container +[BeforeEach] [sig-apps] StatefulSet /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:46:42.644: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename container-probe +Oct 26 16:07:17.026: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename statefulset STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Probing container - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 -[It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64 +[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79 +STEP: Creating service test in namespace statefulset-1932 +[It] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating pod test-webserver-539f0ce6-2a0c-499d-b863-216db81cd8c7 in namespace container-probe-8482 -Sep 21 15:46:46.941: INFO: Started pod test-webserver-539f0ce6-2a0c-499d-b863-216db81cd8c7 in namespace container-probe-8482 -STEP: checking the pod's current state and verifying that restartCount is present -Sep 21 15:46:46.962: INFO: Initial restart count of pod test-webserver-539f0ce6-2a0c-499d-b863-216db81cd8c7 is 0 -STEP: deleting the pod -[AfterEach] [k8s.io] Probing container +STEP: Creating stateful set ss in namespace statefulset-1932 +STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-1932 +Oct 26 16:07:17.238: INFO: Found 0 stateful pods, waiting for 1 +Oct 26 16:07:27.258: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Pending - Ready=false +Oct 26 16:07:37.262: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod +Oct 26 16:07:37.277: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-1932 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Oct 26 16:07:37.717: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Oct 26 16:07:37.717: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Oct 26 16:07:37.717: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Oct 26 16:07:37.733: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true +Oct 26 16:07:47.762: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false +Oct 26 16:07:47.762: INFO: Waiting for statefulset status.replicas updated to 0 +Oct 26 16:07:47.833: INFO: POD NODE PHASE GRACE CONDITIONS +Oct 26 16:07:47.833: INFO: ss-0 10.72.119.74 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:38 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:38 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC }] +Oct 26 16:07:47.833: INFO: +Oct 26 16:07:47.833: INFO: StatefulSet ss has not reached scale 3, at 1 +Oct 26 16:07:48.885: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.981875524s +Oct 26 16:07:49.901: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.929738792s +Oct 26 16:07:50.918: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.913722728s +Oct 26 16:07:51.934: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.896822391s +Oct 26 16:07:52.951: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.880176956s +Oct 26 16:07:53.974: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.863142856s +Oct 26 16:07:54.990: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.840824086s +Oct 26 16:07:56.006: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.824464086s +Oct 26 16:07:57.023: INFO: Verifying statefulset ss doesn't scale past 3 for another 808.264113ms +STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-1932 +Oct 26 16:07:58.041: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-1932 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Oct 26 16:07:58.427: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" +Oct 26 16:07:58.427: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Oct 26 16:07:58.427: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Oct 26 16:07:58.427: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-1932 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Oct 26 16:07:58.821: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" +Oct 26 16:07:58.821: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Oct 26 16:07:58.821: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Oct 26 16:07:58.821: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-1932 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Oct 26 16:07:59.282: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" +Oct 26 16:07:59.282: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Oct 26 16:07:59.282: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Oct 26 16:07:59.302: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +Oct 26 16:07:59.302: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true +Oct 26 16:07:59.302: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Pending - Ready=false +Oct 26 16:08:09.327: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +Oct 26 16:08:09.327: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true +Oct 26 16:08:09.327: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true +STEP: Scale down will not halt with unhealthy stateful pod +Oct 26 16:08:09.345: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-1932 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Oct 26 16:08:09.691: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Oct 26 16:08:09.691: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Oct 26 16:08:09.691: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Oct 26 16:08:09.691: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-1932 ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Oct 26 16:08:10.112: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Oct 26 16:08:10.112: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Oct 26 16:08:10.112: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Oct 26 16:08:10.112: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-1932 ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Oct 26 16:08:10.501: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Oct 26 16:08:10.501: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Oct 26 16:08:10.501: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Oct 26 16:08:10.501: INFO: Waiting for statefulset status.replicas updated to 0 +Oct 26 16:08:10.518: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2 +Oct 26 16:08:20.552: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false +Oct 26 16:08:20.552: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false +Oct 26 16:08:20.552: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false +Oct 26 16:08:20.601: INFO: POD NODE PHASE GRACE CONDITIONS +Oct 26 16:08:20.601: INFO: ss-0 10.72.119.74 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC }] +Oct 26 16:08:20.601: INFO: ss-1 10.72.119.82 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:20.601: INFO: ss-2 10.72.119.76 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:20.601: INFO: +Oct 26 16:08:20.601: INFO: StatefulSet ss has not reached scale 0, at 3 +Oct 26 16:08:21.618: INFO: POD NODE PHASE GRACE CONDITIONS +Oct 26 16:08:21.618: INFO: ss-0 10.72.119.74 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC }] +Oct 26 16:08:21.618: INFO: ss-1 10.72.119.82 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:21.618: INFO: ss-2 10.72.119.76 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:21.618: INFO: +Oct 26 16:08:21.618: INFO: StatefulSet ss has not reached scale 0, at 3 +Oct 26 16:08:22.639: INFO: POD NODE PHASE GRACE CONDITIONS +Oct 26 16:08:22.639: INFO: ss-0 10.72.119.74 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC }] +Oct 26 16:08:22.639: INFO: ss-1 10.72.119.82 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:22.639: INFO: ss-2 10.72.119.76 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:22.639: INFO: +Oct 26 16:08:22.639: INFO: StatefulSet ss has not reached scale 0, at 3 +Oct 26 16:08:23.655: INFO: POD NODE PHASE GRACE CONDITIONS +Oct 26 16:08:23.655: INFO: ss-0 10.72.119.74 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC }] +Oct 26 16:08:23.655: INFO: ss-1 10.72.119.82 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:23.655: INFO: ss-2 10.72.119.76 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:23.655: INFO: +Oct 26 16:08:23.655: INFO: StatefulSet ss has not reached scale 0, at 3 +Oct 26 16:08:24.671: INFO: POD NODE PHASE GRACE CONDITIONS +Oct 26 16:08:24.671: INFO: ss-0 10.72.119.74 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC }] +Oct 26 16:08:24.672: INFO: ss-1 10.72.119.82 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:24.672: INFO: ss-2 10.72.119.76 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:24.672: INFO: +Oct 26 16:08:24.672: INFO: StatefulSet ss has not reached scale 0, at 3 +Oct 26 16:08:25.696: INFO: POD NODE PHASE GRACE CONDITIONS +Oct 26 16:08:25.696: INFO: ss-0 10.72.119.74 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:17 +0000 UTC }] +Oct 26 16:08:25.696: INFO: ss-1 10.72.119.82 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:25.696: INFO: ss-2 10.72.119.76 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:25.696: INFO: +Oct 26 16:08:25.696: INFO: StatefulSet ss has not reached scale 0, at 3 +Oct 26 16:08:26.719: INFO: POD NODE PHASE GRACE CONDITIONS +Oct 26 16:08:26.719: INFO: ss-1 10.72.119.82 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:26.719: INFO: ss-2 10.72.119.76 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:26.719: INFO: +Oct 26 16:08:26.719: INFO: StatefulSet ss has not reached scale 0, at 2 +Oct 26 16:08:27.734: INFO: POD NODE PHASE GRACE CONDITIONS +Oct 26 16:08:27.734: INFO: ss-1 10.72.119.82 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:27.734: INFO: ss-2 10.72.119.76 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:11 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:27.734: INFO: +Oct 26 16:08:27.734: INFO: StatefulSet ss has not reached scale 0, at 2 +Oct 26 16:08:28.752: INFO: POD NODE PHASE GRACE CONDITIONS +Oct 26 16:08:28.752: INFO: ss-1 10.72.119.82 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:08:10 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-10-26 16:07:47 +0000 UTC }] +Oct 26 16:08:28.752: INFO: +Oct 26 16:08:28.752: INFO: StatefulSet ss has not reached scale 0, at 1 +Oct 26 16:08:29.768: INFO: Verifying statefulset ss doesn't scale past 0 for another 833.920134ms +STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-1932 +Oct 26 16:08:30.785: INFO: Scaling statefulset ss to 0 +Oct 26 16:08:30.837: INFO: Waiting for statefulset status.replicas updated to 0 +[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 +Oct 26 16:08:30.852: INFO: Deleting all statefulset in ns statefulset-1932 +Oct 26 16:08:30.868: INFO: Scaling statefulset ss to 0 +Oct 26 16:08:30.923: INFO: Waiting for statefulset status.replicas updated to 0 +Oct 26 16:08:30.939: INFO: Deleting statefulset ss +[AfterEach] [sig-apps] StatefulSet /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:50:47.209: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-8482" for this suite. +Oct 26 16:08:31.019: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-1932" for this suite. -• [SLOW TEST:244.621 seconds] -[k8s.io] Probing container -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +• [SLOW TEST:74.025 seconds] +[sig-apps] StatefulSet +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 + [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 + Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":280,"completed":27,"skipped":434,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]","total":280,"completed":26,"skipped":411,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] +[sig-storage] ConfigMap + should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] EmptyDir volumes +[BeforeEach] [sig-storage] ConfigMap /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:50:47.266: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename emptydir +Oct 26 16:08:31.051: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename configmap STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] +[It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test emptydir 0666 on tmpfs -Sep 21 15:50:47.581: INFO: Waiting up to 5m0s for pod "pod-4868cd15-2f75-4ae0-9007-c1c52c527a9a" in namespace "emptydir-2748" to be "success or failure" -Sep 21 15:50:47.604: INFO: Pod "pod-4868cd15-2f75-4ae0-9007-c1c52c527a9a": Phase="Pending", Reason="", readiness=false. Elapsed: 22.684558ms -Sep 21 15:50:49.621: INFO: Pod "pod-4868cd15-2f75-4ae0-9007-c1c52c527a9a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.04026989s -Sep 21 15:50:51.636: INFO: Pod "pod-4868cd15-2f75-4ae0-9007-c1c52c527a9a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.054795842s +STEP: Creating configMap with name configmap-test-volume-fd491c96-d1b3-4f80-92d3-e86ab80e4d78 +STEP: Creating a pod to test consume configMaps +Oct 26 16:08:31.272: INFO: Waiting up to 5m0s for pod "pod-configmaps-24cf4445-9f39-4e4c-a0ea-6d0af54c2346" in namespace "configmap-380" to be "success or failure" +Oct 26 16:08:31.296: INFO: Pod "pod-configmaps-24cf4445-9f39-4e4c-a0ea-6d0af54c2346": Phase="Pending", Reason="", readiness=false. Elapsed: 24.474277ms +Oct 26 16:08:33.311: INFO: Pod "pod-configmaps-24cf4445-9f39-4e4c-a0ea-6d0af54c2346": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.039542303s STEP: Saw pod success -Sep 21 15:50:51.636: INFO: Pod "pod-4868cd15-2f75-4ae0-9007-c1c52c527a9a" satisfied condition "success or failure" -Sep 21 15:50:51.649: INFO: Trying to get logs from node 10.241.51.147 pod pod-4868cd15-2f75-4ae0-9007-c1c52c527a9a container test-container: +Oct 26 16:08:33.311: INFO: Pod "pod-configmaps-24cf4445-9f39-4e4c-a0ea-6d0af54c2346" satisfied condition "success or failure" +Oct 26 16:08:33.331: INFO: Trying to get logs from node 10.72.119.74 pod pod-configmaps-24cf4445-9f39-4e4c-a0ea-6d0af54c2346 container configmap-volume-test: STEP: delete the pod -Sep 21 15:50:51.775: INFO: Waiting for pod pod-4868cd15-2f75-4ae0-9007-c1c52c527a9a to disappear -Sep 21 15:50:51.799: INFO: Pod pod-4868cd15-2f75-4ae0-9007-c1c52c527a9a no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +Oct 26 16:08:33.442: INFO: Waiting for pod pod-configmaps-24cf4445-9f39-4e4c-a0ea-6d0af54c2346 to disappear +Oct 26 16:08:33.461: INFO: Pod pod-configmaps-24cf4445-9f39-4e4c-a0ea-6d0af54c2346 no longer exists +[AfterEach] [sig-storage] ConfigMap /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:50:51.799: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-2748" for this suite. -•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":28,"skipped":520,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +Oct 26 16:08:33.461: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-380" for this suite. +•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":27,"skipped":436,"failed":0} +SSSSSS ------------------------------ -[sig-api-machinery] Watchers - should receive events on concurrent watches in same order [Conformance] +[sig-apps] Deployment + deployment should support proportional scaling [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] Watchers +[BeforeEach] [sig-apps] Deployment /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:50:51.840: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename watch +Oct 26 16:08:33.497: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename deployment STEP: Waiting for a default service account to be provisioned in namespace -[It] should receive events on concurrent watches in same order [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: starting a background goroutine to produce watch events -STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order -[AfterEach] [sig-api-machinery] Watchers - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:50:56.260: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "watch-1596" for this suite. -•{"msg":"PASSED [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance]","total":280,"completed":29,"skipped":574,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-storage] Projected configMap - should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] +[BeforeEach] [sig-apps] Deployment + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69 +[It] deployment should support proportional scaling [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] Projected configMap - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 15:50:56.366: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename projected -STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating configMap with name projected-configmap-test-volume-map-ff901adb-33a1-4fcb-aba3-ac73e99d1d50 -STEP: Creating a pod to test consume configMaps -Sep 21 15:50:56.596: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-60ca611b-d45a-4958-b586-b0979d08eadf" in namespace "projected-3529" to be "success or failure" -Sep 21 15:50:56.613: INFO: Pod "pod-projected-configmaps-60ca611b-d45a-4958-b586-b0979d08eadf": Phase="Pending", Reason="", readiness=false. Elapsed: 16.284533ms -Sep 21 15:50:58.629: INFO: Pod "pod-projected-configmaps-60ca611b-d45a-4958-b586-b0979d08eadf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.032850351s -Sep 21 15:51:00.645: INFO: Pod "pod-projected-configmaps-60ca611b-d45a-4958-b586-b0979d08eadf": Phase="Pending", Reason="", readiness=false. Elapsed: 4.049027775s -Sep 21 15:51:02.669: INFO: Pod "pod-projected-configmaps-60ca611b-d45a-4958-b586-b0979d08eadf": Phase="Pending", Reason="", readiness=false. Elapsed: 6.072113584s -Sep 21 15:51:04.699: INFO: Pod "pod-projected-configmaps-60ca611b-d45a-4958-b586-b0979d08eadf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.102209569s -STEP: Saw pod success -Sep 21 15:51:04.699: INFO: Pod "pod-projected-configmaps-60ca611b-d45a-4958-b586-b0979d08eadf" satisfied condition "success or failure" -Sep 21 15:51:04.719: INFO: Trying to get logs from node 10.241.51.150 pod pod-projected-configmaps-60ca611b-d45a-4958-b586-b0979d08eadf container projected-configmap-volume-test: -STEP: delete the pod -Sep 21 15:51:04.863: INFO: Waiting for pod pod-projected-configmaps-60ca611b-d45a-4958-b586-b0979d08eadf to disappear -Sep 21 15:51:04.877: INFO: Pod pod-projected-configmaps-60ca611b-d45a-4958-b586-b0979d08eadf no longer exists -[AfterEach] [sig-storage] Projected configMap +Oct 26 16:08:33.620: INFO: Creating deployment "webserver-deployment" +Oct 26 16:08:33.651: INFO: Waiting for observed generation 1 +Oct 26 16:08:35.696: INFO: Waiting for all required pods to come up +Oct 26 16:08:35.713: INFO: Pod name httpd: Found 10 pods out of 10 +STEP: ensuring each pod is running +Oct 26 16:08:37.759: INFO: Waiting for deployment "webserver-deployment" to complete +Oct 26 16:08:37.796: INFO: Updating deployment "webserver-deployment" with a non-existent image +Oct 26 16:08:37.900: INFO: Updating deployment webserver-deployment +Oct 26 16:08:37.900: INFO: Waiting for observed generation 2 +Oct 26 16:08:39.944: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 +Oct 26 16:08:39.963: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 +Oct 26 16:08:39.978: INFO: Waiting for the first rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas +Oct 26 16:08:40.026: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 +Oct 26 16:08:40.026: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 +Oct 26 16:08:40.041: INFO: Waiting for the second rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas +Oct 26 16:08:40.072: INFO: Verifying that deployment "webserver-deployment" has minimum required number of available replicas +Oct 26 16:08:40.072: INFO: Scaling up the deployment "webserver-deployment" from 10 to 30 +Oct 26 16:08:40.156: INFO: Updating deployment webserver-deployment +Oct 26 16:08:40.156: INFO: Waiting for the replicasets of deployment "webserver-deployment" to have desired number of replicas +Oct 26 16:08:40.201: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 +Oct 26 16:08:40.220: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 +[AfterEach] [sig-apps] Deployment + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63 +Oct 26 16:08:40.258: INFO: Deployment "webserver-deployment": +&Deployment{ObjectMeta:{webserver-deployment deployment-4122 /apis/apps/v1/namespaces/deployment-4122/deployments/webserver-deployment 4adda90e-94d0-4289-815f-6d5f4ac4cc98 46236 3 2020-10-26 16:08:33 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:2] [] [] []},Spec:DeploymentSpec{Replicas:*30,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00498a088 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} [] nil default-scheduler [] [] nil [] map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:13,UpdatedReplicas:5,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "webserver-deployment-c7997dcc8" is progressing.,LastUpdateTime:2020-10-26 16:08:38 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,},DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2020-10-26 16:08:40 +0000 UTC,LastTransitionTime:2020-10-26 16:08:40 +0000 UTC,},},ReadyReplicas:8,CollisionCount:nil,},} + +Oct 26 16:08:40.273: INFO: New ReplicaSet "webserver-deployment-c7997dcc8" of Deployment "webserver-deployment": +&ReplicaSet{ObjectMeta:{webserver-deployment-c7997dcc8 deployment-4122 /apis/apps/v1/namespaces/deployment-4122/replicasets/webserver-deployment-c7997dcc8 7639f2f5-70bd-463d-8250-065c8b3bc6dd 46231 3 2020-10-26 16:08:37 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment webserver-deployment 4adda90e-94d0-4289-815f-6d5f4ac4cc98 0xc0049d79d7 0xc0049d79d8}] [] []},Spec:ReplicaSetSpec{Replicas:*13,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: c7997dcc8,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0049d7a48 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:5,FullyLabeledReplicas:5,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} +Oct 26 16:08:40.274: INFO: All old ReplicaSets of Deployment "webserver-deployment": +Oct 26 16:08:40.274: INFO: &ReplicaSet{ObjectMeta:{webserver-deployment-595b5b9587 deployment-4122 /apis/apps/v1/namespaces/deployment-4122/replicasets/webserver-deployment-595b5b9587 d3eaf24a-d135-44ef-bb31-a6db652e7724 46229 3 2020-10-26 16:08:33 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment webserver-deployment 4adda90e-94d0-4289-815f-6d5f4ac4cc98 0xc0049d7917 0xc0049d7918}] [] []},Spec:ReplicaSetSpec{Replicas:*20,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 595b5b9587,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0049d7978 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:8,FullyLabeledReplicas:8,ObservedGeneration:2,ReadyReplicas:8,AvailableReplicas:8,Conditions:[]ReplicaSetCondition{},},} +Oct 26 16:08:40.294: INFO: Pod "webserver-deployment-595b5b9587-2d6r6" is available: +&Pod{ObjectMeta:{webserver-deployment-595b5b9587-2d6r6 webserver-deployment-595b5b9587- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-595b5b9587-2d6r6 e649df21-6564-41f4-8d46-e96dcb9d7e62 46035 0 2020-10-26 16:08:33 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.54.152/32 cni.projectcalico.org/podIPs:172.30.54.152/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.54.152" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 d3eaf24a-d135-44ef-bb31-a6db652e7724 0xc0049d7f37 0xc0049d7f38}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.82,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:35 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:35 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.82,PodIP:172.30.54.152,StartTime:2020-10-26 16:08:33 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:08:35 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://a50787caed937cd617d3633274b46d2e3a5854ec6bd7b75ae4e58207886f6cbb,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.54.152,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.296: INFO: Pod "webserver-deployment-595b5b9587-44ssx" is available: +&Pod{ObjectMeta:{webserver-deployment-595b5b9587-44ssx webserver-deployment-595b5b9587- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-595b5b9587-44ssx 7314b433-3383-420d-9a5d-a1269859f242 46029 0 2020-10-26 16:08:33 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.54.151/32 cni.projectcalico.org/podIPs:172.30.54.151/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.54.151" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 d3eaf24a-d135-44ef-bb31-a6db652e7724 0xc00484e0f7 0xc00484e0f8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.82,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:35 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:35 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.82,PodIP:172.30.54.151,StartTime:2020-10-26 16:08:33 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:08:35 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://6617bb96b6b9552c3da69a08116c528bc324b8c9a959910db9c4384056330a0f,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.54.151,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.297: INFO: Pod "webserver-deployment-595b5b9587-8t88q" is available: +&Pod{ObjectMeta:{webserver-deployment-595b5b9587-8t88q webserver-deployment-595b5b9587- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-595b5b9587-8t88q 6b2d832f-ebe9-46db-a48c-d1eab1c406ad 46041 0 2020-10-26 16:08:33 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.194.88/32 cni.projectcalico.org/podIPs:172.30.194.88/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.194.88" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 d3eaf24a-d135-44ef-bb31-a6db652e7724 0xc00484e2b7 0xc00484e2b8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.74,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.74,PodIP:172.30.194.88,StartTime:2020-10-26 16:08:33 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:08:35 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://d277c15da1d62994b6fd8b0a015e0c3b9330bc687f315160df98cdbef10d02ee,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.194.88,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.297: INFO: Pod "webserver-deployment-595b5b9587-djv49" is available: +&Pod{ObjectMeta:{webserver-deployment-595b5b9587-djv49 webserver-deployment-595b5b9587- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-595b5b9587-djv49 0b00f65d-40eb-4dac-93b2-5c0a29b845e9 46045 0 2020-10-26 16:08:33 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.194.87/32 cni.projectcalico.org/podIPs:172.30.194.87/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.194.87" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 d3eaf24a-d135-44ef-bb31-a6db652e7724 0xc00484e477 0xc00484e478}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.74,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.74,PodIP:172.30.194.87,StartTime:2020-10-26 16:08:33 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:08:35 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://79d043a87a5f657276fde83a0ef987df2583db65aa62911abaa6b2d9550b9f89,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.194.87,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.297: INFO: Pod "webserver-deployment-595b5b9587-dr5tj" is available: +&Pod{ObjectMeta:{webserver-deployment-595b5b9587-dr5tj webserver-deployment-595b5b9587- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-595b5b9587-dr5tj 121d7df1-f321-47ce-9e7e-943693f9f9e9 46062 0 2020-10-26 16:08:33 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.103.242/32 cni.projectcalico.org/podIPs:172.30.103.242/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.103.242" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 d3eaf24a-d135-44ef-bb31-a6db652e7724 0xc00484e617 0xc00484e618}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.76,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.76,PodIP:172.30.103.242,StartTime:2020-10-26 16:08:33 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:08:35 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://b3f224b9dc0c4b33be60ade0425736157baece846e752dec7c484feb1dd9f8b1,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.103.242,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.297: INFO: Pod "webserver-deployment-595b5b9587-g5nrc" is not available: +&Pod{ObjectMeta:{webserver-deployment-595b5b9587-g5nrc webserver-deployment-595b5b9587- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-595b5b9587-g5nrc a2f598ee-ae1a-4580-83e4-21295f4e2487 46245 0 2020-10-26 16:08:40 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 d3eaf24a-d135-44ef-bb31-a6db652e7724 0xc00484e7b7 0xc00484e7b8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.82,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:40 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:40 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:40 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.82,PodIP:,StartTime:2020-10-26 16:08:40 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.297: INFO: Pod "webserver-deployment-595b5b9587-h77tp" is available: +&Pod{ObjectMeta:{webserver-deployment-595b5b9587-h77tp webserver-deployment-595b5b9587- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-595b5b9587-h77tp c0d98daf-eb07-43a6-baf0-6dc0675f6ef3 46066 0 2020-10-26 16:08:33 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.103.241/32 cni.projectcalico.org/podIPs:172.30.103.241/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.103.241" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 d3eaf24a-d135-44ef-bb31-a6db652e7724 0xc00484e937 0xc00484e938}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.76,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.76,PodIP:172.30.103.241,StartTime:2020-10-26 16:08:33 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:08:35 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://d4a016dc4888fd4f9240801617fe029cd0f53f0a350f6bd0571600d046920da2,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.103.241,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.298: INFO: Pod "webserver-deployment-595b5b9587-rz5w6" is available: +&Pod{ObjectMeta:{webserver-deployment-595b5b9587-rz5w6 webserver-deployment-595b5b9587- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-595b5b9587-rz5w6 2751c679-38bf-4405-9262-3c2b6f6ef1bc 46055 0 2020-10-26 16:08:33 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.103.240/32 cni.projectcalico.org/podIPs:172.30.103.240/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.103.240" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 d3eaf24a-d135-44ef-bb31-a6db652e7724 0xc00484ead7 0xc00484ead8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.76,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.76,PodIP:172.30.103.240,StartTime:2020-10-26 16:08:33 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:08:35 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://5bd7a59f7818b3bc4dde9b9877420359b42a4ecdf662ab345f5039234011c54b,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.103.240,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.298: INFO: Pod "webserver-deployment-595b5b9587-thp4p" is available: +&Pod{ObjectMeta:{webserver-deployment-595b5b9587-thp4p webserver-deployment-595b5b9587- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-595b5b9587-thp4p 39e1512d-55ac-42c6-b7c9-1034524c4e5b 46038 0 2020-10-26 16:08:33 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.194.76/32 cni.projectcalico.org/podIPs:172.30.194.76/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.194.76" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 d3eaf24a-d135-44ef-bb31-a6db652e7724 0xc00484ec97 0xc00484ec98}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.74,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:36 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:33 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.74,PodIP:172.30.194.76,StartTime:2020-10-26 16:08:33 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:08:35 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://5dfc0890aa2f1f6bcc0badcb2746ab86703b0994f7c4a46dcf8619de2ffd5651,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.194.76,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.299: INFO: Pod "webserver-deployment-595b5b9587-wbscq" is not available: +&Pod{ObjectMeta:{webserver-deployment-595b5b9587-wbscq webserver-deployment-595b5b9587- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-595b5b9587-wbscq 22f10ace-d73f-41e0-81a8-1577a2478071 46242 0 2020-10-26 16:08:40 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 d3eaf24a-d135-44ef-bb31-a6db652e7724 0xc00484ee37 0xc00484ee38}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.82,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:40 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.299: INFO: Pod "webserver-deployment-595b5b9587-z5snq" is not available: +&Pod{ObjectMeta:{webserver-deployment-595b5b9587-z5snq webserver-deployment-595b5b9587- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-595b5b9587-z5snq 360ac821-fe56-4c6e-ab90-46a28bb40628 46243 0 2020-10-26 16:08:40 +0000 UTC map[name:httpd pod-template-hash:595b5b9587] map[openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 d3eaf24a-d135-44ef-bb31-a6db652e7724 0xc00484ef70 0xc00484ef71}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.74,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:40 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.300: INFO: Pod "webserver-deployment-c7997dcc8-bgp6t" is not available: +&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-bgp6t webserver-deployment-c7997dcc8- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-c7997dcc8-bgp6t e7417755-c897-4041-9f9d-595e08909411 46217 0 2020-10-26 16:08:38 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.54.157/32 cni.projectcalico.org/podIPs:172.30.54.157/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.54.157" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 7639f2f5-70bd-463d-8250-065c8b3bc6dd 0xc00484f0c0 0xc00484f0c1}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.82,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.82,PodIP:,StartTime:2020-10-26 16:08:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.300: INFO: Pod "webserver-deployment-c7997dcc8-czxq8" is not available: +&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-czxq8 webserver-deployment-c7997dcc8- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-c7997dcc8-czxq8 abe8427e-adb3-441e-91a0-57d083dd34b5 46206 0 2020-10-26 16:08:37 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.194.89/32 cni.projectcalico.org/podIPs:172.30.194.89/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.194.89" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 7639f2f5-70bd-463d-8250-065c8b3bc6dd 0xc00484f277 0xc00484f278}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.74,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.74,PodIP:,StartTime:2020-10-26 16:08:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.300: INFO: Pod "webserver-deployment-c7997dcc8-s8fpq" is not available: +&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-s8fpq webserver-deployment-c7997dcc8- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-c7997dcc8-s8fpq 6c14c9ac-52df-40c2-805f-862ce4864a59 46246 0 2020-10-26 16:08:40 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 7639f2f5-70bd-463d-8250-065c8b3bc6dd 0xc00484f417 0xc00484f418}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.301: INFO: Pod "webserver-deployment-c7997dcc8-vblz8" is not available: +&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-vblz8 webserver-deployment-c7997dcc8- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-c7997dcc8-vblz8 4698f174-9b57-4d7e-9cb8-8422782e836a 46225 0 2020-10-26 16:08:38 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.54.158/32 cni.projectcalico.org/podIPs:172.30.54.158/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.54.158" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 7639f2f5-70bd-463d-8250-065c8b3bc6dd 0xc00484f577 0xc00484f578}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.82,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.82,PodIP:,StartTime:2020-10-26 16:08:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.303: INFO: Pod "webserver-deployment-c7997dcc8-wzc9d" is not available: +&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-wzc9d webserver-deployment-c7997dcc8- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-c7997dcc8-wzc9d fe78edab-f242-40b5-a0ec-d3387e042350 46212 0 2020-10-26 16:08:38 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.103.243/32 cni.projectcalico.org/podIPs:172.30.103.243/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.103.243" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 7639f2f5-70bd-463d-8250-065c8b3bc6dd 0xc00484f717 0xc00484f718}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.76,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.76,PodIP:,StartTime:2020-10-26 16:08:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Oct 26 16:08:40.303: INFO: Pod "webserver-deployment-c7997dcc8-znhbw" is not available: +&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-znhbw webserver-deployment-c7997dcc8- deployment-4122 /api/v1/namespaces/deployment-4122/pods/webserver-deployment-c7997dcc8-znhbw 9b5448cc-6742-4070-a753-ea8a6a402129 46220 0 2020-10-26 16:08:38 +0000 UTC map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.194.90/32 cni.projectcalico.org/podIPs:172.30.194.90/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.194.90" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 7639f2f5-70bd-463d-8250-065c8b3bc6dd 0xc00484f8d7 0xc00484f8d8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-rdz2b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-rdz2b,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-rdz2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.74,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-spsbq,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:08:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.74,PodIP:,StartTime:2020-10-26 16:08:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +[AfterEach] [sig-apps] Deployment /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:51:04.877: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-3529" for this suite. +Oct 26 16:08:40.303: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-4122" for this suite. -• [SLOW TEST:8.558 seconds] -[sig-storage] Projected configMap -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 - should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] +• [SLOW TEST:6.845 seconds] +[sig-apps] Deployment +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 + deployment should support proportional scaling [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":30,"skipped":599,"failed":0} -S +{"msg":"PASSED [sig-apps] Deployment deployment should support proportional scaling [Conformance]","total":280,"completed":28,"skipped":442,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] ConfigMap - optional updates should be reflected in volume [NodeConformance] [Conformance] +[sig-scheduling] NoExecuteTaintManager Single Pod [Serial] + removing taint cancels eviction [Disruptive] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] ConfigMap +[BeforeEach] [sig-scheduling] NoExecuteTaintManager Single Pod [Serial] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:51:04.925: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename configmap +Oct 26 16:08:40.347: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename taint-single-pod STEP: Waiting for a default service account to be provisioned in namespace -[It] optional updates should be reflected in volume [NodeConformance] [Conformance] +[BeforeEach] [sig-scheduling] NoExecuteTaintManager Single Pod [Serial] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/taints.go:163 +Oct 26 16:08:40.500: INFO: Waiting up to 1m0s for all nodes to be ready +Oct 26 16:09:40.650: INFO: Waiting for terminating namespaces to be deleted... +[It] removing taint cancels eviction [Disruptive] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 15:51:05.139: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node -STEP: Creating configMap with name cm-test-opt-del-c7a13d4b-29d6-4b1e-9a1a-73f717e2d1ea -STEP: Creating configMap with name cm-test-opt-upd-92f2d9ae-8e42-4b0e-b066-2dd0bfc1e374 -STEP: Creating the pod -STEP: Deleting configmap cm-test-opt-del-c7a13d4b-29d6-4b1e-9a1a-73f717e2d1ea -STEP: Updating configmap cm-test-opt-upd-92f2d9ae-8e42-4b0e-b066-2dd0bfc1e374 -STEP: Creating configMap with name cm-test-opt-create-d6c30b94-4161-4ec5-af47-f777f536befe -STEP: waiting to observe update in volume -[AfterEach] [sig-storage] ConfigMap +Oct 26 16:09:40.665: INFO: Starting informer... +STEP: Starting pod... +Oct 26 16:09:40.930: INFO: Pod is running on 10.72.119.74. Tainting Node +STEP: Trying to apply a taint on the Node +STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute +STEP: Waiting short time to make sure Pod is queued for deletion +Oct 26 16:09:40.980: INFO: Pod wasn't evicted. Proceeding +Oct 26 16:09:40.980: INFO: Removing taint from Node +STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute +STEP: Waiting some time to make sure that toleration time passed. +Oct 26 16:10:56.051: INFO: Pod wasn't evicted. Test successful +[AfterEach] [sig-scheduling] NoExecuteTaintManager Single Pod [Serial] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:52:33.645: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-5639" for this suite. +Oct 26 16:10:56.051: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "taint-single-pod-7620" for this suite. -• [SLOW TEST:88.784 seconds] -[sig-storage] ConfigMap -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 - optional updates should be reflected in volume [NodeConformance] [Conformance] +• [SLOW TEST:135.744 seconds] +[sig-scheduling] NoExecuteTaintManager Single Pod [Serial] +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 + removing taint cancels eviction [Disruptive] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":31,"skipped":600,"failed":0} -SSSSSSSSSSSSS +{"msg":"PASSED [sig-scheduling] NoExecuteTaintManager Single Pod [Serial] removing taint cancels eviction [Disruptive] [Conformance]","total":280,"completed":29,"skipped":466,"failed":0} +SSSSS ------------------------------ -[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook - should execute poststart exec hook properly [NodeConformance] [Conformance] +[sig-cli] Kubectl client Kubectl run deployment + should create a deployment from an image [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Container Lifecycle Hook +[BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:52:33.711: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename container-lifecycle-hook +Oct 26 16:10:56.093: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] when create a pod with lifecycle hook - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64 -STEP: create the container to handle the HTTPGet hook request. -[It] should execute poststart exec hook properly [NodeConformance] [Conformance] +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278 +[BeforeEach] Kubectl run deployment + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1733 +[It] should create a deployment from an image [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: create the pod with lifecycle hook -STEP: check poststart hook -STEP: delete the pod with lifecycle hook -Sep 21 15:52:42.152: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -Sep 21 15:52:42.168: INFO: Pod pod-with-poststart-exec-hook still exists -Sep 21 15:52:44.168: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -Sep 21 15:52:44.183: INFO: Pod pod-with-poststart-exec-hook still exists -Sep 21 15:52:46.168: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -Sep 21 15:52:46.184: INFO: Pod pod-with-poststart-exec-hook still exists -Sep 21 15:52:48.168: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -Sep 21 15:52:48.184: INFO: Pod pod-with-poststart-exec-hook still exists -Sep 21 15:52:50.168: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -Sep 21 15:52:50.184: INFO: Pod pod-with-poststart-exec-hook still exists -Sep 21 15:52:52.168: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -Sep 21 15:52:52.187: INFO: Pod pod-with-poststart-exec-hook no longer exists -[AfterEach] [k8s.io] Container Lifecycle Hook +STEP: running the image docker.io/library/httpd:2.4.38-alpine +Oct 26 16:10:56.226: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 run e2e-test-httpd-deployment --image=docker.io/library/httpd:2.4.38-alpine --generator=deployment/apps.v1 --namespace=kubectl-1978' +Oct 26 16:10:56.373: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" +Oct 26 16:10:56.373: INFO: stdout: "deployment.apps/e2e-test-httpd-deployment created\n" +STEP: verifying the deployment e2e-test-httpd-deployment was created +STEP: verifying the pod controlled by deployment e2e-test-httpd-deployment was created +[AfterEach] Kubectl run deployment + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1738 +Oct 26 16:11:00.438: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete deployment e2e-test-httpd-deployment --namespace=kubectl-1978' +Oct 26 16:11:00.608: INFO: stderr: "" +Oct 26 16:11:00.608: INFO: stdout: "deployment.apps \"e2e-test-httpd-deployment\" deleted\n" +[AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:52:52.187: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-lifecycle-hook-3144" for this suite. - -• [SLOW TEST:18.533 seconds] -[k8s.io] Container Lifecycle Hook -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - when create a pod with lifecycle hook - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 - should execute poststart exec hook properly [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance]","total":280,"completed":32,"skipped":613,"failed":0} -SSSSSSS +Oct 26 16:11:00.608: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-1978" for this suite. +•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run deployment should create a deployment from an image [Conformance]","total":280,"completed":30,"skipped":471,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class - should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] +[sig-storage] EmptyDir volumes + should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] [sig-node] Pods Extended +[BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:52:52.245: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename pods +Oct 26 16:11:00.644: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Pods Set QOS Class - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:178 -[It] should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] +[It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: creating the pod -STEP: submitting the pod to kubernetes -STEP: verifying QOS class is set on the pod -[AfterEach] [k8s.io] [sig-node] Pods Extended +STEP: Creating a pod to test emptydir 0644 on node default medium +Oct 26 16:11:00.869: INFO: Waiting up to 5m0s for pod "pod-5c37c676-686e-4bc7-8743-c29b01af2582" in namespace "emptydir-2660" to be "success or failure" +Oct 26 16:11:00.891: INFO: Pod "pod-5c37c676-686e-4bc7-8743-c29b01af2582": Phase="Pending", Reason="", readiness=false. Elapsed: 21.684946ms +Oct 26 16:11:02.907: INFO: Pod "pod-5c37c676-686e-4bc7-8743-c29b01af2582": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.037438164s +STEP: Saw pod success +Oct 26 16:11:02.907: INFO: Pod "pod-5c37c676-686e-4bc7-8743-c29b01af2582" satisfied condition "success or failure" +Oct 26 16:11:02.921: INFO: Trying to get logs from node 10.72.119.74 pod pod-5c37c676-686e-4bc7-8743-c29b01af2582 container test-container: +STEP: delete the pod +Oct 26 16:11:03.042: INFO: Waiting for pod pod-5c37c676-686e-4bc7-8743-c29b01af2582 to disappear +Oct 26 16:11:03.062: INFO: Pod pod-5c37c676-686e-4bc7-8743-c29b01af2582 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:52:52.467: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-2001" for this suite. -•{"msg":"PASSED [k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]","total":280,"completed":33,"skipped":620,"failed":0} -SSSSSSSSSS +Oct 26 16:11:03.062: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-2660" for this suite. +•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":31,"skipped":495,"failed":0} + ------------------------------ -[sig-api-machinery] Secrets - should be consumable via the environment [NodeConformance] [Conformance] +[sig-storage] Projected downwardAPI + should update labels on modification [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] Secrets +[BeforeEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:52:52.514: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename secrets +Oct 26 16:11:03.094: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable via the environment [NodeConformance] [Conformance] +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40 +[It] should update labels on modification [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: creating secret secrets-7887/secret-test-08530bce-381f-4c34-9b81-787ff2f27515 -STEP: Creating a pod to test consume secrets -Sep 21 15:52:52.736: INFO: Waiting up to 5m0s for pod "pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024" in namespace "secrets-7887" to be "success or failure" -Sep 21 15:52:52.751: INFO: Pod "pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024": Phase="Pending", Reason="", readiness=false. Elapsed: 14.711711ms -Sep 21 15:52:54.767: INFO: Pod "pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030641033s -Sep 21 15:52:56.784: INFO: Pod "pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024": Phase="Pending", Reason="", readiness=false. Elapsed: 4.04721089s -Sep 21 15:52:58.798: INFO: Pod "pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024": Phase="Pending", Reason="", readiness=false. Elapsed: 6.062003783s -Sep 21 15:53:00.815: INFO: Pod "pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024": Phase="Pending", Reason="", readiness=false. Elapsed: 8.078102806s -Sep 21 15:53:02.832: INFO: Pod "pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024": Phase="Pending", Reason="", readiness=false. Elapsed: 10.095820875s -Sep 21 15:53:04.850: INFO: Pod "pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024": Phase="Pending", Reason="", readiness=false. Elapsed: 12.11385581s -Sep 21 15:53:06.866: INFO: Pod "pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024": Phase="Succeeded", Reason="", readiness=false. Elapsed: 14.12909651s -STEP: Saw pod success -Sep 21 15:53:06.866: INFO: Pod "pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024" satisfied condition "success or failure" -Sep 21 15:53:06.882: INFO: Trying to get logs from node 10.241.51.150 pod pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024 container env-test: -STEP: delete the pod -Sep 21 15:53:06.963: INFO: Waiting for pod pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024 to disappear -Sep 21 15:53:06.982: INFO: Pod pod-configmaps-6e682e69-0511-494c-9180-b8eca01cb024 no longer exists -[AfterEach] [sig-api-machinery] Secrets +STEP: Creating the pod +Oct 26 16:11:07.931: INFO: Successfully updated pod "labelsupdate62590a0b-32fd-4618-bae7-836026aa8acd" +[AfterEach] [sig-storage] Projected downwardAPI /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:53:06.982: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-7887" for this suite. +Oct 26 16:11:09.998: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-9118" for this suite. -• [SLOW TEST:14.514 seconds] -[sig-api-machinery] Secrets -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets.go:31 - should be consumable via the environment [NodeConformance] [Conformance] +• [SLOW TEST:6.940 seconds] +[sig-storage] Projected downwardAPI +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:34 + should update labels on modification [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-api-machinery] Secrets should be consumable via the environment [NodeConformance] [Conformance]","total":280,"completed":34,"skipped":630,"failed":0} -SSSSS +{"msg":"PASSED [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance]","total":280,"completed":32,"skipped":495,"failed":0} +SSSSSSS ------------------------------ -[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] +[sig-storage] Projected combined + should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-apps] StatefulSet +[BeforeEach] [sig-storage] Projected combined /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:53:07.028: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename statefulset +Oct 26 16:11:10.036: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] StatefulSet - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64 -[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79 -STEP: Creating service test in namespace statefulset-9660 -[It] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] +[It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating stateful set ss in namespace statefulset-9660 -STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-9660 -Sep 21 15:53:08.247: INFO: Found 0 stateful pods, waiting for 1 -Sep 21 15:53:18.274: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true -STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod -Sep 21 15:53:18.295: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Sep 21 15:53:19.328: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Sep 21 15:53:19.328: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Sep 21 15:53:19.328: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Sep 21 15:53:19.343: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true -Sep 21 15:53:29.362: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false -Sep 21 15:53:29.362: INFO: Waiting for statefulset status.replicas updated to 0 -Sep 21 15:53:29.429: INFO: POD NODE PHASE GRACE CONDITIONS -Sep 21 15:53:29.429: INFO: ss-0 10.241.51.147 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:08 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:08 +0000 UTC }] -Sep 21 15:53:29.429: INFO: -Sep 21 15:53:29.429: INFO: StatefulSet ss has not reached scale 3, at 1 -Sep 21 15:53:30.450: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.980703299s -Sep 21 15:53:31.472: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.959212891s -Sep 21 15:53:32.489: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.936915447s -Sep 21 15:53:33.507: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.920136661s -Sep 21 15:53:34.523: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.902863739s -Sep 21 15:53:35.543: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.886693325s -Sep 21 15:53:36.562: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.865977294s -Sep 21 15:53:37.582: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.846932341s -Sep 21 15:53:38.596: INFO: Verifying statefulset ss doesn't scale past 3 for another 827.855114ms -STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-9660 -Sep 21 15:53:39.628: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:53:52.760: INFO: rc: 1 -Sep 21 15:53:52.760: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (InternalError): Internal error occurred: admission plugin "security.openshift.io/SCCExecRestrictions" failed to complete validation in 13s - -error: -exit status 1 -Sep 21 15:54:02.760: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:54:11.593: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" -Sep 21 15:54:11.593: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" -Sep 21 15:54:11.593: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - -Sep 21 15:54:11.593: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:54:11.988: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" -Sep 21 15:54:11.988: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" -Sep 21 15:54:11.988: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - -Sep 21 15:54:11.988: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:54:15.542: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" -Sep 21 15:54:15.542: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" -Sep 21 15:54:15.542: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - -Sep 21 15:54:15.559: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true -Sep 21 15:54:15.559: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true -Sep 21 15:54:15.559: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true -STEP: Scale down will not halt with unhealthy stateful pod -Sep 21 15:54:15.579: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Sep 21 15:54:16.105: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Sep 21 15:54:16.105: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Sep 21 15:54:16.105: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Sep 21 15:54:16.105: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Sep 21 15:54:16.484: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Sep 21 15:54:16.484: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Sep 21 15:54:16.484: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Sep 21 15:54:16.484: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Sep 21 15:54:18.925: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Sep 21 15:54:18.925: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Sep 21 15:54:18.925: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Sep 21 15:54:18.925: INFO: Waiting for statefulset status.replicas updated to 0 -Sep 21 15:54:18.966: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1 -Sep 21 15:54:29.007: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false -Sep 21 15:54:29.007: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false -Sep 21 15:54:29.007: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false -Sep 21 15:54:29.072: INFO: POD NODE PHASE GRACE CONDITIONS -Sep 21 15:54:29.072: INFO: ss-0 10.241.51.147 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:08 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:16 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:16 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:08 +0000 UTC }] -Sep 21 15:54:29.072: INFO: ss-1 10.241.51.150 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:29.072: INFO: ss-2 10.241.51.149 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:29.072: INFO: -Sep 21 15:54:29.072: INFO: StatefulSet ss has not reached scale 0, at 3 -Sep 21 15:54:30.088: INFO: POD NODE PHASE GRACE CONDITIONS -Sep 21 15:54:30.088: INFO: ss-0 10.241.51.147 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:08 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:16 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:16 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:08 +0000 UTC }] -Sep 21 15:54:30.088: INFO: ss-1 10.241.51.150 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:30.088: INFO: ss-2 10.241.51.149 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:30.088: INFO: -Sep 21 15:54:30.088: INFO: StatefulSet ss has not reached scale 0, at 3 -Sep 21 15:54:31.105: INFO: POD NODE PHASE GRACE CONDITIONS -Sep 21 15:54:31.105: INFO: ss-0 10.241.51.147 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:08 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:16 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:16 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:08 +0000 UTC }] -Sep 21 15:54:31.105: INFO: ss-1 10.241.51.150 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:31.105: INFO: ss-2 10.241.51.149 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:31.105: INFO: -Sep 21 15:54:31.105: INFO: StatefulSet ss has not reached scale 0, at 3 -Sep 21 15:54:32.121: INFO: POD NODE PHASE GRACE CONDITIONS -Sep 21 15:54:32.121: INFO: ss-0 10.241.51.147 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:08 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:16 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:16 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:08 +0000 UTC }] -Sep 21 15:54:32.121: INFO: ss-1 10.241.51.150 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:32.121: INFO: ss-2 10.241.51.149 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:32.121: INFO: -Sep 21 15:54:32.121: INFO: StatefulSet ss has not reached scale 0, at 3 -Sep 21 15:54:33.137: INFO: POD NODE PHASE GRACE CONDITIONS -Sep 21 15:54:33.137: INFO: ss-0 10.241.51.147 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:08 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:16 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:16 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:08 +0000 UTC }] -Sep 21 15:54:33.137: INFO: ss-1 10.241.51.150 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:33.137: INFO: ss-2 10.241.51.149 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:33.137: INFO: -Sep 21 15:54:33.137: INFO: StatefulSet ss has not reached scale 0, at 3 -Sep 21 15:54:34.153: INFO: POD NODE PHASE GRACE CONDITIONS -Sep 21 15:54:34.153: INFO: ss-1 10.241.51.150 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:34.153: INFO: ss-2 10.241.51.149 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:34.153: INFO: -Sep 21 15:54:34.153: INFO: StatefulSet ss has not reached scale 0, at 2 -Sep 21 15:54:35.170: INFO: POD NODE PHASE GRACE CONDITIONS -Sep 21 15:54:35.170: INFO: ss-1 10.241.51.150 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:35.170: INFO: -Sep 21 15:54:35.170: INFO: StatefulSet ss has not reached scale 0, at 1 -Sep 21 15:54:36.191: INFO: POD NODE PHASE GRACE CONDITIONS -Sep 21 15:54:36.191: INFO: ss-1 10.241.51.150 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:36.191: INFO: -Sep 21 15:54:36.191: INFO: StatefulSet ss has not reached scale 0, at 1 -Sep 21 15:54:37.206: INFO: POD NODE PHASE GRACE CONDITIONS -Sep 21 15:54:37.206: INFO: ss-1 10.241.51.150 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:37.206: INFO: -Sep 21 15:54:37.206: INFO: StatefulSet ss has not reached scale 0, at 1 -Sep 21 15:54:38.220: INFO: POD NODE PHASE GRACE CONDITIONS -Sep 21 15:54:38.220: INFO: ss-1 10.241.51.150 Pending 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:54:17 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-09-21 15:53:29 +0000 UTC }] -Sep 21 15:54:38.220: INFO: -Sep 21 15:54:38.220: INFO: StatefulSet ss has not reached scale 0, at 1 -STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-9660 -Sep 21 15:54:39.247: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:54:40.415: INFO: rc: 1 -Sep 21 15:54:40.415: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:54:50.416: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:54:51.109: INFO: rc: 1 -Sep 21 15:54:51.109: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:55:01.109: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:55:01.487: INFO: rc: 1 -Sep 21 15:55:01.487: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:55:11.487: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:55:11.631: INFO: rc: 1 -Sep 21 15:55:11.631: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:55:21.631: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:55:21.846: INFO: rc: 1 -Sep 21 15:55:21.847: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:55:31.847: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:55:32.265: INFO: rc: 1 -Sep 21 15:55:32.265: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:55:42.265: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:55:42.507: INFO: rc: 1 -Sep 21 15:55:42.507: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:55:52.507: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:55:52.723: INFO: rc: 1 -Sep 21 15:55:52.723: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:56:02.723: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:56:02.847: INFO: rc: 1 -Sep 21 15:56:02.847: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:56:12.848: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:56:13.363: INFO: rc: 1 -Sep 21 15:56:13.363: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:56:23.364: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:56:23.531: INFO: rc: 1 -Sep 21 15:56:23.531: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:56:33.531: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:56:34.357: INFO: rc: 1 -Sep 21 15:56:34.357: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:56:44.358: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:56:44.557: INFO: rc: 1 -Sep 21 15:56:44.557: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:56:54.558: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:56:54.715: INFO: rc: 1 -Sep 21 15:56:54.715: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:57:04.716: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:57:04.854: INFO: rc: 1 -Sep 21 15:57:04.854: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:57:14.854: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:57:19.379: INFO: rc: 1 -Sep 21 15:57:19.380: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:57:29.380: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:57:29.530: INFO: rc: 1 -Sep 21 15:57:29.530: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:57:39.530: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:57:43.641: INFO: rc: 1 -Sep 21 15:57:43.642: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:57:53.642: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:57:53.799: INFO: rc: 1 -Sep 21 15:57:53.799: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:58:03.799: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:58:04.238: INFO: rc: 1 -Sep 21 15:58:04.238: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:58:14.239: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:58:14.374: INFO: rc: 1 -Sep 21 15:58:14.374: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:58:24.374: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:58:24.517: INFO: rc: 1 -Sep 21 15:58:24.517: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:58:34.518: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:58:34.760: INFO: rc: 1 -Sep 21 15:58:34.760: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:58:44.760: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:58:44.907: INFO: rc: 1 -Sep 21 15:58:44.907: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:58:54.907: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:58:56.588: INFO: rc: 1 -Sep 21 15:58:56.588: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:59:06.588: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:59:08.078: INFO: rc: 1 -Sep 21 15:59:08.078: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:59:18.078: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:59:18.584: INFO: rc: 1 -Sep 21 15:59:18.585: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:59:28.585: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:59:28.856: INFO: rc: 1 -Sep 21 15:59:28.856: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:59:38.856: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:59:39.067: INFO: rc: 1 -Sep 21 15:59:39.067: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true: -Command stdout: - -stderr: -Error from server (NotFound): pods "ss-1" not found - -error: -exit status 1 -Sep 21 15:59:49.067: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-9660 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Sep 21 15:59:49.236: INFO: rc: 1 -Sep 21 15:59:49.236: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: -Sep 21 15:59:49.236: INFO: Scaling statefulset ss to 0 -Sep 21 15:59:49.779: INFO: Waiting for statefulset status.replicas updated to 0 -[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 -Sep 21 15:59:49.804: INFO: Deleting all statefulset in ns statefulset-9660 -Sep 21 15:59:49.852: INFO: Scaling statefulset ss to 0 -Sep 21 15:59:49.956: INFO: Waiting for statefulset status.replicas updated to 0 -Sep 21 15:59:49.979: INFO: Deleting statefulset ss -[AfterEach] [sig-apps] StatefulSet +STEP: Creating configMap with name configmap-projected-all-test-volume-0798ded1-dddc-4685-b486-9fb045b71d00 +STEP: Creating secret with name secret-projected-all-test-volume-3139db55-8ea6-4233-a0bf-5c9ccb7cd326 +STEP: Creating a pod to test Check all projections for projected volume plugin +Oct 26 16:11:10.250: INFO: Waiting up to 5m0s for pod "projected-volume-dd5f990b-5bd1-4310-a35a-b0e9ce93bda0" in namespace "projected-6549" to be "success or failure" +Oct 26 16:11:10.264: INFO: Pod "projected-volume-dd5f990b-5bd1-4310-a35a-b0e9ce93bda0": Phase="Pending", Reason="", readiness=false. Elapsed: 14.785758ms +Oct 26 16:11:12.284: INFO: Pod "projected-volume-dd5f990b-5bd1-4310-a35a-b0e9ce93bda0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.034690111s +STEP: Saw pod success +Oct 26 16:11:12.284: INFO: Pod "projected-volume-dd5f990b-5bd1-4310-a35a-b0e9ce93bda0" satisfied condition "success or failure" +Oct 26 16:11:12.300: INFO: Trying to get logs from node 10.72.119.74 pod projected-volume-dd5f990b-5bd1-4310-a35a-b0e9ce93bda0 container projected-all-volume-test: +STEP: delete the pod +Oct 26 16:11:12.394: INFO: Waiting for pod projected-volume-dd5f990b-5bd1-4310-a35a-b0e9ce93bda0 to disappear +Oct 26 16:11:12.409: INFO: Pod projected-volume-dd5f990b-5bd1-4310-a35a-b0e9ce93bda0 no longer exists +[AfterEach] [sig-storage] Projected combined /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:59:50.091: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "statefulset-9660" for this suite. - -• [SLOW TEST:403.375 seconds] -[sig-apps] StatefulSet -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 - [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]","total":280,"completed":35,"skipped":635,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +Oct 26 16:11:12.409: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-6549" for this suite. +•{"msg":"PASSED [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance]","total":280,"completed":33,"skipped":502,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[k8s.io] Docker Containers - should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] +[sig-apps] Daemon set [Serial] + should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Docker Containers +[BeforeEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:59:50.404: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename containers +Oct 26 16:11:12.450: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename daemonsets STEP: Waiting for a default service account to be provisioned in namespace -[It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133 +[It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test override command -Sep 21 15:59:52.827: INFO: Waiting up to 5m0s for pod "client-containers-3d3a2eda-5687-4858-a544-6547018e2c9f" in namespace "containers-5944" to be "success or failure" -Sep 21 15:59:52.859: INFO: Pod "client-containers-3d3a2eda-5687-4858-a544-6547018e2c9f": Phase="Pending", Reason="", readiness=false. Elapsed: 32.191632ms -Sep 21 15:59:54.889: INFO: Pod "client-containers-3d3a2eda-5687-4858-a544-6547018e2c9f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.061632169s -Sep 21 15:59:57.132: INFO: Pod "client-containers-3d3a2eda-5687-4858-a544-6547018e2c9f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.304765189s -STEP: Saw pod success -Sep 21 15:59:57.132: INFO: Pod "client-containers-3d3a2eda-5687-4858-a544-6547018e2c9f" satisfied condition "success or failure" -Sep 21 15:59:57.269: INFO: Trying to get logs from node 10.241.51.147 pod client-containers-3d3a2eda-5687-4858-a544-6547018e2c9f container test-container: -STEP: delete the pod -Sep 21 15:59:58.184: INFO: Waiting for pod client-containers-3d3a2eda-5687-4858-a544-6547018e2c9f to disappear -Sep 21 15:59:58.215: INFO: Pod client-containers-3d3a2eda-5687-4858-a544-6547018e2c9f no longer exists -[AfterEach] [k8s.io] Docker Containers +Oct 26 16:11:12.689: INFO: Creating simple daemon set daemon-set +STEP: Check that daemon pods launch on every node of the cluster. +Oct 26 16:11:12.756: INFO: Number of nodes with available pods: 0 +Oct 26 16:11:12.756: INFO: Node 10.72.119.74 is running more than one daemon pod +Oct 26 16:11:13.818: INFO: Number of nodes with available pods: 0 +Oct 26 16:11:13.818: INFO: Node 10.72.119.74 is running more than one daemon pod +Oct 26 16:11:14.788: INFO: Number of nodes with available pods: 2 +Oct 26 16:11:14.788: INFO: Node 10.72.119.76 is running more than one daemon pod +Oct 26 16:11:15.788: INFO: Number of nodes with available pods: 3 +Oct 26 16:11:15.788: INFO: Number of running nodes: 3, number of available pods: 3 +STEP: Update daemon pods image. +STEP: Check that daemon pods images are updated. +Oct 26 16:11:15.918: INFO: Wrong image for pod: daemon-set-g9x7h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:15.918: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:15.918: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:16.963: INFO: Wrong image for pod: daemon-set-g9x7h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:16.963: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:16.963: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:17.957: INFO: Wrong image for pod: daemon-set-g9x7h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:17.957: INFO: Pod daemon-set-g9x7h is not available +Oct 26 16:11:17.957: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:17.957: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:18.957: INFO: Wrong image for pod: daemon-set-g9x7h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:18.957: INFO: Pod daemon-set-g9x7h is not available +Oct 26 16:11:18.957: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:18.957: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:19.958: INFO: Wrong image for pod: daemon-set-g9x7h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:19.958: INFO: Pod daemon-set-g9x7h is not available +Oct 26 16:11:19.958: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:19.958: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:20.959: INFO: Wrong image for pod: daemon-set-g9x7h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:20.959: INFO: Pod daemon-set-g9x7h is not available +Oct 26 16:11:20.959: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:20.959: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:21.957: INFO: Wrong image for pod: daemon-set-g9x7h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:21.957: INFO: Pod daemon-set-g9x7h is not available +Oct 26 16:11:21.957: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:21.957: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:22.956: INFO: Wrong image for pod: daemon-set-g9x7h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:22.956: INFO: Pod daemon-set-g9x7h is not available +Oct 26 16:11:22.956: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:22.957: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:23.958: INFO: Wrong image for pod: daemon-set-g9x7h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:23.958: INFO: Pod daemon-set-g9x7h is not available +Oct 26 16:11:23.958: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:23.958: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:24.959: INFO: Wrong image for pod: daemon-set-g9x7h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:24.959: INFO: Pod daemon-set-g9x7h is not available +Oct 26 16:11:24.959: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:24.959: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:25.956: INFO: Wrong image for pod: daemon-set-g9x7h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:25.956: INFO: Pod daemon-set-g9x7h is not available +Oct 26 16:11:25.956: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:25.956: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:26.958: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:26.958: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:26.958: INFO: Pod daemon-set-vbr4c is not available +Oct 26 16:11:27.956: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:27.956: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:27.956: INFO: Pod daemon-set-vbr4c is not available +Oct 26 16:11:28.958: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:28.959: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:29.967: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:29.967: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:30.958: INFO: Wrong image for pod: daemon-set-h8b2h. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:30.958: INFO: Pod daemon-set-h8b2h is not available +Oct 26 16:11:30.958: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:31.957: INFO: Pod daemon-set-bn64n is not available +Oct 26 16:11:31.957: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:32.958: INFO: Pod daemon-set-bn64n is not available +Oct 26 16:11:32.958: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:33.956: INFO: Pod daemon-set-bn64n is not available +Oct 26 16:11:33.956: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:34.961: INFO: Pod daemon-set-bn64n is not available +Oct 26 16:11:34.961: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:35.958: INFO: Pod daemon-set-bn64n is not available +Oct 26 16:11:35.958: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:36.958: INFO: Pod daemon-set-bn64n is not available +Oct 26 16:11:36.959: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:37.956: INFO: Pod daemon-set-bn64n is not available +Oct 26 16:11:37.956: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:38.964: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:39.962: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:39.962: INFO: Pod daemon-set-qq94w is not available +Oct 26 16:11:40.967: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:40.968: INFO: Pod daemon-set-qq94w is not available +Oct 26 16:11:41.958: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:41.958: INFO: Pod daemon-set-qq94w is not available +Oct 26 16:11:42.958: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:42.958: INFO: Pod daemon-set-qq94w is not available +Oct 26 16:11:43.957: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:43.957: INFO: Pod daemon-set-qq94w is not available +Oct 26 16:11:44.956: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:44.956: INFO: Pod daemon-set-qq94w is not available +Oct 26 16:11:45.957: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:45.957: INFO: Pod daemon-set-qq94w is not available +Oct 26 16:11:46.963: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:46.963: INFO: Pod daemon-set-qq94w is not available +Oct 26 16:11:47.956: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:47.957: INFO: Pod daemon-set-qq94w is not available +Oct 26 16:11:48.966: INFO: Wrong image for pod: daemon-set-qq94w. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Oct 26 16:11:48.966: INFO: Pod daemon-set-qq94w is not available +Oct 26 16:11:49.958: INFO: Pod daemon-set-hfkgv is not available +STEP: Check that daemon pods are still running on every node of the cluster. +Oct 26 16:11:50.004: INFO: Number of nodes with available pods: 2 +Oct 26 16:11:50.004: INFO: Node 10.72.119.82 is running more than one daemon pod +Oct 26 16:11:51.035: INFO: Number of nodes with available pods: 2 +Oct 26 16:11:51.035: INFO: Node 10.72.119.82 is running more than one daemon pod +Oct 26 16:11:52.038: INFO: Number of nodes with available pods: 2 +Oct 26 16:11:52.038: INFO: Node 10.72.119.82 is running more than one daemon pod +Oct 26 16:11:53.036: INFO: Number of nodes with available pods: 2 +Oct 26 16:11:53.037: INFO: Node 10.72.119.82 is running more than one daemon pod +Oct 26 16:11:54.037: INFO: Number of nodes with available pods: 2 +Oct 26 16:11:54.037: INFO: Node 10.72.119.82 is running more than one daemon pod +Oct 26 16:11:55.036: INFO: Number of nodes with available pods: 2 +Oct 26 16:11:55.036: INFO: Node 10.72.119.82 is running more than one daemon pod +Oct 26 16:11:56.056: INFO: Number of nodes with available pods: 3 +Oct 26 16:11:56.057: INFO: Number of running nodes: 3, number of available pods: 3 +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99 +STEP: Deleting DaemonSet "daemon-set" +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-7966, will wait for the garbage collector to delete the pods +Oct 26 16:11:56.236: INFO: Deleting DaemonSet.extensions daemon-set took: 36.634809ms +Oct 26 16:11:56.637: INFO: Terminating DaemonSet.extensions daemon-set pods took: 400.271813ms +Oct 26 16:12:09.554: INFO: Number of nodes with available pods: 0 +Oct 26 16:12:09.554: INFO: Number of running nodes: 0, number of available pods: 0 +Oct 26 16:12:09.569: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-7966/daemonsets","resourceVersion":"49403"},"items":null} + +Oct 26 16:12:09.585: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-7966/pods","resourceVersion":"49403"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 15:59:58.216: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "containers-5944" for this suite. +Oct 26 16:12:09.638: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-7966" for this suite. -• [SLOW TEST:7.968 seconds] -[k8s.io] Docker Containers -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance] +• [SLOW TEST:57.226 seconds] +[sig-apps] Daemon set [Serial] +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 + should update pod when spec was updated and update strategy is RollingUpdate [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]","total":280,"completed":36,"skipped":674,"failed":0} -SSSSSSSS +{"msg":"PASSED [sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]","total":280,"completed":34,"skipped":534,"failed":0} +SS ------------------------------ -[sig-storage] ConfigMap - should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] ConfigMap +[sig-api-machinery] Watchers + should observe add, update, and delete watch notifications on configmaps [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-api-machinery] Watchers /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 15:59:58.372: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename configmap +Oct 26 16:12:09.676: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename watch STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] +[It] should observe add, update, and delete watch notifications on configmaps [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating configMap with name configmap-test-volume-a66cfac3-91da-4780-848d-570a4e693ba3 -STEP: Creating a pod to test consume configMaps -Sep 21 15:59:59.127: INFO: Waiting up to 5m0s for pod "pod-configmaps-8f31a2d9-24d3-48f0-a84a-8c20c246e40a" in namespace "configmap-4973" to be "success or failure" -Sep 21 15:59:59.438: INFO: Pod "pod-configmaps-8f31a2d9-24d3-48f0-a84a-8c20c246e40a": Phase="Pending", Reason="", readiness=false. Elapsed: 310.69188ms -Sep 21 16:00:01.478: INFO: Pod "pod-configmaps-8f31a2d9-24d3-48f0-a84a-8c20c246e40a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.350158435s -Sep 21 16:00:03.937: INFO: Pod "pod-configmaps-8f31a2d9-24d3-48f0-a84a-8c20c246e40a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.809514208s -Sep 21 16:00:05.968: INFO: Pod "pod-configmaps-8f31a2d9-24d3-48f0-a84a-8c20c246e40a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.840729428s -STEP: Saw pod success -Sep 21 16:00:05.969: INFO: Pod "pod-configmaps-8f31a2d9-24d3-48f0-a84a-8c20c246e40a" satisfied condition "success or failure" -Sep 21 16:00:05.998: INFO: Trying to get logs from node 10.241.51.150 pod pod-configmaps-8f31a2d9-24d3-48f0-a84a-8c20c246e40a container configmap-volume-test: -STEP: delete the pod -Sep 21 16:00:06.258: INFO: Waiting for pod pod-configmaps-8f31a2d9-24d3-48f0-a84a-8c20c246e40a to disappear -Sep 21 16:00:06.285: INFO: Pod pod-configmaps-8f31a2d9-24d3-48f0-a84a-8c20c246e40a no longer exists -[AfterEach] [sig-storage] ConfigMap +STEP: creating a watch on configmaps with label A +STEP: creating a watch on configmaps with label B +STEP: creating a watch on configmaps with label A or B +STEP: creating a configmap with label A and ensuring the correct watchers observe the notification +Oct 26 16:12:09.850: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-1200 /api/v1/namespaces/watch-1200/configmaps/e2e-watch-test-configmap-a de95273b-9a99-41de-bc60-991617f477fd 49433 0 2020-10-26 16:12:09 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},} +Oct 26 16:12:09.850: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-1200 /api/v1/namespaces/watch-1200/configmaps/e2e-watch-test-configmap-a de95273b-9a99-41de-bc60-991617f477fd 49433 0 2020-10-26 16:12:09 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},} +STEP: modifying configmap A and ensuring the correct watchers observe the notification +Oct 26 16:12:19.876: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-1200 /api/v1/namespaces/watch-1200/configmaps/e2e-watch-test-configmap-a de95273b-9a99-41de-bc60-991617f477fd 49555 0 2020-10-26 16:12:09 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} +Oct 26 16:12:19.876: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-1200 /api/v1/namespaces/watch-1200/configmaps/e2e-watch-test-configmap-a de95273b-9a99-41de-bc60-991617f477fd 49555 0 2020-10-26 16:12:09 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} +STEP: modifying configmap A again and ensuring the correct watchers observe the notification +Oct 26 16:12:29.907: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-1200 /api/v1/namespaces/watch-1200/configmaps/e2e-watch-test-configmap-a de95273b-9a99-41de-bc60-991617f477fd 49598 0 2020-10-26 16:12:09 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +Oct 26 16:12:29.907: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-1200 /api/v1/namespaces/watch-1200/configmaps/e2e-watch-test-configmap-a de95273b-9a99-41de-bc60-991617f477fd 49598 0 2020-10-26 16:12:09 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +STEP: deleting configmap A and ensuring the correct watchers observe the notification +Oct 26 16:12:39.940: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-1200 /api/v1/namespaces/watch-1200/configmaps/e2e-watch-test-configmap-a de95273b-9a99-41de-bc60-991617f477fd 49641 0 2020-10-26 16:12:09 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +Oct 26 16:12:39.940: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-1200 /api/v1/namespaces/watch-1200/configmaps/e2e-watch-test-configmap-a de95273b-9a99-41de-bc60-991617f477fd 49641 0 2020-10-26 16:12:09 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +STEP: creating a configmap with label B and ensuring the correct watchers observe the notification +Oct 26 16:12:49.963: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-1200 /api/v1/namespaces/watch-1200/configmaps/e2e-watch-test-configmap-b a1429487-6c04-4f59-ae96-2a115215f658 49686 0 2020-10-26 16:12:49 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},} +Oct 26 16:12:49.963: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-1200 /api/v1/namespaces/watch-1200/configmaps/e2e-watch-test-configmap-b a1429487-6c04-4f59-ae96-2a115215f658 49686 0 2020-10-26 16:12:49 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},} +STEP: deleting configmap B and ensuring the correct watchers observe the notification +Oct 26 16:12:59.989: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-1200 /api/v1/namespaces/watch-1200/configmaps/e2e-watch-test-configmap-b a1429487-6c04-4f59-ae96-2a115215f658 49725 0 2020-10-26 16:12:49 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},} +Oct 26 16:12:59.989: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-1200 /api/v1/namespaces/watch-1200/configmaps/e2e-watch-test-configmap-b a1429487-6c04-4f59-ae96-2a115215f658 49725 0 2020-10-26 16:12:49 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},} +[AfterEach] [sig-api-machinery] Watchers /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:00:06.285: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-4973" for this suite. +Oct 26 16:13:09.990: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "watch-1200" for this suite. -• [SLOW TEST:7.987 seconds] -[sig-storage] ConfigMap -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33 - should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance] +• [SLOW TEST:60.351 seconds] +[sig-api-machinery] Watchers +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 + should observe add, update, and delete watch notifications on configmaps [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":37,"skipped":682,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance]","total":280,"completed":35,"skipped":536,"failed":0} +SSS ------------------------------ -[k8s.io] Security Context When creating a pod with readOnlyRootFilesystem - should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] +[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook + should execute prestop exec hook properly [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Security Context +[BeforeEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:00:06.364: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename security-context-test +Oct 26 16:13:10.028: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename container-lifecycle-hook STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Security Context - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39 -[It] should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] +[BeforeEach] when create a pod with lifecycle hook + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64 +STEP: create the container to handle the HTTPGet hook request. +[It] should execute prestop exec hook properly [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 16:00:07.480: INFO: Waiting up to 5m0s for pod "busybox-readonly-false-8b654eb3-6ff2-46f7-a641-622148ca38ae" in namespace "security-context-test-430" to be "success or failure" -Sep 21 16:00:07.563: INFO: Pod "busybox-readonly-false-8b654eb3-6ff2-46f7-a641-622148ca38ae": Phase="Pending", Reason="", readiness=false. Elapsed: 82.180775ms -Sep 21 16:00:09.600: INFO: Pod "busybox-readonly-false-8b654eb3-6ff2-46f7-a641-622148ca38ae": Phase="Pending", Reason="", readiness=false. Elapsed: 2.119072825s -Sep 21 16:00:11.632: INFO: Pod "busybox-readonly-false-8b654eb3-6ff2-46f7-a641-622148ca38ae": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.151975113s -Sep 21 16:00:11.633: INFO: Pod "busybox-readonly-false-8b654eb3-6ff2-46f7-a641-622148ca38ae" satisfied condition "success or failure" -[AfterEach] [k8s.io] Security Context +STEP: create the pod with lifecycle hook +STEP: delete the pod with lifecycle hook +Oct 26 16:13:14.381: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Oct 26 16:13:14.402: INFO: Pod pod-with-prestop-exec-hook still exists +Oct 26 16:13:16.402: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Oct 26 16:13:16.418: INFO: Pod pod-with-prestop-exec-hook still exists +Oct 26 16:13:18.402: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Oct 26 16:13:18.419: INFO: Pod pod-with-prestop-exec-hook still exists +Oct 26 16:13:20.402: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Oct 26 16:13:20.419: INFO: Pod pod-with-prestop-exec-hook still exists +Oct 26 16:13:22.402: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Oct 26 16:13:22.417: INFO: Pod pod-with-prestop-exec-hook still exists +Oct 26 16:13:24.402: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Oct 26 16:13:24.418: INFO: Pod pod-with-prestop-exec-hook still exists +Oct 26 16:13:26.403: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Oct 26 16:13:26.418: INFO: Pod pod-with-prestop-exec-hook still exists +Oct 26 16:13:28.402: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Oct 26 16:13:28.421: INFO: Pod pod-with-prestop-exec-hook no longer exists +STEP: check prestop hook +[AfterEach] [k8s.io] Container Lifecycle Hook /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:00:11.633: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "security-context-test-430" for this suite. +Oct 26 16:13:28.484: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-lifecycle-hook-5125" for this suite. -• [SLOW TEST:5.368 seconds] -[k8s.io] Security Context +• [SLOW TEST:18.492 seconds] +[k8s.io] Container Lifecycle Hook /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - When creating a pod with readOnlyRootFilesystem - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:164 - should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] + when create a pod with lifecycle hook + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 + should execute prestop exec hook properly [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [k8s.io] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]","total":280,"completed":38,"skipped":735,"failed":0} +{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance]","total":280,"completed":36,"skipped":539,"failed":0} SSSSS ------------------------------ +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + should be able to deny custom resource creation, update and deletion [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:13:28.521: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename webhook +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 +STEP: Setting up server cert +STEP: Create role binding to let webhook read extension-apiserver-authentication +STEP: Deploying the webhook pod +STEP: Wait for the deployment to be ready +Oct 26 16:13:29.527: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +STEP: Deploying the webhook service +STEP: Verifying the service has paired with the endpoint +Oct 26 16:13:32.607: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should be able to deny custom resource creation, update and deletion [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +Oct 26 16:13:32.631: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Registering the custom resource webhook via the AdmissionRegistration API +STEP: Creating a custom resource that should be denied by the webhook +STEP: Creating a custom resource whose deletion would be denied by the webhook +STEP: Updating the custom resource with disallowed data should be denied +STEP: Deleting the custom resource should be denied +STEP: Remove the offending key and value from the custom resource data +STEP: Deleting the updated custom resource should be successful +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:13:34.132: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-8577" for this suite. +STEP: Destroying namespace "webhook-8577-markers" for this suite. +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 + +• [SLOW TEST:5.871 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 + should be able to deny custom resource creation, update and deletion [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +------------------------------ +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance]","total":280,"completed":37,"skipped":544,"failed":0} +S +------------------------------ +[sig-storage] Downward API volume + should update labels on modification [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:13:34.415: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename downward-api +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40 +[It] should update labels on modification [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating the pod +Oct 26 16:13:37.286: INFO: Successfully updated pod "labelsupdated0e32da7-0b62-4dae-a24d-8f8411109306" +[AfterEach] [sig-storage] Downward API volume + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:13:39.356: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-8843" for this suite. +•{"msg":"PASSED [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance]","total":280,"completed":38,"skipped":545,"failed":0} +SSSSSSSSSSSSS +------------------------------ [sig-api-machinery] ResourceQuota - should be able to update and delete ResourceQuota. [Conformance] + should create a ResourceQuota and capture the life of a configMap. [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:00:11.733: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:13:39.406: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace -[It] should be able to update and delete ResourceQuota. [Conformance] +[It] should create a ResourceQuota and capture the life of a configMap. [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota -STEP: Getting a ResourceQuota -STEP: Updating a ResourceQuota -STEP: Verifying a ResourceQuota was modified -STEP: Deleting a ResourceQuota -STEP: Verifying the deleted ResourceQuota +STEP: Ensuring resource quota status is calculated +STEP: Creating a ConfigMap +STEP: Ensuring resource quota status captures configMap creation +STEP: Deleting a ConfigMap +STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:00:13.630: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-6296" for this suite. -•{"msg":"PASSED [sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance]","total":280,"completed":39,"skipped":740,"failed":0} -SSSSSSSSSSSSSSSSSSSSSS +Oct 26 16:13:55.716: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-1168" for this suite. + +• [SLOW TEST:16.355 seconds] +[sig-api-machinery] ResourceQuota +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 + should create a ResourceQuota and capture the life of a configMap. [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -[sig-storage] Downward API volume - should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] +{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance]","total":280,"completed":39,"skipped":558,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-node] Downward API + should provide pod UID as env vars [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] Downward API volume +[BeforeEach] [sig-node] Downward API /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:00:14.242: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:13:55.763: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-storage] Downward API volume - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40 -[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] +[It] should provide pod UID as env vars [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test downward API volume plugin -Sep 21 16:00:15.966: INFO: Waiting up to 5m0s for pod "downwardapi-volume-11d9ddb5-8bba-495d-bd91-d485c0a26e5f" in namespace "downward-api-6321" to be "success or failure" -Sep 21 16:00:15.998: INFO: Pod "downwardapi-volume-11d9ddb5-8bba-495d-bd91-d485c0a26e5f": Phase="Pending", Reason="", readiness=false. Elapsed: 31.797362ms -Sep 21 16:00:18.026: INFO: Pod "downwardapi-volume-11d9ddb5-8bba-495d-bd91-d485c0a26e5f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.059493457s -Sep 21 16:00:20.056: INFO: Pod "downwardapi-volume-11d9ddb5-8bba-495d-bd91-d485c0a26e5f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.09018717s +STEP: Creating a pod to test downward api env vars +Oct 26 16:13:55.959: INFO: Waiting up to 5m0s for pod "downward-api-20733acb-aec2-4eba-9f1c-daa3ef16e6da" in namespace "downward-api-7073" to be "success or failure" +Oct 26 16:13:55.973: INFO: Pod "downward-api-20733acb-aec2-4eba-9f1c-daa3ef16e6da": Phase="Pending", Reason="", readiness=false. Elapsed: 13.810091ms +Oct 26 16:13:57.988: INFO: Pod "downward-api-20733acb-aec2-4eba-9f1c-daa3ef16e6da": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.028748669s STEP: Saw pod success -Sep 21 16:00:20.057: INFO: Pod "downwardapi-volume-11d9ddb5-8bba-495d-bd91-d485c0a26e5f" satisfied condition "success or failure" -Sep 21 16:00:20.097: INFO: Trying to get logs from node 10.241.51.150 pod downwardapi-volume-11d9ddb5-8bba-495d-bd91-d485c0a26e5f container client-container: +Oct 26 16:13:57.989: INFO: Pod "downward-api-20733acb-aec2-4eba-9f1c-daa3ef16e6da" satisfied condition "success or failure" +Oct 26 16:13:58.003: INFO: Trying to get logs from node 10.72.119.74 pod downward-api-20733acb-aec2-4eba-9f1c-daa3ef16e6da container dapi-container: STEP: delete the pod -Sep 21 16:00:20.730: INFO: Waiting for pod downwardapi-volume-11d9ddb5-8bba-495d-bd91-d485c0a26e5f to disappear -Sep 21 16:00:20.965: INFO: Pod downwardapi-volume-11d9ddb5-8bba-495d-bd91-d485c0a26e5f no longer exists -[AfterEach] [sig-storage] Downward API volume +Oct 26 16:13:58.094: INFO: Waiting for pod downward-api-20733acb-aec2-4eba-9f1c-daa3ef16e6da to disappear +Oct 26 16:13:58.108: INFO: Pod downward-api-20733acb-aec2-4eba-9f1c-daa3ef16e6da no longer exists +[AfterEach] [sig-node] Downward API + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:13:58.108: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-7073" for this suite. +•{"msg":"PASSED [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance]","total":280,"completed":40,"skipped":582,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + should perform rolling updates and roll backs of template modifications [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:13:58.151: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename statefulset +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64 +[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79 +STEP: Creating service test in namespace statefulset-7519 +[It] should perform rolling updates and roll backs of template modifications [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating a new StatefulSet +Oct 26 16:13:58.351: INFO: Found 0 stateful pods, waiting for 3 +Oct 26 16:14:08.368: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true +Oct 26 16:14:08.368: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true +Oct 26 16:14:08.368: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true +Oct 26 16:14:08.418: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-7519 ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Oct 26 16:14:08.993: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Oct 26 16:14:08.993: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Oct 26 16:14:08.993: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +STEP: Updating StatefulSet template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine +Oct 26 16:14:19.106: INFO: Updating stateful set ss2 +STEP: Creating a new revision +STEP: Updating Pods in reverse ordinal order +Oct 26 16:14:29.194: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-7519 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Oct 26 16:14:29.639: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" +Oct 26 16:14:29.639: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Oct 26 16:14:29.639: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Oct 26 16:14:39.752: INFO: Waiting for StatefulSet statefulset-7519/ss2 to complete update +Oct 26 16:14:39.752: INFO: Waiting for Pod statefulset-7519/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 +Oct 26 16:14:39.752: INFO: Waiting for Pod statefulset-7519/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 +Oct 26 16:14:49.788: INFO: Waiting for StatefulSet statefulset-7519/ss2 to complete update +Oct 26 16:14:49.788: INFO: Waiting for Pod statefulset-7519/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 +Oct 26 16:14:49.788: INFO: Waiting for Pod statefulset-7519/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 +Oct 26 16:14:59.784: INFO: Waiting for StatefulSet statefulset-7519/ss2 to complete update +Oct 26 16:14:59.784: INFO: Waiting for Pod statefulset-7519/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 +Oct 26 16:15:09.792: INFO: Waiting for StatefulSet statefulset-7519/ss2 to complete update +Oct 26 16:15:09.792: INFO: Waiting for Pod statefulset-7519/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94 +Oct 26 16:15:19.788: INFO: Waiting for StatefulSet statefulset-7519/ss2 to complete update +STEP: Rolling back to a previous revision +Oct 26 16:15:29.792: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-7519 ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Oct 26 16:15:30.168: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Oct 26 16:15:30.168: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Oct 26 16:15:30.168: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Oct 26 16:15:40.282: INFO: Updating stateful set ss2 +STEP: Rolling back update in reverse ordinal order +Oct 26 16:15:50.364: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-7519 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Oct 26 16:15:50.740: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" +Oct 26 16:15:50.740: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Oct 26 16:15:50.740: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Oct 26 16:16:00.840: INFO: Waiting for StatefulSet statefulset-7519/ss2 to complete update +Oct 26 16:16:00.840: INFO: Waiting for Pod statefulset-7519/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57 +Oct 26 16:16:00.840: INFO: Waiting for Pod statefulset-7519/ss2-1 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57 +Oct 26 16:16:10.909: INFO: Waiting for StatefulSet statefulset-7519/ss2 to complete update +Oct 26 16:16:10.909: INFO: Waiting for Pod statefulset-7519/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57 +[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 +Oct 26 16:16:20.874: INFO: Deleting all statefulset in ns statefulset-7519 +Oct 26 16:16:20.890: INFO: Scaling statefulset ss2 to 0 +Oct 26 16:16:50.958: INFO: Waiting for statefulset status.replicas updated to 0 +Oct 26 16:16:50.975: INFO: Deleting statefulset ss2 +[AfterEach] [sig-apps] StatefulSet /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:00:20.966: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-6321" for this suite. +Oct 26 16:16:51.048: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-7519" for this suite. -• [SLOW TEST:7.066 seconds] -[sig-storage] Downward API volume -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:35 - should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] +• [SLOW TEST:172.931 seconds] +[sig-apps] StatefulSet +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 + [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 + should perform rolling updates and roll backs of template modifications [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +------------------------------ +{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance]","total":280,"completed":41,"skipped":630,"failed":0} +SSSSSSSS +------------------------------ +[sig-apps] Deployment + deployment should delete old replica sets [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-apps] Deployment + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:16:51.083: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename deployment +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-apps] Deployment + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69 +[It] deployment should delete old replica sets [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +Oct 26 16:16:51.323: INFO: Pod name cleanup-pod: Found 0 pods out of 1 +Oct 26 16:16:56.340: INFO: Pod name cleanup-pod: Found 1 pods out of 1 +STEP: ensuring each pod is running +Oct 26 16:16:56.340: INFO: Creating deployment test-cleanup-deployment +STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up +[AfterEach] [sig-apps] Deployment + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63 +Oct 26 16:16:58.477: INFO: Deployment "test-cleanup-deployment": +&Deployment{ObjectMeta:{test-cleanup-deployment deployment-3858 /apis/apps/v1/namespaces/deployment-3858/deployments/test-cleanup-deployment cb932463-150c-485e-bace-e6649ff56946 51756 1 2020-10-26 16:16:56 +0000 UTC map[name:cleanup-pod] map[deployment.kubernetes.io/revision:1] [] [] []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:cleanup-pod] map[] [] [] []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0032776d8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} [] nil default-scheduler [] [] nil [] map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-10-26 16:16:56 +0000 UTC,LastTransitionTime:2020-10-26 16:16:56 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-cleanup-deployment-55ffc6b7b6" has successfully progressed.,LastUpdateTime:2020-10-26 16:16:58 +0000 UTC,LastTransitionTime:2020-10-26 16:16:56 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} + +Oct 26 16:16:58.496: INFO: New ReplicaSet "test-cleanup-deployment-55ffc6b7b6" of Deployment "test-cleanup-deployment": +&ReplicaSet{ObjectMeta:{test-cleanup-deployment-55ffc6b7b6 deployment-3858 /apis/apps/v1/namespaces/deployment-3858/replicasets/test-cleanup-deployment-55ffc6b7b6 868ec5f7-78da-4749-8eba-2e0748a3e6f3 51742 1 2020-10-26 16:16:56 +0000 UTC map[name:cleanup-pod pod-template-hash:55ffc6b7b6] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-cleanup-deployment cb932463-150c-485e-bace-e6649ff56946 0xc0047149a7 0xc0047149a8}] [] []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod-template-hash: 55ffc6b7b6,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:cleanup-pod pod-template-hash:55ffc6b7b6] map[] [] [] []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc004714a18 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} +Oct 26 16:16:58.515: INFO: Pod "test-cleanup-deployment-55ffc6b7b6-mpbl6" is available: +&Pod{ObjectMeta:{test-cleanup-deployment-55ffc6b7b6-mpbl6 test-cleanup-deployment-55ffc6b7b6- deployment-3858 /api/v1/namespaces/deployment-3858/pods/test-cleanup-deployment-55ffc6b7b6-mpbl6 deb2c69e-a914-4be3-9923-a8be1834efd4 51741 0 2020-10-26 16:16:56 +0000 UTC map[name:cleanup-pod pod-template-hash:55ffc6b7b6] map[cni.projectcalico.org/podIP:172.30.194.126/32 cni.projectcalico.org/podIPs:172.30.194.126/32 k8s.v1.cni.cncf.io/networks-status:[{ + "name": "k8s-pod-network", + "ips": [ + "172.30.194.126" + ], + "dns": {} +}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet test-cleanup-deployment-55ffc6b7b6 868ec5f7-78da-4749-8eba-2e0748a3e6f3 0xc004714df7 0xc004714df8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-2n5rb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-2n5rb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-2n5rb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.74,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-7qvp4,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:16:56 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:16:58 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:16:58 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:16:56 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.74,PodIP:172.30.194.126,StartTime:2020-10-26 16:16:56 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:16:57 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:gcr.io/kubernetes-e2e-test-images/agnhost@sha256:02b9e10ea1f7439a03d9bc3ee16d984172ca7eac1818792ad4d721c5c8f72ff4,ContainerID:cri-o://9d50a38e076288fcd8063dcf6c81b302b6ed50cab69259dde2ab61fa69f1e41b,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.194.126,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +[AfterEach] [sig-apps] Deployment + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:16:58.515: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-3858" for this suite. + +• [SLOW TEST:7.467 seconds] +[sig-apps] Deployment +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 + deployment should delete old replica sets [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":40,"skipped":762,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [sig-apps] Deployment deployment should delete old replica sets [Conformance]","total":280,"completed":42,"skipped":638,"failed":0} +SSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-cli] Kubectl client Guestbook application - should create and stop a working application [Conformance] +[sig-cli] Kubectl client Proxy server + should support proxy with --port 0 [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:00:21.308: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:16:58.551: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278 -[It] should create and stop a working application [Conformance] +[It] should support proxy with --port 0 [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: creating all guestbook components -Sep 21 16:00:21.678: INFO: apiVersion: v1 -kind: Service -metadata: - name: agnhost-slave - labels: - app: agnhost - role: slave - tier: backend -spec: - ports: - - port: 6379 - selector: - app: agnhost - role: slave - tier: backend +STEP: starting the proxy server +Oct 26 16:16:58.689: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-871182394 proxy -p 0 --disable-filter' +STEP: curling proxy /api/ output +[AfterEach] [sig-cli] Kubectl client + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:16:58.816: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-9184" for this suite. +•{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support proxy with --port 0 [Conformance]","total":280,"completed":43,"skipped":659,"failed":0} +SSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-apps] ReplicationController + should release no longer matching pods [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-apps] ReplicationController + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:16:58.851: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename replication-controller +STEP: Waiting for a default service account to be provisioned in namespace +[It] should release no longer matching pods [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Given a ReplicationController is created +STEP: When the matched label of one of its pods change +Oct 26 16:16:59.058: INFO: Pod name pod-release: Found 0 pods out of 1 +Oct 26 16:17:04.074: INFO: Pod name pod-release: Found 1 pods out of 1 +STEP: Then the pod is released +[AfterEach] [sig-apps] ReplicationController + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:17:05.155: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replication-controller-2330" for this suite. -Sep 21 16:00:21.678: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-2534' -Sep 21 16:00:23.715: INFO: stderr: "" -Sep 21 16:00:23.715: INFO: stdout: "service/agnhost-slave created\n" -Sep 21 16:00:23.715: INFO: apiVersion: v1 -kind: Service -metadata: - name: agnhost-master - labels: - app: agnhost - role: master - tier: backend -spec: - ports: - - port: 6379 - targetPort: 6379 - selector: - app: agnhost - role: master - tier: backend - -Sep 21 16:00:23.715: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-2534' -Sep 21 16:00:24.336: INFO: stderr: "" -Sep 21 16:00:24.336: INFO: stdout: "service/agnhost-master created\n" -Sep 21 16:00:24.336: INFO: apiVersion: v1 -kind: Service -metadata: - name: frontend - labels: - app: guestbook - tier: frontend -spec: - # if your cluster supports it, uncomment the following to automatically create - # an external load-balanced IP for the frontend service. - # type: LoadBalancer - ports: - - port: 80 - selector: - app: guestbook - tier: frontend - -Sep 21 16:00:24.336: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-2534' -Sep 21 16:00:24.731: INFO: stderr: "" -Sep 21 16:00:24.731: INFO: stdout: "service/frontend created\n" -Sep 21 16:00:24.731: INFO: apiVersion: apps/v1 -kind: Deployment -metadata: - name: frontend -spec: - replicas: 3 - selector: - matchLabels: - app: guestbook - tier: frontend - template: - metadata: - labels: - app: guestbook - tier: frontend - spec: - containers: - - name: guestbook-frontend - image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8 - args: [ "guestbook", "--backend-port", "6379" ] - resources: - requests: - cpu: 100m - memory: 100Mi - ports: - - containerPort: 80 - -Sep 21 16:00:24.731: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-2534' -Sep 21 16:00:25.387: INFO: stderr: "" -Sep 21 16:00:25.387: INFO: stdout: "deployment.apps/frontend created\n" -Sep 21 16:00:25.387: INFO: apiVersion: apps/v1 -kind: Deployment -metadata: - name: agnhost-master -spec: - replicas: 1 - selector: - matchLabels: - app: agnhost - role: master - tier: backend - template: - metadata: - labels: - app: agnhost - role: master - tier: backend - spec: - containers: - - name: master - image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8 - args: [ "guestbook", "--http-port", "6379" ] - resources: - requests: - cpu: 100m - memory: 100Mi - ports: - - containerPort: 6379 - -Sep 21 16:00:25.387: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-2534' -Sep 21 16:00:25.933: INFO: stderr: "" -Sep 21 16:00:25.933: INFO: stdout: "deployment.apps/agnhost-master created\n" -Sep 21 16:00:25.933: INFO: apiVersion: apps/v1 -kind: Deployment -metadata: - name: agnhost-slave -spec: - replicas: 2 - selector: - matchLabels: - app: agnhost - role: slave - tier: backend - template: - metadata: - labels: - app: agnhost - role: slave - tier: backend - spec: - containers: - - name: slave - image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8 - args: [ "guestbook", "--slaveof", "agnhost-master", "--http-port", "6379" ] - resources: - requests: - cpu: 100m - memory: 100Mi - ports: - - containerPort: 6379 - -Sep 21 16:00:25.933: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-2534' -Sep 21 16:00:26.244: INFO: stderr: "" -Sep 21 16:00:26.244: INFO: stdout: "deployment.apps/agnhost-slave created\n" -STEP: validating guestbook app -Sep 21 16:00:26.244: INFO: Waiting for all frontend pods to be Running. -Sep 21 16:00:31.294: INFO: Waiting for frontend to serve content. -Sep 21 16:00:31.822: INFO: Trying to add a new entry to the guestbook. -Sep 21 16:00:31.896: INFO: Verifying that added entry can be retrieved. -Sep 21 16:00:32.037: INFO: Failed to get response from guestbook. err: , response: {"data":""} -STEP: using delete to clean up resources -Sep 21 16:00:37.874: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete --grace-period=0 --force -f - --namespace=kubectl-2534' -Sep 21 16:00:40.044: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Sep 21 16:00:40.044: INFO: stdout: "service \"agnhost-slave\" force deleted\n" -STEP: using delete to clean up resources -Sep 21 16:00:40.044: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete --grace-period=0 --force -f - --namespace=kubectl-2534' -Sep 21 16:00:41.060: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Sep 21 16:00:41.060: INFO: stdout: "service \"agnhost-master\" force deleted\n" -STEP: using delete to clean up resources -Sep 21 16:00:41.060: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete --grace-period=0 --force -f - --namespace=kubectl-2534' -Sep 21 16:00:41.815: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Sep 21 16:00:41.815: INFO: stdout: "service \"frontend\" force deleted\n" -STEP: using delete to clean up resources -Sep 21 16:00:41.815: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete --grace-period=0 --force -f - --namespace=kubectl-2534' -Sep 21 16:00:42.041: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Sep 21 16:00:42.042: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" -STEP: using delete to clean up resources -Sep 21 16:00:42.042: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete --grace-period=0 --force -f - --namespace=kubectl-2534' -Sep 21 16:00:42.200: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Sep 21 16:00:42.200: INFO: stdout: "deployment.apps \"agnhost-master\" force deleted\n" -STEP: using delete to clean up resources -Sep 21 16:00:42.200: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete --grace-period=0 --force -f - --namespace=kubectl-2534' -Sep 21 16:00:42.336: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Sep 21 16:00:42.336: INFO: stdout: "deployment.apps \"agnhost-slave\" force deleted\n" +• [SLOW TEST:6.336 seconds] +[sig-apps] ReplicationController +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 + should release no longer matching pods [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +------------------------------ +{"msg":"PASSED [sig-apps] ReplicationController should release no longer matching pods [Conformance]","total":280,"completed":44,"skipped":681,"failed":0} +SSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-cli] Kubectl client Kubectl rolling-update + should support rolling-update to same image [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:17:05.191: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename kubectl +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278 +[BeforeEach] Kubectl rolling-update + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1692 +[It] should support rolling-update to same image [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: running the image docker.io/library/httpd:2.4.38-alpine +Oct 26 16:17:05.342: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 run e2e-test-httpd-rc --image=docker.io/library/httpd:2.4.38-alpine --generator=run/v1 --namespace=kubectl-309' +Oct 26 16:17:05.540: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n" +Oct 26 16:17:05.540: INFO: stdout: "replicationcontroller/e2e-test-httpd-rc created\n" +STEP: verifying the rc e2e-test-httpd-rc was created +Oct 26 16:17:05.562: INFO: Waiting for rc e2e-test-httpd-rc to stabilize, generation 1 observed generation 0 spec.replicas 1 status.replicas 0 +Oct 26 16:17:05.577: INFO: Waiting for rc e2e-test-httpd-rc to stabilize, generation 1 observed generation 1 spec.replicas 1 status.replicas 0 +STEP: rolling-update to same image controller +Oct 26 16:17:05.629: INFO: scanned /root for discovery docs: +Oct 26 16:17:05.630: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 rolling-update e2e-test-httpd-rc --update-period=1s --image=docker.io/library/httpd:2.4.38-alpine --image-pull-policy=IfNotPresent --namespace=kubectl-309' +Oct 26 16:17:21.839: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n" +Oct 26 16:17:21.840: INFO: stdout: "Created e2e-test-httpd-rc-730ea48ddf6890f865bf31a9388c940f\nScaling up e2e-test-httpd-rc-730ea48ddf6890f865bf31a9388c940f from 0 to 1, scaling down e2e-test-httpd-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-httpd-rc-730ea48ddf6890f865bf31a9388c940f up to 1\nScaling e2e-test-httpd-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-httpd-rc\nRenaming e2e-test-httpd-rc-730ea48ddf6890f865bf31a9388c940f to e2e-test-httpd-rc\nreplicationcontroller/e2e-test-httpd-rc rolling updated\n" +Oct 26 16:17:21.840: INFO: stdout: "Created e2e-test-httpd-rc-730ea48ddf6890f865bf31a9388c940f\nScaling up e2e-test-httpd-rc-730ea48ddf6890f865bf31a9388c940f from 0 to 1, scaling down e2e-test-httpd-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-httpd-rc-730ea48ddf6890f865bf31a9388c940f up to 1\nScaling e2e-test-httpd-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-httpd-rc\nRenaming e2e-test-httpd-rc-730ea48ddf6890f865bf31a9388c940f to e2e-test-httpd-rc\nreplicationcontroller/e2e-test-httpd-rc rolling updated\n" +STEP: waiting for all containers in run=e2e-test-httpd-rc pods to come up. +Oct 26 16:17:21.840: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-httpd-rc --namespace=kubectl-309' +Oct 26 16:17:21.978: INFO: stderr: "" +Oct 26 16:17:21.978: INFO: stdout: "e2e-test-httpd-rc-730ea48ddf6890f865bf31a9388c940f-t6s6t " +Oct 26 16:17:21.978: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods e2e-test-httpd-rc-730ea48ddf6890f865bf31a9388c940f-t6s6t -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "e2e-test-httpd-rc") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-309' +Oct 26 16:17:22.103: INFO: stderr: "" +Oct 26 16:17:22.103: INFO: stdout: "true" +Oct 26 16:17:22.103: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods e2e-test-httpd-rc-730ea48ddf6890f865bf31a9388c940f-t6s6t -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "e2e-test-httpd-rc"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-309' +Oct 26 16:17:22.245: INFO: stderr: "" +Oct 26 16:17:22.245: INFO: stdout: "docker.io/library/httpd:2.4.38-alpine" +Oct 26 16:17:22.245: INFO: e2e-test-httpd-rc-730ea48ddf6890f865bf31a9388c940f-t6s6t is verified up and running +[AfterEach] Kubectl rolling-update + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1698 +Oct 26 16:17:22.245: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete rc e2e-test-httpd-rc --namespace=kubectl-309' +Oct 26 16:17:22.407: INFO: stderr: "" +Oct 26 16:17:22.407: INFO: stdout: "replicationcontroller \"e2e-test-httpd-rc\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:00:42.336: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-2534" for this suite. +Oct 26 16:17:22.407: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-309" for this suite. -• [SLOW TEST:21.685 seconds] +• [SLOW TEST:17.249 seconds] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 - Guestbook application - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:386 - should create and stop a working application [Conformance] + Kubectl rolling-update + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1687 + should support rolling-update to same image [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-cli] Kubectl client Guestbook application should create and stop a working application [Conformance]","total":280,"completed":41,"skipped":785,"failed":0} -SSSSSS +{"msg":"PASSED [sig-cli] Kubectl client Kubectl rolling-update should support rolling-update to same image [Conformance]","total":280,"completed":45,"skipped":701,"failed":0} +SSSS ------------------------------ -[sig-storage] Secrets - should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] +[sig-storage] Projected configMap + should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] Secrets +[BeforeEach] [sig-storage] Projected configMap /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:00:42.994: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename secrets +Oct 26 16:17:22.440: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] +[It] should be consumable from pods in volume [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating secret with name secret-test-74f598a0-8652-4352-a7e9-7c1cbc936a49 -STEP: Creating a pod to test consume secrets -Sep 21 16:00:43.766: INFO: Waiting up to 5m0s for pod "pod-secrets-2f7a68d8-c791-4147-a899-6bbae257cd56" in namespace "secrets-8758" to be "success or failure" -Sep 21 16:00:43.796: INFO: Pod "pod-secrets-2f7a68d8-c791-4147-a899-6bbae257cd56": Phase="Pending", Reason="", readiness=false. Elapsed: 30.050597ms -Sep 21 16:00:45.836: INFO: Pod "pod-secrets-2f7a68d8-c791-4147-a899-6bbae257cd56": Phase="Pending", Reason="", readiness=false. Elapsed: 2.070036284s -Sep 21 16:00:47.905: INFO: Pod "pod-secrets-2f7a68d8-c791-4147-a899-6bbae257cd56": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.138741183s +STEP: Creating configMap with name projected-configmap-test-volume-0266885c-6158-4988-8d42-71005c3e88a1 +STEP: Creating a pod to test consume configMaps +Oct 26 16:17:22.673: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-1b3e707f-11a9-4aa2-8c46-b7b75da8ab06" in namespace "projected-441" to be "success or failure" +Oct 26 16:17:22.688: INFO: Pod "pod-projected-configmaps-1b3e707f-11a9-4aa2-8c46-b7b75da8ab06": Phase="Pending", Reason="", readiness=false. Elapsed: 14.148546ms +Oct 26 16:17:24.703: INFO: Pod "pod-projected-configmaps-1b3e707f-11a9-4aa2-8c46-b7b75da8ab06": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02991783s +Oct 26 16:17:26.719: INFO: Pod "pod-projected-configmaps-1b3e707f-11a9-4aa2-8c46-b7b75da8ab06": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.045505478s STEP: Saw pod success -Sep 21 16:00:47.905: INFO: Pod "pod-secrets-2f7a68d8-c791-4147-a899-6bbae257cd56" satisfied condition "success or failure" -Sep 21 16:00:48.143: INFO: Trying to get logs from node 10.241.51.150 pod pod-secrets-2f7a68d8-c791-4147-a899-6bbae257cd56 container secret-volume-test: +Oct 26 16:17:26.719: INFO: Pod "pod-projected-configmaps-1b3e707f-11a9-4aa2-8c46-b7b75da8ab06" satisfied condition "success or failure" +Oct 26 16:17:26.734: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-configmaps-1b3e707f-11a9-4aa2-8c46-b7b75da8ab06 container projected-configmap-volume-test: STEP: delete the pod -Sep 21 16:00:48.615: INFO: Waiting for pod pod-secrets-2f7a68d8-c791-4147-a899-6bbae257cd56 to disappear -Sep 21 16:00:48.641: INFO: Pod pod-secrets-2f7a68d8-c791-4147-a899-6bbae257cd56 no longer exists -[AfterEach] [sig-storage] Secrets +Oct 26 16:17:26.847: INFO: Waiting for pod pod-projected-configmaps-1b3e707f-11a9-4aa2-8c46-b7b75da8ab06 to disappear +Oct 26 16:17:26.862: INFO: Pod pod-projected-configmaps-1b3e707f-11a9-4aa2-8c46-b7b75da8ab06 no longer exists +[AfterEach] [sig-storage] Projected configMap /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:00:48.641: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-8758" for this suite. - -• [SLOW TEST:5.717 seconds] -[sig-storage] Secrets -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34 - should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":280,"completed":42,"skipped":791,"failed":0} -SSSSSSSSSSSSSSSSSSS +Oct 26 16:17:26.862: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-441" for this suite. +•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":46,"skipped":705,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - patching/updating a validating webhook should work [Conformance] +[sig-storage] Downward API volume + should update annotations on modification [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +[BeforeEach] [sig-storage] Downward API volume /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:00:48.710: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename webhook +Oct 26 16:17:26.896: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 -STEP: Setting up server cert -STEP: Create role binding to let webhook read extension-apiserver-authentication -STEP: Deploying the webhook pod -STEP: Wait for the deployment to be ready -Sep 21 16:00:50.081: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -Sep 21 16:00:52.260: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300850, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300850, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300850, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300850, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)} -Sep 21 16:00:54.307: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300850, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300850, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300850, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300850, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)} -STEP: Deploying the webhook service -STEP: Verifying the service has paired with the endpoint -Sep 21 16:00:57.382: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] patching/updating a validating webhook should work [Conformance] +[BeforeEach] [sig-storage] Downward API volume + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40 +[It] should update annotations on modification [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a validating webhook configuration -STEP: Creating a configMap that does not comply to the validation webhook rules -STEP: Updating a validating webhook configuration's rules to not include the create operation -STEP: Creating a configMap that does not comply to the validation webhook rules -STEP: Patching a validating webhook configuration's rules to include the create operation -STEP: Creating a configMap that does not comply to the validation webhook rules -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +STEP: Creating the pod +Oct 26 16:17:31.765: INFO: Successfully updated pod "annotationupdate997e2a0a-3f80-45fc-9739-c51413f6641c" +[AfterEach] [sig-storage] Downward API volume /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:00:58.168: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-9595" for this suite. -STEP: Destroying namespace "webhook-9595-markers" for this suite. -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 +Oct 26 16:17:33.837: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-5158" for this suite. -• [SLOW TEST:10.287 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - patching/updating a validating webhook should work [Conformance] +• [SLOW TEST:6.980 seconds] +[sig-storage] Downward API volume +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:35 + should update annotations on modification [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance]","total":280,"completed":43,"skipped":810,"failed":0} -SSSSSSS +{"msg":"PASSED [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance]","total":280,"completed":47,"skipped":734,"failed":0} +SS ------------------------------ -[sig-storage] Subpath Atomic writer volumes - should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] +[k8s.io] Container Runtime blackbox test on terminated container + should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] Subpath +[BeforeEach] [k8s.io] Container Runtime /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:00:58.998: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename subpath +Oct 26 16:17:33.875: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] Atomic writer volumes - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37 -STEP: Setting up data -[It] should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] +[It] should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating pod pod-subpath-test-configmap-8pjs -STEP: Creating a pod to test atomic-volume-subpath -Sep 21 16:00:59.970: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-8pjs" in namespace "subpath-5263" to be "success or failure" -Sep 21 16:01:00.005: INFO: Pod "pod-subpath-test-configmap-8pjs": Phase="Pending", Reason="", readiness=false. Elapsed: 34.192635ms -Sep 21 16:01:02.057: INFO: Pod "pod-subpath-test-configmap-8pjs": Phase="Running", Reason="", readiness=true. Elapsed: 2.086970319s -Sep 21 16:01:04.086: INFO: Pod "pod-subpath-test-configmap-8pjs": Phase="Running", Reason="", readiness=true. Elapsed: 4.115053855s -Sep 21 16:01:06.111: INFO: Pod "pod-subpath-test-configmap-8pjs": Phase="Running", Reason="", readiness=true. Elapsed: 6.140739968s -Sep 21 16:01:08.137: INFO: Pod "pod-subpath-test-configmap-8pjs": Phase="Running", Reason="", readiness=true. Elapsed: 8.166590243s -Sep 21 16:01:10.172: INFO: Pod "pod-subpath-test-configmap-8pjs": Phase="Running", Reason="", readiness=true. Elapsed: 10.201369402s -Sep 21 16:01:12.200: INFO: Pod "pod-subpath-test-configmap-8pjs": Phase="Running", Reason="", readiness=true. Elapsed: 12.229200105s -Sep 21 16:01:14.239: INFO: Pod "pod-subpath-test-configmap-8pjs": Phase="Running", Reason="", readiness=true. Elapsed: 14.268686378s -Sep 21 16:01:16.267: INFO: Pod "pod-subpath-test-configmap-8pjs": Phase="Running", Reason="", readiness=true. Elapsed: 16.296894639s -Sep 21 16:01:18.291: INFO: Pod "pod-subpath-test-configmap-8pjs": Phase="Running", Reason="", readiness=true. Elapsed: 18.321029366s -Sep 21 16:01:20.317: INFO: Pod "pod-subpath-test-configmap-8pjs": Phase="Running", Reason="", readiness=true. Elapsed: 20.346370957s -Sep 21 16:01:22.348: INFO: Pod "pod-subpath-test-configmap-8pjs": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.377918697s -STEP: Saw pod success -Sep 21 16:01:22.348: INFO: Pod "pod-subpath-test-configmap-8pjs" satisfied condition "success or failure" -Sep 21 16:01:22.382: INFO: Trying to get logs from node 10.241.51.147 pod pod-subpath-test-configmap-8pjs container test-container-subpath-configmap-8pjs: -STEP: delete the pod -Sep 21 16:01:23.118: INFO: Waiting for pod pod-subpath-test-configmap-8pjs to disappear -Sep 21 16:01:23.358: INFO: Pod pod-subpath-test-configmap-8pjs no longer exists -STEP: Deleting pod pod-subpath-test-configmap-8pjs -Sep 21 16:01:23.358: INFO: Deleting pod "pod-subpath-test-configmap-8pjs" in namespace "subpath-5263" -[AfterEach] [sig-storage] Subpath +STEP: create the container +STEP: wait for the container to reach Failed +STEP: get the container status +STEP: the container should be terminated +STEP: the termination message should be set +Oct 26 16:17:37.171: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- +STEP: delete the container +[AfterEach] [k8s.io] Container Runtime /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:01:23.639: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "subpath-5263" for this suite. - -• [SLOW TEST:24.706 seconds] -[sig-storage] Subpath -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23 - Atomic writer volumes - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33 - should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]","total":280,"completed":44,"skipped":817,"failed":0} -SSS +Oct 26 16:17:37.241: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-runtime-2835" for this suite. +•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":280,"completed":48,"skipped":736,"failed":0} +SSSSSSSSS ------------------------------ -[k8s.io] Lease - lease API should be available [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Lease - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 16:01:23.705: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename lease-test -STEP: Waiting for a default service account to be provisioned in namespace -[It] lease API should be available [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[AfterEach] [k8s.io] Lease - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:01:24.350: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "lease-test-3122" for this suite. -•{"msg":"PASSED [k8s.io] Lease lease API should be available [Conformance]","total":280,"completed":45,"skipped":820,"failed":0} -SSSSSSS ------------------------------- -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should mutate configmap [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 16:01:24.468: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename webhook -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 -STEP: Setting up server cert -STEP: Create role binding to let webhook read extension-apiserver-authentication -STEP: Deploying the webhook pod -STEP: Wait for the deployment to be ready -Sep 21 16:01:25.599: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -Sep 21 16:01:27.883: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300885, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300885, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300885, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300885, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)} -STEP: Deploying the webhook service -STEP: Verifying the service has paired with the endpoint -Sep 21 16:01:31.007: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should mutate configmap [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Registering the mutating configmap webhook via the AdmissionRegistration API -STEP: create a configmap that should be updated by the webhook -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:01:31.506: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-1711" for this suite. -STEP: Destroying namespace "webhook-1711-markers" for this suite. -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 - -• [SLOW TEST:8.279 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - should mutate configmap [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance]","total":280,"completed":46,"skipped":827,"failed":0} -SSSSSSSSSSSS ------------------------------- -[sig-cli] Kubectl client Proxy server - should support proxy with --port 0 [Conformance] +[sig-cli] Kubectl client Kubectl logs + should be able to retrieve and filter logs [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:01:32.747: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:17:37.281: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278 -[It] should support proxy with --port 0 [Conformance] +[BeforeEach] Kubectl logs + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1464 +STEP: creating an pod +Oct 26 16:17:37.429: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 run logs-generator --generator=run-pod/v1 --image=gcr.io/kubernetes-e2e-test-images/agnhost:2.8 --namespace=kubectl-8604 -- logs-generator --log-lines-total 100 --run-duration 20s' +Oct 26 16:17:37.632: INFO: stderr: "" +Oct 26 16:17:37.632: INFO: stdout: "pod/logs-generator created\n" +[It] should be able to retrieve and filter logs [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: starting the proxy server -Sep 21 16:01:33.054: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-468631458 proxy -p 0 --disable-filter' -STEP: curling proxy /api/ output +STEP: Waiting for log generator to start. +Oct 26 16:17:37.632: INFO: Waiting up to 5m0s for 1 pods to be running and ready, or succeeded: [logs-generator] +Oct 26 16:17:37.632: INFO: Waiting up to 5m0s for pod "logs-generator" in namespace "kubectl-8604" to be "running and ready, or succeeded" +Oct 26 16:17:37.647: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 14.245337ms +Oct 26 16:17:39.662: INFO: Pod "logs-generator": Phase="Running", Reason="", readiness=true. Elapsed: 2.02987295s +Oct 26 16:17:39.662: INFO: Pod "logs-generator" satisfied condition "running and ready, or succeeded" +Oct 26 16:17:39.662: INFO: Wanted all 1 pods to be running and ready, or succeeded. Result: true. Pods: [logs-generator] +STEP: checking for a matching strings +Oct 26 16:17:39.663: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 logs logs-generator logs-generator --namespace=kubectl-8604' +Oct 26 16:17:39.893: INFO: stderr: "" +Oct 26 16:17:39.893: INFO: stdout: "I1026 16:17:39.042820 1 logs_generator.go:76] 0 PUT /api/v1/namespaces/kube-system/pods/wrw 330\nI1026 16:17:39.242960 1 logs_generator.go:76] 1 GET /api/v1/namespaces/ns/pods/5s66 345\nI1026 16:17:39.443020 1 logs_generator.go:76] 2 GET /api/v1/namespaces/kube-system/pods/cxc 591\nI1026 16:17:39.643000 1 logs_generator.go:76] 3 PUT /api/v1/namespaces/kube-system/pods/6fws 597\nI1026 16:17:39.842983 1 logs_generator.go:76] 4 PUT /api/v1/namespaces/kube-system/pods/7vbq 598\n" +STEP: limiting log lines +Oct 26 16:17:39.893: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 logs logs-generator logs-generator --namespace=kubectl-8604 --tail=1' +Oct 26 16:17:40.072: INFO: stderr: "" +Oct 26 16:17:40.072: INFO: stdout: "I1026 16:17:40.042992 1 logs_generator.go:76] 5 PUT /api/v1/namespaces/default/pods/nfj 512\n" +Oct 26 16:17:40.072: INFO: got output "I1026 16:17:40.042992 1 logs_generator.go:76] 5 PUT /api/v1/namespaces/default/pods/nfj 512\n" +STEP: limiting log bytes +Oct 26 16:17:40.072: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 logs logs-generator logs-generator --namespace=kubectl-8604 --limit-bytes=1' +Oct 26 16:17:40.287: INFO: stderr: "" +Oct 26 16:17:40.287: INFO: stdout: "I" +Oct 26 16:17:40.287: INFO: got output "I" +STEP: exposing timestamps +Oct 26 16:17:40.287: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 logs logs-generator logs-generator --namespace=kubectl-8604 --tail=1 --timestamps' +Oct 26 16:17:40.480: INFO: stderr: "" +Oct 26 16:17:40.480: INFO: stdout: "2020-10-26T11:17:40.443040238-05:00 I1026 16:17:40.442979 1 logs_generator.go:76] 7 GET /api/v1/namespaces/ns/pods/9kzd 491\n" +Oct 26 16:17:40.480: INFO: got output "2020-10-26T11:17:40.443040238-05:00 I1026 16:17:40.442979 1 logs_generator.go:76] 7 GET /api/v1/namespaces/ns/pods/9kzd 491\n" +STEP: restricting to a time range +Oct 26 16:17:42.980: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 logs logs-generator logs-generator --namespace=kubectl-8604 --since=1s' +Oct 26 16:17:43.160: INFO: stderr: "" +Oct 26 16:17:43.160: INFO: stdout: "I1026 16:17:42.243001 1 logs_generator.go:76] 16 GET /api/v1/namespaces/kube-system/pods/swc 544\nI1026 16:17:42.443004 1 logs_generator.go:76] 17 GET /api/v1/namespaces/default/pods/d6lp 519\nI1026 16:17:42.643003 1 logs_generator.go:76] 18 PUT /api/v1/namespaces/kube-system/pods/7mrr 341\nI1026 16:17:42.843000 1 logs_generator.go:76] 19 POST /api/v1/namespaces/default/pods/z7m 347\nI1026 16:17:43.042977 1 logs_generator.go:76] 20 POST /api/v1/namespaces/default/pods/bfgf 339\n" +Oct 26 16:17:43.160: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 logs logs-generator logs-generator --namespace=kubectl-8604 --since=24h' +Oct 26 16:17:43.353: INFO: stderr: "" +Oct 26 16:17:43.353: INFO: stdout: "I1026 16:17:39.042820 1 logs_generator.go:76] 0 PUT /api/v1/namespaces/kube-system/pods/wrw 330\nI1026 16:17:39.242960 1 logs_generator.go:76] 1 GET /api/v1/namespaces/ns/pods/5s66 345\nI1026 16:17:39.443020 1 logs_generator.go:76] 2 GET /api/v1/namespaces/kube-system/pods/cxc 591\nI1026 16:17:39.643000 1 logs_generator.go:76] 3 PUT /api/v1/namespaces/kube-system/pods/6fws 597\nI1026 16:17:39.842983 1 logs_generator.go:76] 4 PUT /api/v1/namespaces/kube-system/pods/7vbq 598\nI1026 16:17:40.042992 1 logs_generator.go:76] 5 PUT /api/v1/namespaces/default/pods/nfj 512\nI1026 16:17:40.242969 1 logs_generator.go:76] 6 GET /api/v1/namespaces/ns/pods/g4n 469\nI1026 16:17:40.442979 1 logs_generator.go:76] 7 GET /api/v1/namespaces/ns/pods/9kzd 491\nI1026 16:17:40.642961 1 logs_generator.go:76] 8 PUT /api/v1/namespaces/kube-system/pods/wfp 549\nI1026 16:17:40.842958 1 logs_generator.go:76] 9 POST /api/v1/namespaces/kube-system/pods/csn 347\nI1026 16:17:41.043018 1 logs_generator.go:76] 10 PUT /api/v1/namespaces/default/pods/cbkp 347\nI1026 16:17:41.242976 1 logs_generator.go:76] 11 GET /api/v1/namespaces/default/pods/hmt 238\nI1026 16:17:41.443007 1 logs_generator.go:76] 12 GET /api/v1/namespaces/default/pods/n9f 531\nI1026 16:17:41.642983 1 logs_generator.go:76] 13 PUT /api/v1/namespaces/kube-system/pods/zwrf 557\nI1026 16:17:41.842991 1 logs_generator.go:76] 14 GET /api/v1/namespaces/default/pods/t74d 386\nI1026 16:17:42.042996 1 logs_generator.go:76] 15 POST /api/v1/namespaces/ns/pods/9jnf 224\nI1026 16:17:42.243001 1 logs_generator.go:76] 16 GET /api/v1/namespaces/kube-system/pods/swc 544\nI1026 16:17:42.443004 1 logs_generator.go:76] 17 GET /api/v1/namespaces/default/pods/d6lp 519\nI1026 16:17:42.643003 1 logs_generator.go:76] 18 PUT /api/v1/namespaces/kube-system/pods/7mrr 341\nI1026 16:17:42.843000 1 logs_generator.go:76] 19 POST /api/v1/namespaces/default/pods/z7m 347\nI1026 16:17:43.042977 1 logs_generator.go:76] 20 POST /api/v1/namespaces/default/pods/bfgf 339\nI1026 16:17:43.242962 1 logs_generator.go:76] 21 PUT /api/v1/namespaces/default/pods/8fb 403\n" +[AfterEach] Kubectl logs + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1470 +Oct 26 16:17:43.354: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete pod logs-generator --namespace=kubectl-8604' +Oct 26 16:17:56.689: INFO: stderr: "" +Oct 26 16:17:56.689: INFO: stdout: "pod \"logs-generator\" deleted\n" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:01:33.155: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-1922" for this suite. -•{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support proxy with --port 0 [Conformance]","total":280,"completed":47,"skipped":839,"failed":0} -SSSSSS +Oct 26 16:17:56.689: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-8604" for this suite. + +• [SLOW TEST:19.459 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + Kubectl logs + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1460 + should be able to retrieve and filter logs [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -[k8s.io] Docker Containers - should use the image defaults if command and args are blank [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Docker Containers - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 16:01:33.276: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename containers -STEP: Waiting for a default service account to be provisioned in namespace -[It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[AfterEach] [k8s.io] Docker Containers - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:01:37.937: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "containers-8783" for this suite. -•{"msg":"PASSED [k8s.io] Docker Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance]","total":280,"completed":48,"skipped":845,"failed":0} -SSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs [Conformance]","total":280,"completed":49,"skipped":745,"failed":0} +SSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] ConfigMap - should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] +[k8s.io] Container Runtime blackbox test when starting a container that exits + should run with the expected status [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] ConfigMap +[BeforeEach] [k8s.io] Container Runtime /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:01:38.061: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename configmap +Oct 26 16:17:56.741: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename container-runtime STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] +[It] should run with the expected status [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating configMap with name configmap-test-volume-57f99fe3-c925-4389-9617-3fa5e24946fa -STEP: Creating a pod to test consume configMaps -Sep 21 16:01:38.844: INFO: Waiting up to 5m0s for pod "pod-configmaps-7dfdc870-3f47-49d3-a08a-b1a5d1bcc893" in namespace "configmap-2734" to be "success or failure" -Sep 21 16:01:39.573: INFO: Pod "pod-configmaps-7dfdc870-3f47-49d3-a08a-b1a5d1bcc893": Phase="Pending", Reason="", readiness=false. Elapsed: 729.011105ms -Sep 21 16:01:41.602: INFO: Pod "pod-configmaps-7dfdc870-3f47-49d3-a08a-b1a5d1bcc893": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.757878668s -STEP: Saw pod success -Sep 21 16:01:41.602: INFO: Pod "pod-configmaps-7dfdc870-3f47-49d3-a08a-b1a5d1bcc893" satisfied condition "success or failure" -Sep 21 16:01:41.634: INFO: Trying to get logs from node 10.241.51.147 pod pod-configmaps-7dfdc870-3f47-49d3-a08a-b1a5d1bcc893 container configmap-volume-test: -STEP: delete the pod -Sep 21 16:01:41.820: INFO: Waiting for pod pod-configmaps-7dfdc870-3f47-49d3-a08a-b1a5d1bcc893 to disappear -Sep 21 16:01:41.849: INFO: Pod pod-configmaps-7dfdc870-3f47-49d3-a08a-b1a5d1bcc893 no longer exists -[AfterEach] [sig-storage] ConfigMap +STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount' +STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase' +STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition +STEP: Container 'terminate-cmd-rpa': should get the expected 'State' +STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance] +STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount' +STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase' +STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition +STEP: Container 'terminate-cmd-rpof': should get the expected 'State' +STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance] +STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount' +STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase' +STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition +STEP: Container 'terminate-cmd-rpn': should get the expected 'State' +STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance] +[AfterEach] [k8s.io] Container Runtime /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:01:41.849: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-2734" for this suite. -•{"msg":"PASSED [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":280,"completed":49,"skipped":866,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +Oct 26 16:18:22.965: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-runtime-1404" for this suite. + +• [SLOW TEST:26.287 seconds] +[k8s.io] Container Runtime +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 + blackbox test + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38 + when starting a container that exits + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:39 + should run with the expected status [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -[sig-storage] EmptyDir volumes - should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] +{"msg":"PASSED [k8s.io] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance]","total":280,"completed":50,"skipped":763,"failed":0} +SSSSSSSSSSSSSSSSSS +------------------------------ +[sig-network] Service endpoints latency + should not be very high [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] EmptyDir volumes +[BeforeEach] [sig-network] Service endpoints latency /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:01:41.932: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename emptydir +Oct 26 16:18:23.028: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename svc-latency STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] +[It] should not be very high [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test emptydir 0777 on tmpfs -Sep 21 16:01:43.027: INFO: Waiting up to 5m0s for pod "pod-65712ce7-4ffc-469e-b8d4-9ebb5d9d29ef" in namespace "emptydir-3697" to be "success or failure" -Sep 21 16:01:43.055: INFO: Pod "pod-65712ce7-4ffc-469e-b8d4-9ebb5d9d29ef": Phase="Pending", Reason="", readiness=false. Elapsed: 28.078799ms -Sep 21 16:01:45.090: INFO: Pod "pod-65712ce7-4ffc-469e-b8d4-9ebb5d9d29ef": Phase="Pending", Reason="", readiness=false. Elapsed: 2.062665785s -Sep 21 16:01:47.115: INFO: Pod "pod-65712ce7-4ffc-469e-b8d4-9ebb5d9d29ef": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.087898906s -STEP: Saw pod success -Sep 21 16:01:47.115: INFO: Pod "pod-65712ce7-4ffc-469e-b8d4-9ebb5d9d29ef" satisfied condition "success or failure" -Sep 21 16:01:47.142: INFO: Trying to get logs from node 10.241.51.147 pod pod-65712ce7-4ffc-469e-b8d4-9ebb5d9d29ef container test-container: -STEP: delete the pod -Sep 21 16:01:48.060: INFO: Waiting for pod pod-65712ce7-4ffc-469e-b8d4-9ebb5d9d29ef to disappear -Sep 21 16:01:48.093: INFO: Pod pod-65712ce7-4ffc-469e-b8d4-9ebb5d9d29ef no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +Oct 26 16:18:23.188: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: creating replication controller svc-latency-rc in namespace svc-latency-4784 +I1026 16:18:23.226369 26 runners.go:189] Created replication controller with name: svc-latency-rc, namespace: svc-latency-4784, replica count: 1 +I1026 16:18:24.276924 26 runners.go:189] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I1026 16:18:25.277181 26 runners.go:189] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I1026 16:18:26.277407 26 runners.go:189] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Oct 26 16:18:26.408: INFO: Created: latency-svc-zk4bn +Oct 26 16:18:26.428: INFO: Got endpoints: latency-svc-zk4bn [51.151244ms] +Oct 26 16:18:26.463: INFO: Created: latency-svc-sgg2h +Oct 26 16:18:26.479: INFO: Created: latency-svc-lt8f6 +Oct 26 16:18:26.494: INFO: Got endpoints: latency-svc-sgg2h [64.458135ms] +Oct 26 16:18:26.500: INFO: Got endpoints: latency-svc-lt8f6 [69.925557ms] +Oct 26 16:18:26.500: INFO: Created: latency-svc-gzjvb +Oct 26 16:18:26.513: INFO: Created: latency-svc-x7m9w +Oct 26 16:18:26.519: INFO: Got endpoints: latency-svc-gzjvb [89.272233ms] +Oct 26 16:18:26.534: INFO: Got endpoints: latency-svc-x7m9w [104.541918ms] +Oct 26 16:18:26.535: INFO: Created: latency-svc-h2fc2 +Oct 26 16:18:26.543: INFO: Created: latency-svc-xxfqc +Oct 26 16:18:26.549: INFO: Got endpoints: latency-svc-h2fc2 [120.704724ms] +Oct 26 16:18:26.559: INFO: Created: latency-svc-xk97h +Oct 26 16:18:26.566: INFO: Got endpoints: latency-svc-xxfqc [135.453304ms] +Oct 26 16:18:26.571: INFO: Got endpoints: latency-svc-xk97h [139.417549ms] +Oct 26 16:18:26.576: INFO: Created: latency-svc-2bvnj +Oct 26 16:18:26.588: INFO: Created: latency-svc-zgkqb +Oct 26 16:18:26.588: INFO: Got endpoints: latency-svc-2bvnj [157.199948ms] +Oct 26 16:18:26.601: INFO: Created: latency-svc-r4zhl +Oct 26 16:18:26.606: INFO: Got endpoints: latency-svc-zgkqb [176.126969ms] +Oct 26 16:18:26.612: INFO: Created: latency-svc-8qk9z +Oct 26 16:18:26.617: INFO: Got endpoints: latency-svc-r4zhl [185.959335ms] +Oct 26 16:18:26.629: INFO: Created: latency-svc-mqgnb +Oct 26 16:18:26.629: INFO: Got endpoints: latency-svc-8qk9z [198.039534ms] +Oct 26 16:18:26.646: INFO: Created: latency-svc-sxxn7 +Oct 26 16:18:26.652: INFO: Got endpoints: latency-svc-mqgnb [222.626321ms] +Oct 26 16:18:26.663: INFO: Got endpoints: latency-svc-sxxn7 [230.927291ms] +Oct 26 16:18:26.671: INFO: Created: latency-svc-tv6ds +Oct 26 16:18:26.695: INFO: Got endpoints: latency-svc-tv6ds [263.710908ms] +Oct 26 16:18:26.700: INFO: Created: latency-svc-sxfs4 +Oct 26 16:18:26.715: INFO: Created: latency-svc-ngvqx +Oct 26 16:18:26.717: INFO: Got endpoints: latency-svc-sxfs4 [65.693812ms] +Oct 26 16:18:26.735: INFO: Got endpoints: latency-svc-ngvqx [303.44432ms] +Oct 26 16:18:26.736: INFO: Created: latency-svc-wlkqv +Oct 26 16:18:26.746: INFO: Created: latency-svc-5c6sx +Oct 26 16:18:26.752: INFO: Got endpoints: latency-svc-wlkqv [257.716652ms] +Oct 26 16:18:26.758: INFO: Got endpoints: latency-svc-5c6sx [258.28771ms] +Oct 26 16:18:26.759: INFO: Created: latency-svc-ndkns +Oct 26 16:18:26.776: INFO: Created: latency-svc-vj58b +Oct 26 16:18:26.777: INFO: Got endpoints: latency-svc-ndkns [258.052748ms] +Oct 26 16:18:26.795: INFO: Created: latency-svc-bpp69 +Oct 26 16:18:26.795: INFO: Got endpoints: latency-svc-vj58b [261.475502ms] +Oct 26 16:18:26.807: INFO: Got endpoints: latency-svc-bpp69 [257.950744ms] +Oct 26 16:18:26.808: INFO: Created: latency-svc-psxcc +Oct 26 16:18:26.821: INFO: Created: latency-svc-2s8sv +Oct 26 16:18:26.827: INFO: Got endpoints: latency-svc-psxcc [261.335435ms] +Oct 26 16:18:26.838: INFO: Created: latency-svc-t7kpk +Oct 26 16:18:26.845: INFO: Got endpoints: latency-svc-2s8sv [274.133042ms] +Oct 26 16:18:26.851: INFO: Created: latency-svc-tpz8n +Oct 26 16:18:26.858: INFO: Got endpoints: latency-svc-t7kpk [269.506807ms] +Oct 26 16:18:26.870: INFO: Got endpoints: latency-svc-tpz8n [264.583686ms] +Oct 26 16:18:26.871: INFO: Created: latency-svc-tld6h +Oct 26 16:18:26.883: INFO: Got endpoints: latency-svc-tld6h [266.048887ms] +Oct 26 16:18:26.887: INFO: Created: latency-svc-mf6sr +Oct 26 16:18:26.906: INFO: Got endpoints: latency-svc-mf6sr [277.755848ms] +Oct 26 16:18:26.907: INFO: Created: latency-svc-cjvbt +Oct 26 16:18:26.914: INFO: Created: latency-svc-kl4mb +Oct 26 16:18:26.920: INFO: Got endpoints: latency-svc-cjvbt [257.416107ms] +Oct 26 16:18:26.929: INFO: Created: latency-svc-kgn6t +Oct 26 16:18:26.931: INFO: Got endpoints: latency-svc-kl4mb [236.119987ms] +Oct 26 16:18:26.944: INFO: Created: latency-svc-k886z +Oct 26 16:18:26.948: INFO: Got endpoints: latency-svc-kgn6t [230.273005ms] +Oct 26 16:18:26.961: INFO: Created: latency-svc-s6krt +Oct 26 16:18:26.961: INFO: Got endpoints: latency-svc-k886z [225.619596ms] +Oct 26 16:18:26.972: INFO: Created: latency-svc-xwpmk +Oct 26 16:18:26.975: INFO: Got endpoints: latency-svc-s6krt [223.443756ms] +Oct 26 16:18:26.987: INFO: Created: latency-svc-bs2r9 +Oct 26 16:18:26.991: INFO: Got endpoints: latency-svc-xwpmk [232.727903ms] +Oct 26 16:18:26.999: INFO: Created: latency-svc-stv9k +Oct 26 16:18:27.002: INFO: Got endpoints: latency-svc-bs2r9 [224.986398ms] +Oct 26 16:18:27.038: INFO: Created: latency-svc-qr7x8 +Oct 26 16:18:27.038: INFO: Created: latency-svc-gsbkv +Oct 26 16:18:27.038: INFO: Got endpoints: latency-svc-qr7x8 [231.205233ms] +Oct 26 16:18:27.039: INFO: Got endpoints: latency-svc-stv9k [243.346649ms] +Oct 26 16:18:27.042: INFO: Created: latency-svc-gbtsw +Oct 26 16:18:27.047: INFO: Got endpoints: latency-svc-gsbkv [220.118811ms] +Oct 26 16:18:27.054: INFO: Created: latency-svc-dwr8l +Oct 26 16:18:27.055: INFO: Got endpoints: latency-svc-gbtsw [210.146513ms] +Oct 26 16:18:27.067: INFO: Created: latency-svc-ncqt9 +Oct 26 16:18:27.071: INFO: Got endpoints: latency-svc-dwr8l [212.554387ms] +Oct 26 16:18:27.078: INFO: Created: latency-svc-bgkf7 +Oct 26 16:18:27.085: INFO: Got endpoints: latency-svc-ncqt9 [214.612087ms] +Oct 26 16:18:27.092: INFO: Created: latency-svc-nppwm +Oct 26 16:18:27.093: INFO: Got endpoints: latency-svc-bgkf7 [210.158289ms] +Oct 26 16:18:27.103: INFO: Created: latency-svc-2t7zr +Oct 26 16:18:27.106: INFO: Got endpoints: latency-svc-nppwm [199.188607ms] +Oct 26 16:18:27.117: INFO: Got endpoints: latency-svc-2t7zr [196.655344ms] +Oct 26 16:18:27.117: INFO: Created: latency-svc-r4k42 +Oct 26 16:18:27.134: INFO: Got endpoints: latency-svc-r4k42 [202.944498ms] +Oct 26 16:18:27.135: INFO: Created: latency-svc-wrqnm +Oct 26 16:18:27.145: INFO: Got endpoints: latency-svc-wrqnm [197.572453ms] +Oct 26 16:18:27.146: INFO: Created: latency-svc-sc8l5 +Oct 26 16:18:27.158: INFO: Created: latency-svc-ccbm6 +Oct 26 16:18:27.171: INFO: Created: latency-svc-d6bpw +Oct 26 16:18:27.171: INFO: Got endpoints: latency-svc-sc8l5 [209.838444ms] +Oct 26 16:18:27.174: INFO: Got endpoints: latency-svc-ccbm6 [198.348252ms] +Oct 26 16:18:27.182: INFO: Got endpoints: latency-svc-d6bpw [191.240446ms] +Oct 26 16:18:27.183: INFO: Created: latency-svc-n9lfc +Oct 26 16:18:27.207: INFO: Created: latency-svc-d2fwz +Oct 26 16:18:27.207: INFO: Created: latency-svc-59fmr +Oct 26 16:18:27.207: INFO: Got endpoints: latency-svc-n9lfc [205.437296ms] +Oct 26 16:18:27.208: INFO: Got endpoints: latency-svc-d2fwz [169.820195ms] +Oct 26 16:18:27.223: INFO: Got endpoints: latency-svc-59fmr [184.756684ms] +Oct 26 16:18:27.225: INFO: Created: latency-svc-xhn9m +Oct 26 16:18:27.234: INFO: Created: latency-svc-skksh +Oct 26 16:18:27.241: INFO: Got endpoints: latency-svc-xhn9m [193.848764ms] +Oct 26 16:18:27.246: INFO: Created: latency-svc-5khn6 +Oct 26 16:18:27.251: INFO: Got endpoints: latency-svc-skksh [196.306924ms] +Oct 26 16:18:27.262: INFO: Created: latency-svc-tfvlz +Oct 26 16:18:27.262: INFO: Got endpoints: latency-svc-5khn6 [191.620443ms] +Oct 26 16:18:27.275: INFO: Created: latency-svc-f6p8r +Oct 26 16:18:27.277: INFO: Got endpoints: latency-svc-tfvlz [191.972238ms] +Oct 26 16:18:27.290: INFO: Created: latency-svc-tvhkj +Oct 26 16:18:27.291: INFO: Got endpoints: latency-svc-f6p8r [197.513771ms] +Oct 26 16:18:27.321: INFO: Got endpoints: latency-svc-tvhkj [215.515652ms] +Oct 26 16:18:27.325: INFO: Created: latency-svc-26nhm +Oct 26 16:18:27.346: INFO: Created: latency-svc-bmvw5 +Oct 26 16:18:27.346: INFO: Got endpoints: latency-svc-26nhm [229.309675ms] +Oct 26 16:18:27.355: INFO: Created: latency-svc-27t4v +Oct 26 16:18:27.360: INFO: Got endpoints: latency-svc-bmvw5 [225.832224ms] +Oct 26 16:18:27.370: INFO: Created: latency-svc-sgw87 +Oct 26 16:18:27.373: INFO: Got endpoints: latency-svc-27t4v [227.99184ms] +Oct 26 16:18:27.382: INFO: Created: latency-svc-tzns9 +Oct 26 16:18:27.391: INFO: Got endpoints: latency-svc-sgw87 [220.149004ms] +Oct 26 16:18:27.399: INFO: Created: latency-svc-n85wn +Oct 26 16:18:27.400: INFO: Got endpoints: latency-svc-tzns9 [226.385604ms] +Oct 26 16:18:27.413: INFO: Created: latency-svc-g92vv +Oct 26 16:18:27.417: INFO: Got endpoints: latency-svc-n85wn [234.013026ms] +Oct 26 16:18:27.428: INFO: Created: latency-svc-rxqg4 +Oct 26 16:18:27.432: INFO: Got endpoints: latency-svc-g92vv [224.321957ms] +Oct 26 16:18:27.442: INFO: Created: latency-svc-gtzc2 +Oct 26 16:18:27.451: INFO: Got endpoints: latency-svc-rxqg4 [242.809666ms] +Oct 26 16:18:27.463: INFO: Created: latency-svc-xwwwz +Oct 26 16:18:27.469: INFO: Got endpoints: latency-svc-gtzc2 [245.657994ms] +Oct 26 16:18:27.475: INFO: Created: latency-svc-xwd4t +Oct 26 16:18:27.483: INFO: Got endpoints: latency-svc-xwwwz [241.956045ms] +Oct 26 16:18:27.490: INFO: Created: latency-svc-ln4ds +Oct 26 16:18:27.502: INFO: Got endpoints: latency-svc-ln4ds [239.317951ms] +Oct 26 16:18:27.502: INFO: Got endpoints: latency-svc-xwd4t [250.570865ms] +Oct 26 16:18:27.504: INFO: Created: latency-svc-zxnws +Oct 26 16:18:27.514: INFO: Created: latency-svc-rqzkp +Oct 26 16:18:27.521: INFO: Got endpoints: latency-svc-zxnws [244.162258ms] +Oct 26 16:18:27.531: INFO: Created: latency-svc-zgs7q +Oct 26 16:18:27.531: INFO: Got endpoints: latency-svc-rqzkp [240.150789ms] +Oct 26 16:18:27.563: INFO: Created: latency-svc-gz8jj +Oct 26 16:18:27.564: INFO: Got endpoints: latency-svc-zgs7q [243.150554ms] +Oct 26 16:18:27.565: INFO: Created: latency-svc-cb4fw +Oct 26 16:18:27.566: INFO: Got endpoints: latency-svc-cb4fw [219.414313ms] +Oct 26 16:18:27.578: INFO: Created: latency-svc-jqn9x +Oct 26 16:18:27.581: INFO: Got endpoints: latency-svc-gz8jj [221.033591ms] +Oct 26 16:18:27.592: INFO: Got endpoints: latency-svc-jqn9x [218.734707ms] +Oct 26 16:18:27.596: INFO: Created: latency-svc-pwf6m +Oct 26 16:18:27.611: INFO: Created: latency-svc-kcwh8 +Oct 26 16:18:27.614: INFO: Got endpoints: latency-svc-pwf6m [222.912133ms] +Oct 26 16:18:27.625: INFO: Got endpoints: latency-svc-kcwh8 [224.54588ms] +Oct 26 16:18:27.629: INFO: Created: latency-svc-p825q +Oct 26 16:18:27.648: INFO: Created: latency-svc-qj4f8 +Oct 26 16:18:27.658: INFO: Created: latency-svc-glvtg +Oct 26 16:18:27.658: INFO: Got endpoints: latency-svc-p825q [241.622379ms] +Oct 26 16:18:27.659: INFO: Got endpoints: latency-svc-qj4f8 [227.541475ms] +Oct 26 16:18:27.676: INFO: Created: latency-svc-nnmwr +Oct 26 16:18:27.685: INFO: Got endpoints: latency-svc-glvtg [234.221893ms] +Oct 26 16:18:27.688: INFO: Created: latency-svc-l9mkm +Oct 26 16:18:27.689: INFO: Got endpoints: latency-svc-nnmwr [220.099892ms] +Oct 26 16:18:27.705: INFO: Created: latency-svc-fnwjg +Oct 26 16:18:27.706: INFO: Got endpoints: latency-svc-l9mkm [222.578714ms] +Oct 26 16:18:27.720: INFO: Created: latency-svc-d2hqw +Oct 26 16:18:27.721: INFO: Got endpoints: latency-svc-fnwjg [219.439185ms] +Oct 26 16:18:27.737: INFO: Got endpoints: latency-svc-d2hqw [235.072376ms] +Oct 26 16:18:27.740: INFO: Created: latency-svc-vg8fr +Oct 26 16:18:27.748: INFO: Created: latency-svc-s9zpx +Oct 26 16:18:27.763: INFO: Got endpoints: latency-svc-vg8fr [242.15ms] +Oct 26 16:18:27.769: INFO: Created: latency-svc-6q89f +Oct 26 16:18:27.769: INFO: Got endpoints: latency-svc-s9zpx [237.871409ms] +Oct 26 16:18:27.785: INFO: Got endpoints: latency-svc-6q89f [220.209189ms] +Oct 26 16:18:27.787: INFO: Created: latency-svc-fhlnp +Oct 26 16:18:27.796: INFO: Got endpoints: latency-svc-fhlnp [230.041747ms] +Oct 26 16:18:27.797: INFO: Created: latency-svc-7s8z2 +Oct 26 16:18:27.818: INFO: Created: latency-svc-hhsmk +Oct 26 16:18:27.818: INFO: Got endpoints: latency-svc-7s8z2 [236.859982ms] +Oct 26 16:18:27.825: INFO: Created: latency-svc-mmtpg +Oct 26 16:18:27.825: INFO: Got endpoints: latency-svc-hhsmk [233.340731ms] +Oct 26 16:18:27.840: INFO: Created: latency-svc-76vqb +Oct 26 16:18:27.843: INFO: Got endpoints: latency-svc-mmtpg [228.707773ms] +Oct 26 16:18:27.855: INFO: Created: latency-svc-tprrr +Oct 26 16:18:27.856: INFO: Got endpoints: latency-svc-76vqb [231.073994ms] +Oct 26 16:18:27.869: INFO: Created: latency-svc-k2t4s +Oct 26 16:18:27.871: INFO: Got endpoints: latency-svc-tprrr [211.892714ms] +Oct 26 16:18:27.881: INFO: Created: latency-svc-2w99n +Oct 26 16:18:27.883: INFO: Got endpoints: latency-svc-k2t4s [224.444392ms] +Oct 26 16:18:27.893: INFO: Created: latency-svc-zd95s +Oct 26 16:18:27.899: INFO: Got endpoints: latency-svc-2w99n [213.358943ms] +Oct 26 16:18:27.905: INFO: Created: latency-svc-kwp76 +Oct 26 16:18:27.908: INFO: Got endpoints: latency-svc-zd95s [218.664871ms] +Oct 26 16:18:27.920: INFO: Got endpoints: latency-svc-kwp76 [214.075803ms] +Oct 26 16:18:27.921: INFO: Created: latency-svc-j7wfx +Oct 26 16:18:27.937: INFO: Created: latency-svc-s2l5k +Oct 26 16:18:27.937: INFO: Got endpoints: latency-svc-j7wfx [215.849097ms] +Oct 26 16:18:27.948: INFO: Created: latency-svc-k9g6k +Oct 26 16:18:27.952: INFO: Got endpoints: latency-svc-s2l5k [214.532441ms] +Oct 26 16:18:27.959: INFO: Created: latency-svc-tr625 +Oct 26 16:18:27.960: INFO: Got endpoints: latency-svc-k9g6k [196.650692ms] +Oct 26 16:18:27.971: INFO: Got endpoints: latency-svc-tr625 [201.92656ms] +Oct 26 16:18:27.974: INFO: Created: latency-svc-thk4h +Oct 26 16:18:27.991: INFO: Created: latency-svc-xzgxl +Oct 26 16:18:27.996: INFO: Got endpoints: latency-svc-thk4h [208.209009ms] +Oct 26 16:18:28.003: INFO: Created: latency-svc-psr2k +Oct 26 16:18:28.005: INFO: Got endpoints: latency-svc-xzgxl [208.976671ms] +Oct 26 16:18:28.016: INFO: Created: latency-svc-lbfj9 +Oct 26 16:18:28.020: INFO: Got endpoints: latency-svc-psr2k [201.599917ms] +Oct 26 16:18:28.030: INFO: Got endpoints: latency-svc-lbfj9 [200.877258ms] +Oct 26 16:18:28.037: INFO: Created: latency-svc-hjp4k +Oct 26 16:18:28.047: INFO: Got endpoints: latency-svc-hjp4k [203.473796ms] +Oct 26 16:18:28.047: INFO: Created: latency-svc-d525v +Oct 26 16:18:28.061: INFO: Created: latency-svc-j7r57 +Oct 26 16:18:28.061: INFO: Got endpoints: latency-svc-d525v [205.452649ms] +Oct 26 16:18:28.075: INFO: Got endpoints: latency-svc-j7r57 [203.756611ms] +Oct 26 16:18:28.076: INFO: Created: latency-svc-sd2dw +Oct 26 16:18:28.091: INFO: Got endpoints: latency-svc-sd2dw [207.449108ms] +Oct 26 16:18:28.092: INFO: Created: latency-svc-ntktb +Oct 26 16:18:28.101: INFO: Created: latency-svc-brmwr +Oct 26 16:18:28.106: INFO: Got endpoints: latency-svc-ntktb [207.169712ms] +Oct 26 16:18:28.113: INFO: Created: latency-svc-kckg9 +Oct 26 16:18:28.118: INFO: Got endpoints: latency-svc-brmwr [209.580416ms] +Oct 26 16:18:28.126: INFO: Created: latency-svc-x8cqw +Oct 26 16:18:28.126: INFO: Got endpoints: latency-svc-kckg9 [205.695074ms] +Oct 26 16:18:28.142: INFO: Got endpoints: latency-svc-x8cqw [204.764243ms] +Oct 26 16:18:28.144: INFO: Created: latency-svc-x9gjh +Oct 26 16:18:28.162: INFO: Created: latency-svc-sbvgh +Oct 26 16:18:28.162: INFO: Got endpoints: latency-svc-x9gjh [209.683755ms] +Oct 26 16:18:28.192: INFO: Created: latency-svc-l9ch6 +Oct 26 16:18:28.192: INFO: Created: latency-svc-qbtqj +Oct 26 16:18:28.193: INFO: Got endpoints: latency-svc-qbtqj [221.038655ms] +Oct 26 16:18:28.193: INFO: Got endpoints: latency-svc-sbvgh [232.564231ms] +Oct 26 16:18:28.205: INFO: Created: latency-svc-rtpsf +Oct 26 16:18:28.205: INFO: Got endpoints: latency-svc-l9ch6 [208.805099ms] +Oct 26 16:18:28.218: INFO: Created: latency-svc-9gmsh +Oct 26 16:18:28.218: INFO: Got endpoints: latency-svc-rtpsf [210.978264ms] +Oct 26 16:18:28.234: INFO: Got endpoints: latency-svc-9gmsh [213.808746ms] +Oct 26 16:18:28.234: INFO: Created: latency-svc-nff2b +Oct 26 16:18:28.244: INFO: Created: latency-svc-9mk95 +Oct 26 16:18:28.249: INFO: Got endpoints: latency-svc-nff2b [218.26996ms] +Oct 26 16:18:28.257: INFO: Created: latency-svc-h9b6j +Oct 26 16:18:28.263: INFO: Got endpoints: latency-svc-9mk95 [216.498021ms] +Oct 26 16:18:28.270: INFO: Created: latency-svc-wbcss +Oct 26 16:18:28.272: INFO: Got endpoints: latency-svc-h9b6j [210.200612ms] +Oct 26 16:18:28.286: INFO: Got endpoints: latency-svc-wbcss [210.364096ms] +Oct 26 16:18:28.287: INFO: Created: latency-svc-6t9kh +Oct 26 16:18:28.300: INFO: Created: latency-svc-4p7q9 +Oct 26 16:18:28.305: INFO: Got endpoints: latency-svc-6t9kh [214.449684ms] +Oct 26 16:18:28.312: INFO: Created: latency-svc-8wqhq +Oct 26 16:18:28.315: INFO: Got endpoints: latency-svc-4p7q9 [208.972675ms] +Oct 26 16:18:28.329: INFO: Got endpoints: latency-svc-8wqhq [210.833647ms] +Oct 26 16:18:28.329: INFO: Created: latency-svc-wl256 +Oct 26 16:18:28.353: INFO: Created: latency-svc-rxrwr +Oct 26 16:18:28.353: INFO: Got endpoints: latency-svc-wl256 [226.799749ms] +Oct 26 16:18:28.365: INFO: Got endpoints: latency-svc-rxrwr [222.978694ms] +Oct 26 16:18:28.367: INFO: Created: latency-svc-l9kqc +Oct 26 16:18:28.380: INFO: Created: latency-svc-4s6nq +Oct 26 16:18:28.380: INFO: Got endpoints: latency-svc-l9kqc [218.301108ms] +Oct 26 16:18:28.393: INFO: Created: latency-svc-ngwvt +Oct 26 16:18:28.395: INFO: Got endpoints: latency-svc-4s6nq [201.895646ms] +Oct 26 16:18:28.408: INFO: Got endpoints: latency-svc-ngwvt [214.906939ms] +Oct 26 16:18:28.409: INFO: Created: latency-svc-8w8nf +Oct 26 16:18:28.424: INFO: Created: latency-svc-jqtxm +Oct 26 16:18:28.426: INFO: Got endpoints: latency-svc-8w8nf [221.034743ms] +Oct 26 16:18:28.438: INFO: Created: latency-svc-6v8dj +Oct 26 16:18:28.445: INFO: Got endpoints: latency-svc-jqtxm [226.105456ms] +Oct 26 16:18:28.455: INFO: Created: latency-svc-f4rd2 +Oct 26 16:18:28.459: INFO: Got endpoints: latency-svc-6v8dj [225.330884ms] +Oct 26 16:18:28.468: INFO: Got endpoints: latency-svc-f4rd2 [219.635374ms] +Oct 26 16:18:28.474: INFO: Created: latency-svc-k7c54 +Oct 26 16:18:28.489: INFO: Created: latency-svc-ld4tl +Oct 26 16:18:28.490: INFO: Got endpoints: latency-svc-k7c54 [226.301389ms] +Oct 26 16:18:28.509: INFO: Got endpoints: latency-svc-ld4tl [236.632027ms] +Oct 26 16:18:28.510: INFO: Created: latency-svc-txm57 +Oct 26 16:18:28.523: INFO: Got endpoints: latency-svc-txm57 [236.794166ms] +Oct 26 16:18:28.528: INFO: Created: latency-svc-gkckn +Oct 26 16:18:28.541: INFO: Created: latency-svc-56b6t +Oct 26 16:18:28.545: INFO: Got endpoints: latency-svc-gkckn [239.646654ms] +Oct 26 16:18:28.561: INFO: Created: latency-svc-lbsql +Oct 26 16:18:28.561: INFO: Got endpoints: latency-svc-56b6t [245.962722ms] +Oct 26 16:18:28.576: INFO: Got endpoints: latency-svc-lbsql [246.92418ms] +Oct 26 16:18:28.580: INFO: Created: latency-svc-4pp87 +Oct 26 16:18:28.597: INFO: Got endpoints: latency-svc-4pp87 [243.795286ms] +Oct 26 16:18:28.601: INFO: Created: latency-svc-xt6zq +Oct 26 16:18:28.614: INFO: Created: latency-svc-cr66v +Oct 26 16:18:28.617: INFO: Got endpoints: latency-svc-xt6zq [251.675213ms] +Oct 26 16:18:28.624: INFO: Created: latency-svc-27lgg +Oct 26 16:18:28.636: INFO: Got endpoints: latency-svc-cr66v [255.294572ms] +Oct 26 16:18:28.641: INFO: Created: latency-svc-qm52q +Oct 26 16:18:28.641: INFO: Got endpoints: latency-svc-27lgg [246.754711ms] +Oct 26 16:18:28.658: INFO: Got endpoints: latency-svc-qm52q [250.00621ms] +Oct 26 16:18:28.659: INFO: Created: latency-svc-vj4n7 +Oct 26 16:18:28.670: INFO: Got endpoints: latency-svc-vj4n7 [243.444745ms] +Oct 26 16:18:28.674: INFO: Created: latency-svc-whqmr +Oct 26 16:18:28.692: INFO: Got endpoints: latency-svc-whqmr [247.37024ms] +Oct 26 16:18:28.703: INFO: Created: latency-svc-86d4x +Oct 26 16:18:28.703: INFO: Created: latency-svc-gh59d +Oct 26 16:18:28.706: INFO: Got endpoints: latency-svc-gh59d [246.79934ms] +Oct 26 16:18:28.714: INFO: Got endpoints: latency-svc-86d4x [245.896116ms] +Oct 26 16:18:28.716: INFO: Created: latency-svc-9klc5 +Oct 26 16:18:28.729: INFO: Created: latency-svc-vtkh7 +Oct 26 16:18:28.730: INFO: Got endpoints: latency-svc-9klc5 [240.66474ms] +Oct 26 16:18:28.743: INFO: Created: latency-svc-29qr2 +Oct 26 16:18:28.745: INFO: Got endpoints: latency-svc-vtkh7 [236.309209ms] +Oct 26 16:18:28.757: INFO: Created: latency-svc-t6hlz +Oct 26 16:18:28.759: INFO: Got endpoints: latency-svc-29qr2 [236.501226ms] +Oct 26 16:18:28.767: INFO: Created: latency-svc-dt9z7 +Oct 26 16:18:28.773: INFO: Got endpoints: latency-svc-t6hlz [228.104682ms] +Oct 26 16:18:28.782: INFO: Created: latency-svc-8qf2k +Oct 26 16:18:28.783: INFO: Got endpoints: latency-svc-dt9z7 [221.072875ms] +Oct 26 16:18:28.792: INFO: Got endpoints: latency-svc-8qf2k [215.82259ms] +Oct 26 16:18:28.793: INFO: Created: latency-svc-n2t6k +Oct 26 16:18:28.809: INFO: Created: latency-svc-v9tvl +Oct 26 16:18:28.815: INFO: Got endpoints: latency-svc-n2t6k [217.476639ms] +Oct 26 16:18:28.825: INFO: Created: latency-svc-65f9v +Oct 26 16:18:28.830: INFO: Got endpoints: latency-svc-v9tvl [213.06128ms] +Oct 26 16:18:28.839: INFO: Created: latency-svc-g4t95 +Oct 26 16:18:28.840: INFO: Got endpoints: latency-svc-65f9v [203.783588ms] +Oct 26 16:18:28.852: INFO: Got endpoints: latency-svc-g4t95 [210.071119ms] +Oct 26 16:18:28.856: INFO: Created: latency-svc-5b2k6 +Oct 26 16:18:28.874: INFO: Created: latency-svc-t49xv +Oct 26 16:18:28.883: INFO: Got endpoints: latency-svc-5b2k6 [224.159929ms] +Oct 26 16:18:28.889: INFO: Created: latency-svc-nc8xh +Oct 26 16:18:28.890: INFO: Got endpoints: latency-svc-t49xv [220.182841ms] +Oct 26 16:18:28.903: INFO: Created: latency-svc-pdwf9 +Oct 26 16:18:28.903: INFO: Got endpoints: latency-svc-nc8xh [206.919156ms] +Oct 26 16:18:28.916: INFO: Got endpoints: latency-svc-pdwf9 [205.590786ms] +Oct 26 16:18:28.920: INFO: Created: latency-svc-6kp74 +Oct 26 16:18:28.933: INFO: Created: latency-svc-7kzwj +Oct 26 16:18:28.933: INFO: Got endpoints: latency-svc-6kp74 [217.850893ms] +Oct 26 16:18:28.970: INFO: Created: latency-svc-xpmvd +Oct 26 16:18:28.970: INFO: Created: latency-svc-vc4mz +Oct 26 16:18:28.970: INFO: Got endpoints: latency-svc-vc4mz [224.980185ms] +Oct 26 16:18:28.970: INFO: Got endpoints: latency-svc-7kzwj [239.392953ms] +Oct 26 16:18:28.974: INFO: Created: latency-svc-dx8v7 +Oct 26 16:18:28.974: INFO: Got endpoints: latency-svc-xpmvd [214.930906ms] +Oct 26 16:18:28.989: INFO: Got endpoints: latency-svc-dx8v7 [215.767994ms] +Oct 26 16:18:28.990: INFO: Created: latency-svc-55p8h +Oct 26 16:18:29.004: INFO: Got endpoints: latency-svc-55p8h [220.292093ms] +Oct 26 16:18:29.006: INFO: Created: latency-svc-fvmw6 +Oct 26 16:18:29.018: INFO: Created: latency-svc-rtxsh +Oct 26 16:18:29.021: INFO: Got endpoints: latency-svc-fvmw6 [228.785875ms] +Oct 26 16:18:29.031: INFO: Created: latency-svc-tfsf5 +Oct 26 16:18:29.035: INFO: Got endpoints: latency-svc-rtxsh [219.267775ms] +Oct 26 16:18:29.044: INFO: Created: latency-svc-rhqlh +Oct 26 16:18:29.044: INFO: Got endpoints: latency-svc-tfsf5 [213.871169ms] +Oct 26 16:18:29.053: INFO: Created: latency-svc-rqsjb +Oct 26 16:18:29.058: INFO: Got endpoints: latency-svc-rhqlh [217.928682ms] +Oct 26 16:18:29.067: INFO: Created: latency-svc-pv47k +Oct 26 16:18:29.068: INFO: Got endpoints: latency-svc-rqsjb [216.437262ms] +Oct 26 16:18:29.081: INFO: Got endpoints: latency-svc-pv47k [198.51434ms] +Oct 26 16:18:29.086: INFO: Created: latency-svc-cwxks +Oct 26 16:18:29.094: INFO: Created: latency-svc-txpfq +Oct 26 16:18:29.100: INFO: Got endpoints: latency-svc-cwxks [209.862304ms] +Oct 26 16:18:29.106: INFO: Created: latency-svc-z89wv +Oct 26 16:18:29.119: INFO: Created: latency-svc-7qjfh +Oct 26 16:18:29.132: INFO: Got endpoints: latency-svc-z89wv [215.447066ms] +Oct 26 16:18:29.133: INFO: Got endpoints: latency-svc-txpfq [229.730191ms] +Oct 26 16:18:29.142: INFO: Created: latency-svc-rx67k +Oct 26 16:18:29.156: INFO: Got endpoints: latency-svc-7qjfh [223.115003ms] +Oct 26 16:18:29.162: INFO: Created: latency-svc-dtnsv +Oct 26 16:18:29.181: INFO: Created: latency-svc-dp6fc +Oct 26 16:18:29.202: INFO: Created: latency-svc-gp9d8 +Oct 26 16:18:29.211: INFO: Created: latency-svc-267v7 +Oct 26 16:18:29.223: INFO: Created: latency-svc-9n5bt +Oct 26 16:18:29.232: INFO: Got endpoints: latency-svc-gp9d8 [243.20897ms] +Oct 26 16:18:29.233: INFO: Got endpoints: latency-svc-dtnsv [262.382001ms] +Oct 26 16:18:29.233: INFO: Got endpoints: latency-svc-dp6fc [259.093786ms] +Oct 26 16:18:29.234: INFO: Got endpoints: latency-svc-rx67k [263.12633ms] +Oct 26 16:18:29.237: INFO: Got endpoints: latency-svc-267v7 [232.9944ms] +Oct 26 16:18:29.241: INFO: Created: latency-svc-jrd9f +Oct 26 16:18:29.245: INFO: Got endpoints: latency-svc-9n5bt [224.101531ms] +Oct 26 16:18:29.264: INFO: Got endpoints: latency-svc-jrd9f [228.771503ms] +Oct 26 16:18:29.274: INFO: Created: latency-svc-cwqhx +Oct 26 16:18:29.286: INFO: Got endpoints: latency-svc-cwqhx [241.28823ms] +Oct 26 16:18:29.305: INFO: Created: latency-svc-57z7w +Oct 26 16:18:29.316: INFO: Created: latency-svc-lkn6x +Oct 26 16:18:29.318: INFO: Got endpoints: latency-svc-57z7w [259.905517ms] +Oct 26 16:18:29.333: INFO: Created: latency-svc-xcclt +Oct 26 16:18:29.333: INFO: Got endpoints: latency-svc-lkn6x [264.902976ms] +Oct 26 16:18:29.351: INFO: Created: latency-svc-prw74 +Oct 26 16:18:29.352: INFO: Got endpoints: latency-svc-xcclt [270.158254ms] +Oct 26 16:18:29.359: INFO: Created: latency-svc-vkrlr +Oct 26 16:18:29.360: INFO: Got endpoints: latency-svc-prw74 [259.518733ms] +Oct 26 16:18:29.370: INFO: Created: latency-svc-5vkrg +Oct 26 16:18:29.375: INFO: Got endpoints: latency-svc-vkrlr [242.468024ms] +Oct 26 16:18:29.387: INFO: Got endpoints: latency-svc-5vkrg [254.22092ms] +Oct 26 16:18:29.387: INFO: Created: latency-svc-xwv54 +Oct 26 16:18:29.404: INFO: Got endpoints: latency-svc-xwv54 [248.202499ms] +Oct 26 16:18:29.406: INFO: Created: latency-svc-nb5sc +Oct 26 16:18:29.430: INFO: Got endpoints: latency-svc-nb5sc [196.606945ms] +Oct 26 16:18:29.431: INFO: Created: latency-svc-sx7jw +Oct 26 16:18:29.435: INFO: Created: latency-svc-qclbm +Oct 26 16:18:29.438: INFO: Got endpoints: latency-svc-sx7jw [205.135711ms] +Oct 26 16:18:29.453: INFO: Created: latency-svc-wb75c +Oct 26 16:18:29.453: INFO: Got endpoints: latency-svc-qclbm [219.870887ms] +Oct 26 16:18:29.482: INFO: Created: latency-svc-7pt2m +Oct 26 16:18:29.483: INFO: Got endpoints: latency-svc-7pt2m [245.81354ms] +Oct 26 16:18:29.482: INFO: Got endpoints: latency-svc-wb75c [249.225377ms] +Oct 26 16:18:29.483: INFO: Latencies: [64.458135ms 65.693812ms 69.925557ms 89.272233ms 104.541918ms 120.704724ms 135.453304ms 139.417549ms 157.199948ms 169.820195ms 176.126969ms 184.756684ms 185.959335ms 191.240446ms 191.620443ms 191.972238ms 193.848764ms 196.306924ms 196.606945ms 196.650692ms 196.655344ms 197.513771ms 197.572453ms 198.039534ms 198.348252ms 198.51434ms 199.188607ms 200.877258ms 201.599917ms 201.895646ms 201.92656ms 202.944498ms 203.473796ms 203.756611ms 203.783588ms 204.764243ms 205.135711ms 205.437296ms 205.452649ms 205.590786ms 205.695074ms 206.919156ms 207.169712ms 207.449108ms 208.209009ms 208.805099ms 208.972675ms 208.976671ms 209.580416ms 209.683755ms 209.838444ms 209.862304ms 210.071119ms 210.146513ms 210.158289ms 210.200612ms 210.364096ms 210.833647ms 210.978264ms 211.892714ms 212.554387ms 213.06128ms 213.358943ms 213.808746ms 213.871169ms 214.075803ms 214.449684ms 214.532441ms 214.612087ms 214.906939ms 214.930906ms 215.447066ms 215.515652ms 215.767994ms 215.82259ms 215.849097ms 216.437262ms 216.498021ms 217.476639ms 217.850893ms 217.928682ms 218.26996ms 218.301108ms 218.664871ms 218.734707ms 219.267775ms 219.414313ms 219.439185ms 219.635374ms 219.870887ms 220.099892ms 220.118811ms 220.149004ms 220.182841ms 220.209189ms 220.292093ms 221.033591ms 221.034743ms 221.038655ms 221.072875ms 222.578714ms 222.626321ms 222.912133ms 222.978694ms 223.115003ms 223.443756ms 224.101531ms 224.159929ms 224.321957ms 224.444392ms 224.54588ms 224.980185ms 224.986398ms 225.330884ms 225.619596ms 225.832224ms 226.105456ms 226.301389ms 226.385604ms 226.799749ms 227.541475ms 227.99184ms 228.104682ms 228.707773ms 228.771503ms 228.785875ms 229.309675ms 229.730191ms 230.041747ms 230.273005ms 230.927291ms 231.073994ms 231.205233ms 232.564231ms 232.727903ms 232.9944ms 233.340731ms 234.013026ms 234.221893ms 235.072376ms 236.119987ms 236.309209ms 236.501226ms 236.632027ms 236.794166ms 236.859982ms 237.871409ms 239.317951ms 239.392953ms 239.646654ms 240.150789ms 240.66474ms 241.28823ms 241.622379ms 241.956045ms 242.15ms 242.468024ms 242.809666ms 243.150554ms 243.20897ms 243.346649ms 243.444745ms 243.795286ms 244.162258ms 245.657994ms 245.81354ms 245.896116ms 245.962722ms 246.754711ms 246.79934ms 246.92418ms 247.37024ms 248.202499ms 249.225377ms 250.00621ms 250.570865ms 251.675213ms 254.22092ms 255.294572ms 257.416107ms 257.716652ms 257.950744ms 258.052748ms 258.28771ms 259.093786ms 259.518733ms 259.905517ms 261.335435ms 261.475502ms 262.382001ms 263.12633ms 263.710908ms 264.583686ms 264.902976ms 266.048887ms 269.506807ms 270.158254ms 274.133042ms 277.755848ms 303.44432ms] +Oct 26 16:18:29.483: INFO: 50 %ile: 222.578714ms +Oct 26 16:18:29.483: INFO: 90 %ile: 257.716652ms +Oct 26 16:18:29.484: INFO: 99 %ile: 277.755848ms +Oct 26 16:18:29.484: INFO: Total sample count: 200 +[AfterEach] [sig-network] Service endpoints latency /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:01:48.093: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-3697" for this suite. +Oct 26 16:18:29.484: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svc-latency-4784" for this suite. -• [SLOW TEST:6.271 seconds] -[sig-storage] EmptyDir volumes -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 - should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] +• [SLOW TEST:6.496 seconds] +[sig-network] Service endpoints latency +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 + should not be very high [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":50,"skipped":898,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [sig-network] Service endpoints latency should not be very high [Conformance]","total":280,"completed":51,"skipped":781,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] ResourceQuota - should create a ResourceQuota and capture the life of a configMap. [Conformance] + should create a ResourceQuota and capture the life of a replica set. [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:01:48.203: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:18:29.527: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace -[It] should create a ResourceQuota and capture the life of a configMap. [Conformance] +[It] should create a ResourceQuota and capture the life of a replica set. [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 STEP: Counting existing ResourceQuota STEP: Creating a ResourceQuota STEP: Ensuring resource quota status is calculated -STEP: Creating a ConfigMap -STEP: Ensuring resource quota status captures configMap creation -STEP: Deleting a ConfigMap +STEP: Creating a ReplicaSet +STEP: Ensuring resource quota status captures replicaset creation +STEP: Deleting a ReplicaSet STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:02:06.063: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-9737" for this suite. +Oct 26 16:18:40.896: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-6979" for this suite. -• [SLOW TEST:17.940 seconds] +• [SLOW TEST:11.411 seconds] [sig-api-machinery] ResourceQuota /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - should create a ResourceQuota and capture the life of a configMap. [Conformance] + should create a ResourceQuota and capture the life of a replica set. [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance]","total":280,"completed":51,"skipped":924,"failed":0} -SSSSSSSSS ------------------------------- -[k8s.io] Security Context When creating a pod with privileged - should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Security Context - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 16:02:06.144: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename security-context-test -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Security Context - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39 -[It] should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 16:02:06.673: INFO: Waiting up to 5m0s for pod "busybox-privileged-false-92513281-7414-49a0-b616-56487f2b9dbc" in namespace "security-context-test-7620" to be "success or failure" -Sep 21 16:02:06.702: INFO: Pod "busybox-privileged-false-92513281-7414-49a0-b616-56487f2b9dbc": Phase="Pending", Reason="", readiness=false. Elapsed: 28.874062ms -Sep 21 16:02:08.733: INFO: Pod "busybox-privileged-false-92513281-7414-49a0-b616-56487f2b9dbc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.059431962s -Sep 21 16:02:08.733: INFO: Pod "busybox-privileged-false-92513281-7414-49a0-b616-56487f2b9dbc" satisfied condition "success or failure" -Sep 21 16:02:08.836: INFO: Got logs for pod "busybox-privileged-false-92513281-7414-49a0-b616-56487f2b9dbc": "ip: RTNETLINK answers: Operation not permitted\n" -[AfterEach] [k8s.io] Security Context - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:02:08.836: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "security-context-test-7620" for this suite. -•{"msg":"PASSED [k8s.io] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":52,"skipped":933,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance]","total":280,"completed":52,"skipped":808,"failed":0} +SSSSSS ------------------------------ -[sig-apps] Deployment - RollingUpdateDeployment should delete old pods and create new ones [Conformance] +[k8s.io] Probing container + with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-apps] Deployment +[BeforeEach] [k8s.io] Probing container /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:02:08.938: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename deployment +Oct 26 16:18:40.943: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename container-probe STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69 -[It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] +[BeforeEach] [k8s.io] Probing container + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 +[It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 16:02:09.215: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) -Sep 21 16:02:09.349: INFO: Pod name sample-pod: Found 1 pods out of 1 -STEP: ensuring each pod is running -Sep 21 16:02:13.405: INFO: Creating deployment "test-rolling-update-deployment" -Sep 21 16:02:13.439: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has -Sep 21 16:02:13.664: INFO: deployment "test-rolling-update-deployment" doesn't have the required revision set -Sep 21 16:02:15.737: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected -Sep 21 16:02:15.772: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300933, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300933, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300933, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736300933, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-67cf4f6444\" is progressing."}}, CollisionCount:(*int32)(nil)} -Sep 21 16:02:17.811: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63 -Sep 21 16:02:17.931: INFO: Deployment "test-rolling-update-deployment": -&Deployment{ObjectMeta:{test-rolling-update-deployment deployment-9110 /apis/apps/v1/namespaces/deployment-9110/deployments/test-rolling-update-deployment 57fc4fc5-7a44-4d55-b842-07ff28af0bd0 53352 1 2020-09-21 16:02:13 +0000 UTC map[name:sample-pod] map[deployment.kubernetes.io/revision:3546343826724305833] [] [] []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod] map[] [] [] []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc003258838 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} [] nil default-scheduler [] [] nil [] map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-09-21 16:02:13 +0000 UTC,LastTransitionTime:2020-09-21 16:02:13 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rolling-update-deployment-67cf4f6444" has successfully progressed.,LastUpdateTime:2020-09-21 16:02:16 +0000 UTC,LastTransitionTime:2020-09-21 16:02:13 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} - -Sep 21 16:02:17.972: INFO: New ReplicaSet "test-rolling-update-deployment-67cf4f6444" of Deployment "test-rolling-update-deployment": -&ReplicaSet{ObjectMeta:{test-rolling-update-deployment-67cf4f6444 deployment-9110 /apis/apps/v1/namespaces/deployment-9110/replicasets/test-rolling-update-deployment-67cf4f6444 887bcb37-c57b-4cad-ad6a-4f649770ebdc 53340 1 2020-09-21 16:02:13 +0000 UTC map[name:sample-pod pod-template-hash:67cf4f6444] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305833] [{apps/v1 Deployment test-rolling-update-deployment 57fc4fc5-7a44-4d55-b842-07ff28af0bd0 0xc003258d17 0xc003258d18}] [] []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 67cf4f6444,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod-template-hash:67cf4f6444] map[] [] [] []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc003258da8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} -Sep 21 16:02:17.972: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": -Sep 21 16:02:17.973: INFO: &ReplicaSet{ObjectMeta:{test-rolling-update-controller deployment-9110 /apis/apps/v1/namespaces/deployment-9110/replicasets/test-rolling-update-controller c53a8dec-a535-4a6f-8fbe-5c013955a258 53350 2 2020-09-21 16:02:09 +0000 UTC map[name:sample-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305832] [{apps/v1 Deployment test-rolling-update-deployment 57fc4fc5-7a44-4d55-b842-07ff28af0bd0 0xc003258c47 0xc003258c48}] [] []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod:httpd] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc003258ca8 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} -Sep 21 16:02:18.000: INFO: Pod "test-rolling-update-deployment-67cf4f6444-tj54d" is available: -&Pod{ObjectMeta:{test-rolling-update-deployment-67cf4f6444-tj54d test-rolling-update-deployment-67cf4f6444- deployment-9110 /api/v1/namespaces/deployment-9110/pods/test-rolling-update-deployment-67cf4f6444-tj54d 3f2505f3-8dc3-4f1e-9286-1f82b6b10ab8 53338 0 2020-09-21 16:02:13 +0000 UTC map[name:sample-pod pod-template-hash:67cf4f6444] map[cni.projectcalico.org/podIP:172.30.81.29/32 cni.projectcalico.org/podIPs:172.30.81.29/32 k8s.v1.cni.cncf.io/networks-status:[{ - "name": "k8s-pod-network", - "ips": [ - "172.30.81.29" - ], - "dns": {} -}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet test-rolling-update-deployment-67cf4f6444 887bcb37-c57b-4cad-ad6a-4f649770ebdc 0xc0032593f7 0xc0032593f8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-79nk8,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-79nk8,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-79nk8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-5q6df,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:02:13 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:02:16 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:02:16 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:02:13 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:172.30.81.29,StartTime:2020-09-21 16:02:13 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:02:15 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:gcr.io/kubernetes-e2e-test-images/agnhost@sha256:02b9e10ea1f7439a03d9bc3ee16d984172ca7eac1818792ad4d721c5c8f72ff4,ContainerID:cri-o://2a87c6cdcc5c23085341729e2c8a5e2d539529490a75c9bea2e2b238a10f19db,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.81.29,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -[AfterEach] [sig-apps] Deployment +[AfterEach] [k8s.io] Probing container /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:02:18.000: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "deployment-9110" for this suite. +Oct 26 16:19:41.170: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-9876" for this suite. -• [SLOW TEST:9.167 seconds] -[sig-apps] Deployment -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 - RollingUpdateDeployment should delete old pods and create new ones [Conformance] +• [SLOW TEST:60.269 seconds] +[k8s.io] Probing container +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 + with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance]","total":280,"completed":53,"skipped":968,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [k8s.io] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]","total":280,"completed":53,"skipped":814,"failed":0} +S ------------------------------ -[k8s.io] [sig-node] Events - should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] +[sig-storage] Projected secret + should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] [sig-node] Events +[BeforeEach] [sig-storage] Projected secret /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:02:18.106: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename events +Oct 26 16:19:41.211: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace -[It] should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] +[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: creating the pod -STEP: submitting the pod to kubernetes -STEP: verifying the pod is in kubernetes -STEP: retrieving the pod -Sep 21 16:02:20.925: INFO: &Pod{ObjectMeta:{send-events-b835b0c5-2790-4803-8784-fbc42fe5b4dc events-9539 /api/v1/namespaces/events-9539/pods/send-events-b835b0c5-2790-4803-8784-fbc42fe5b4dc 66053ec8-8404-4626-a8f4-d671b2b60a2b 53414 0 2020-09-21 16:02:18 +0000 UTC map[name:foo time:530979693] map[cni.projectcalico.org/podIP:172.30.237.119/32 cni.projectcalico.org/podIPs:172.30.237.119/32 k8s.v1.cni.cncf.io/networks-status:[{ - "name": "k8s-pod-network", - "ips": [ - "172.30.237.119" - ], - "dns": {} -}] openshift.io/scc:anyuid] [] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-95lgf,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-95lgf,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:p,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[serve-hostname],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:,HostPort:0,ContainerPort:80,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-95lgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:&SELinuxOptions{User:,Role:,Type:,Level:s0:c48,c32,},RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:02:18 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:02:20 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:02:20 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:02:18 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:172.30.237.119,StartTime:2020-09-21 16:02:18 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:p,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:02:20 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:gcr.io/kubernetes-e2e-test-images/agnhost@sha256:02b9e10ea1f7439a03d9bc3ee16d984172ca7eac1818792ad4d721c5c8f72ff4,ContainerID:cri-o://82257164c1c8f3906aeb22f0d013dcba50a6b66a59877ba3521603cef76fda11,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.237.119,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - -STEP: checking for scheduler event about the pod -Sep 21 16:02:22.949: INFO: Saw scheduler event for our pod. -STEP: checking for kubelet event about the pod -Sep 21 16:02:24.972: INFO: Saw kubelet event for our pod. -STEP: deleting the pod -[AfterEach] [k8s.io] [sig-node] Events +STEP: Creating secret with name projected-secret-test-4df6cbc8-4e52-41b5-86ea-33b6f2bc3fa0 +STEP: Creating a pod to test consume secrets +Oct 26 16:19:41.503: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-861ca831-2830-4e1f-893d-59a31d87db56" in namespace "projected-825" to be "success or failure" +Oct 26 16:19:41.518: INFO: Pod "pod-projected-secrets-861ca831-2830-4e1f-893d-59a31d87db56": Phase="Pending", Reason="", readiness=false. Elapsed: 15.176869ms +Oct 26 16:19:43.544: INFO: Pod "pod-projected-secrets-861ca831-2830-4e1f-893d-59a31d87db56": Phase="Pending", Reason="", readiness=false. Elapsed: 2.041172549s +Oct 26 16:19:45.561: INFO: Pod "pod-projected-secrets-861ca831-2830-4e1f-893d-59a31d87db56": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.057277294s +STEP: Saw pod success +Oct 26 16:19:45.561: INFO: Pod "pod-projected-secrets-861ca831-2830-4e1f-893d-59a31d87db56" satisfied condition "success or failure" +Oct 26 16:19:45.576: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-secrets-861ca831-2830-4e1f-893d-59a31d87db56 container secret-volume-test: +STEP: delete the pod +Oct 26 16:19:45.699: INFO: Waiting for pod pod-projected-secrets-861ca831-2830-4e1f-893d-59a31d87db56 to disappear +Oct 26 16:19:45.714: INFO: Pod pod-projected-secrets-861ca831-2830-4e1f-893d-59a31d87db56 no longer exists +[AfterEach] [sig-storage] Projected secret /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:02:25.256: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "events-9539" for this suite. - -• [SLOW TEST:7.272 seconds] -[k8s.io] [sig-node] Events -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - should be sent by kubelets and the scheduler about pods scheduling and running [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [k8s.io] [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running [Conformance]","total":280,"completed":54,"skipped":994,"failed":0} -SSSSS +Oct 26 16:19:45.714: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-825" for this suite. +•{"msg":"PASSED [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":280,"completed":54,"skipped":815,"failed":0} +SSSSSSSSSSSSSS ------------------------------ -[sig-apps] Job - should delete a job [Conformance] +[sig-node] Downward API + should provide host IP as an env var [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-apps] Job +[BeforeEach] [sig-node] Downward API /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:02:25.378: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename job +Oct 26 16:19:45.753: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename downward-api STEP: Waiting for a default service account to be provisioned in namespace -[It] should delete a job [Conformance] +[It] should provide host IP as an env var [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a job -STEP: Ensuring active pods == parallelism -STEP: delete a job -STEP: deleting Job.batch foo in namespace job-348, will wait for the garbage collector to delete the pods -Sep 21 16:02:29.835: INFO: Deleting Job.batch foo took: 67.806707ms -Sep 21 16:02:29.935: INFO: Terminating Job.batch foo pods took: 100.410333ms -STEP: Ensuring job was deleted -[AfterEach] [sig-apps] Job - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:03:09.866: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "job-348" for this suite. - -• [SLOW TEST:44.588 seconds] -[sig-apps] Job -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 - should delete a job [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-apps] Job should delete a job [Conformance]","total":280,"completed":55,"skipped":999,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-network] Networking Granular Checks: Pods - should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-network] Networking - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 16:03:09.966: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename pod-network-test -STEP: Waiting for a default service account to be provisioned in namespace -[It] should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Performing setup for networking test in namespace pod-network-test-3492 -STEP: creating a selector -STEP: Creating the service pods in kubernetes -Sep 21 16:03:10.580: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable -STEP: Creating test pods -Sep 21 16:03:32.170: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.237.122:8080/dial?request=hostname&protocol=http&host=172.30.237.121&port=8080&tries=1'] Namespace:pod-network-test-3492 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -Sep 21 16:03:32.170: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -Sep 21 16:03:32.924: INFO: Waiting for responses: map[] -Sep 21 16:03:32.950: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.237.122:8080/dial?request=hostname&protocol=http&host=172.30.139.38&port=8080&tries=1'] Namespace:pod-network-test-3492 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -Sep 21 16:03:32.950: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -Sep 21 16:03:34.106: INFO: Waiting for responses: map[] -Sep 21 16:03:34.130: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.237.122:8080/dial?request=hostname&protocol=http&host=172.30.81.32&port=8080&tries=1'] Namespace:pod-network-test-3492 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} -Sep 21 16:03:34.130: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -Sep 21 16:03:34.443: INFO: Waiting for responses: map[] -[AfterEach] [sig-network] Networking +STEP: Creating a pod to test downward api env vars +Oct 26 16:19:45.948: INFO: Waiting up to 5m0s for pod "downward-api-a0b26b16-4a0a-474f-a318-de8d198a616d" in namespace "downward-api-9373" to be "success or failure" +Oct 26 16:19:45.964: INFO: Pod "downward-api-a0b26b16-4a0a-474f-a318-de8d198a616d": Phase="Pending", Reason="", readiness=false. Elapsed: 16.082318ms +Oct 26 16:19:47.980: INFO: Pod "downward-api-a0b26b16-4a0a-474f-a318-de8d198a616d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031977583s +STEP: Saw pod success +Oct 26 16:19:47.980: INFO: Pod "downward-api-a0b26b16-4a0a-474f-a318-de8d198a616d" satisfied condition "success or failure" +Oct 26 16:19:47.997: INFO: Trying to get logs from node 10.72.119.74 pod downward-api-a0b26b16-4a0a-474f-a318-de8d198a616d container dapi-container: +STEP: delete the pod +Oct 26 16:19:48.135: INFO: Waiting for pod downward-api-a0b26b16-4a0a-474f-a318-de8d198a616d to disappear +Oct 26 16:19:48.156: INFO: Pod downward-api-a0b26b16-4a0a-474f-a318-de8d198a616d no longer exists +[AfterEach] [sig-node] Downward API /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:03:34.444: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pod-network-test-3492" for this suite. - -• [SLOW TEST:24.546 seconds] -[sig-network] Networking -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26 - Granular Checks: Pods - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29 - should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":56,"skipped":1023,"failed":0} -SSSSSSSSSS +Oct 26 16:19:48.157: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-9373" for this suite. +•{"msg":"PASSED [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance]","total":280,"completed":55,"skipped":829,"failed":0} +SSSSSSSSSSSSSS ------------------------------ [sig-storage] EmptyDir volumes - should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + pod should support shared volumes between containers [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 [BeforeEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:03:34.515: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:19:48.203: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename emptydir STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] +[It] pod should support shared volumes between containers [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test emptydir 0644 on tmpfs -Sep 21 16:03:35.093: INFO: Waiting up to 5m0s for pod "pod-c0c6797d-46be-4292-b1f5-f2138c168c74" in namespace "emptydir-9956" to be "success or failure" -Sep 21 16:03:35.120: INFO: Pod "pod-c0c6797d-46be-4292-b1f5-f2138c168c74": Phase="Pending", Reason="", readiness=false. Elapsed: 27.329514ms -Sep 21 16:03:37.158: INFO: Pod "pod-c0c6797d-46be-4292-b1f5-f2138c168c74": Phase="Pending", Reason="", readiness=false. Elapsed: 2.065010494s -Sep 21 16:03:39.204: INFO: Pod "pod-c0c6797d-46be-4292-b1f5-f2138c168c74": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.111547547s -STEP: Saw pod success -Sep 21 16:03:39.243: INFO: Pod "pod-c0c6797d-46be-4292-b1f5-f2138c168c74" satisfied condition "success or failure" -Sep 21 16:03:39.294: INFO: Trying to get logs from node 10.241.51.150 pod pod-c0c6797d-46be-4292-b1f5-f2138c168c74 container test-container: -STEP: delete the pod -Sep 21 16:03:39.565: INFO: Waiting for pod pod-c0c6797d-46be-4292-b1f5-f2138c168c74 to disappear -Sep 21 16:03:39.607: INFO: Pod pod-c0c6797d-46be-4292-b1f5-f2138c168c74 no longer exists +STEP: Creating Pod +STEP: Waiting for the pod running +STEP: Geting the pod +STEP: Reading file content from the nginx-container +Oct 26 16:19:53.531: INFO: ExecWithOptions {Command:[/bin/sh -c cat /usr/share/volumeshare/shareddata.txt] Namespace:emptydir-9555 PodName:pod-sharedvolume-cedd745b-7faf-4ec7-a6c4-55df5466e857 ContainerName:busybox-main-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} +Oct 26 16:19:53.531: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +Oct 26 16:19:53.780: INFO: Exec stderr: "" [AfterEach] [sig-storage] EmptyDir volumes /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:03:39.607: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-9956" for this suite. +Oct 26 16:19:53.780: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-9555" for this suite. -• [SLOW TEST:5.229 seconds] +• [SLOW TEST:5.613 seconds] [sig-storage] EmptyDir volumes /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 - should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + pod should support shared volumes between containers [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":57,"skipped":1033,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-storage] Projected combined - should project all components that make up the projection API [Projection][NodeConformance] [Conformance] +{"msg":"PASSED [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance]","total":280,"completed":56,"skipped":843,"failed":0} +[k8s.io] Pods + should be submitted and removed [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] Projected combined +[BeforeEach] [k8s.io] Pods /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:03:39.745: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename projected +Oct 26 16:19:53.817: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename pods STEP: Waiting for a default service account to be provisioned in namespace -[It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] +[BeforeEach] [k8s.io] Pods + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177 +[It] should be submitted and removed [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating configMap with name configmap-projected-all-test-volume-e9676904-e728-4c6d-94d5-bd33cab27890 -STEP: Creating secret with name secret-projected-all-test-volume-4a719e07-2722-4868-aece-7a2267d39066 -STEP: Creating a pod to test Check all projections for projected volume plugin -Sep 21 16:03:40.340: INFO: Waiting up to 5m0s for pod "projected-volume-be09acbe-1ba8-4e4a-b7ee-acf209860f32" in namespace "projected-4131" to be "success or failure" -Sep 21 16:03:40.368: INFO: Pod "projected-volume-be09acbe-1ba8-4e4a-b7ee-acf209860f32": Phase="Pending", Reason="", readiness=false. Elapsed: 27.605968ms -Sep 21 16:03:42.413: INFO: Pod "projected-volume-be09acbe-1ba8-4e4a-b7ee-acf209860f32": Phase="Pending", Reason="", readiness=false. Elapsed: 2.072582762s -Sep 21 16:03:44.441: INFO: Pod "projected-volume-be09acbe-1ba8-4e4a-b7ee-acf209860f32": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.101400143s -STEP: Saw pod success -Sep 21 16:03:44.442: INFO: Pod "projected-volume-be09acbe-1ba8-4e4a-b7ee-acf209860f32" satisfied condition "success or failure" -Sep 21 16:03:44.475: INFO: Trying to get logs from node 10.241.51.150 pod projected-volume-be09acbe-1ba8-4e4a-b7ee-acf209860f32 container projected-all-volume-test: -STEP: delete the pod -Sep 21 16:03:45.124: INFO: Waiting for pod projected-volume-be09acbe-1ba8-4e4a-b7ee-acf209860f32 to disappear -Sep 21 16:03:45.158: INFO: Pod projected-volume-be09acbe-1ba8-4e4a-b7ee-acf209860f32 no longer exists -[AfterEach] [sig-storage] Projected combined +STEP: creating the pod +STEP: setting up watch +STEP: submitting the pod to kubernetes +Oct 26 16:19:53.978: INFO: observed the pod list +STEP: verifying the pod is in kubernetes +STEP: verifying pod creation was observed +STEP: deleting the pod gracefully +STEP: verifying the kubelet observed the termination notice +STEP: verifying pod deletion was observed +[AfterEach] [k8s.io] Pods /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:03:45.158: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-4131" for this suite. +Oct 26 16:20:06.693: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-8062" for this suite. -• [SLOW TEST:6.117 seconds] -[sig-storage] Projected combined -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_combined.go:31 - should project all components that make up the projection API [Projection][NodeConformance] [Conformance] +• [SLOW TEST:12.935 seconds] +[k8s.io] Pods +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 + should be submitted and removed [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance]","total":280,"completed":58,"skipped":1072,"failed":0} -SSSSSSS +{"msg":"PASSED [k8s.io] Pods should be submitted and removed [NodeConformance] [Conformance]","total":280,"completed":57,"skipped":843,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + should mutate custom resource [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] EmptyDir volumes +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:03:45.862: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename emptydir +Oct 26 16:20:06.752: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename webhook STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 +STEP: Setting up server cert +STEP: Create role binding to let webhook read extension-apiserver-authentication +STEP: Deploying the webhook pod +STEP: Wait for the deployment to be ready +Oct 26 16:20:07.461: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +Oct 26 16:20:09.510: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326007, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326007, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326007, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326007, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service +STEP: Verifying the service has paired with the endpoint +Oct 26 16:20:12.582: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should mutate custom resource [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test emptydir 0644 on node default medium -Sep 21 16:03:46.313: INFO: Waiting up to 5m0s for pod "pod-5fcb325c-723e-4e93-b8a3-ddc5cc3fc216" in namespace "emptydir-7205" to be "success or failure" -Sep 21 16:03:46.347: INFO: Pod "pod-5fcb325c-723e-4e93-b8a3-ddc5cc3fc216": Phase="Pending", Reason="", readiness=false. Elapsed: 33.161668ms -Sep 21 16:03:48.394: INFO: Pod "pod-5fcb325c-723e-4e93-b8a3-ddc5cc3fc216": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.080088366s -STEP: Saw pod success -Sep 21 16:03:48.394: INFO: Pod "pod-5fcb325c-723e-4e93-b8a3-ddc5cc3fc216" satisfied condition "success or failure" -Sep 21 16:03:48.440: INFO: Trying to get logs from node 10.241.51.147 pod pod-5fcb325c-723e-4e93-b8a3-ddc5cc3fc216 container test-container: -STEP: delete the pod -Sep 21 16:03:48.682: INFO: Waiting for pod pod-5fcb325c-723e-4e93-b8a3-ddc5cc3fc216 to disappear -Sep 21 16:03:48.708: INFO: Pod pod-5fcb325c-723e-4e93-b8a3-ddc5cc3fc216 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +Oct 26 16:20:12.597: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Registering the mutating webhook for custom resource e2e-test-webhook-5088-crds.webhook.example.com via the AdmissionRegistration API +STEP: Creating a custom resource that should be mutated by the webhook +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:03:48.708: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-7205" for this suite. -•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":59,"skipped":1079,"failed":0} -S +Oct 26 16:20:13.894: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-9476" for this suite. +STEP: Destroying namespace "webhook-9476-markers" for this suite. +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 + +• [SLOW TEST:7.397 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 + should mutate custom resource [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - should have a working scale subresource [Conformance] +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance]","total":280,"completed":58,"skipped":879,"failed":0} +SSSSSSSSSS +------------------------------ +[sig-apps] ReplicationController + should serve a basic image on each replica with a public image [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-apps] StatefulSet +[BeforeEach] [sig-apps] ReplicationController /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:03:48.818: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename statefulset +Oct 26 16:20:14.149: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename replication-controller STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] StatefulSet - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64 -[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79 -STEP: Creating service test in namespace statefulset-1605 -[It] should have a working scale subresource [Conformance] +[It] should serve a basic image on each replica with a public image [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating statefulset ss in namespace statefulset-1605 -Sep 21 16:03:49.115: INFO: Found 0 stateful pods, waiting for 1 -Sep 21 16:03:59.144: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true -STEP: getting scale subresource -STEP: updating a scale subresource -STEP: verifying the statefulset Spec.Replicas was modified -[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 -Sep 21 16:03:59.365: INFO: Deleting all statefulset in ns statefulset-1605 -Sep 21 16:03:59.411: INFO: Scaling statefulset ss to 0 -Sep 21 16:04:19.523: INFO: Waiting for statefulset status.replicas updated to 0 -Sep 21 16:04:19.574: INFO: Deleting statefulset ss -[AfterEach] [sig-apps] StatefulSet +STEP: Creating replication controller my-hostname-basic-6902f5e8-6e79-430f-9f46-918861cbadfe +Oct 26 16:20:14.330: INFO: Pod name my-hostname-basic-6902f5e8-6e79-430f-9f46-918861cbadfe: Found 0 pods out of 1 +Oct 26 16:20:19.356: INFO: Pod name my-hostname-basic-6902f5e8-6e79-430f-9f46-918861cbadfe: Found 1 pods out of 1 +Oct 26 16:20:19.357: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-6902f5e8-6e79-430f-9f46-918861cbadfe" are running +Oct 26 16:20:19.380: INFO: Pod "my-hostname-basic-6902f5e8-6e79-430f-9f46-918861cbadfe-lrzwv" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 16:20:14 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 16:20:15 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 16:20:15 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 16:20:14 +0000 UTC Reason: Message:}]) +Oct 26 16:20:19.380: INFO: Trying to dial the pod +Oct 26 16:20:24.444: INFO: Controller my-hostname-basic-6902f5e8-6e79-430f-9f46-918861cbadfe: Got expected result from replica 1 [my-hostname-basic-6902f5e8-6e79-430f-9f46-918861cbadfe-lrzwv]: "my-hostname-basic-6902f5e8-6e79-430f-9f46-918861cbadfe-lrzwv", 1 of 1 required successes so far +[AfterEach] [sig-apps] ReplicationController /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:04:19.866: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "statefulset-1605" for this suite. +Oct 26 16:20:24.444: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replication-controller-1496" for this suite. -• [SLOW TEST:31.114 seconds] -[sig-apps] StatefulSet +• [SLOW TEST:10.331 seconds] +[sig-apps] ReplicationController /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 - [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - should have a working scale subresource [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 + should serve a basic image on each replica with a public image [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance]","total":280,"completed":60,"skipped":1080,"failed":0} -SS +{"msg":"PASSED [sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance]","total":280,"completed":59,"skipped":889,"failed":0} +SSSSSSSSS ------------------------------ -[sig-network] DNS - should provide DNS for services [Conformance] +[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod + should be possible to delete [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-network] DNS +[BeforeEach] [k8s.io] Kubelet /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:04:19.932: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename dns +Oct 26 16:20:24.480: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename kubelet-test STEP: Waiting for a default service account to be provisioned in namespace -[It] should provide DNS for services [Conformance] +[BeforeEach] [k8s.io] Kubelet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 +[BeforeEach] when scheduling a busybox command that always fails in a pod + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81 +[It] should be possible to delete [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a test headless service -STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-7379.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-7379.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-7379.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-7379.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-7379.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-7379.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-7379.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-7379.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-7379.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-7379.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-7379.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-7379.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-7379.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 48.43.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.43.48_udp@PTR;check="$$(dig +tcp +noall +answer +search 48.43.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.43.48_tcp@PTR;sleep 1; done - -STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-7379.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-7379.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-7379.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-7379.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-7379.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-7379.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-7379.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-7379.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-7379.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-7379.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-7379.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-7379.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-7379.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 48.43.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.43.48_udp@PTR;check="$$(dig +tcp +noall +answer +search 48.43.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.43.48_tcp@PTR;sleep 1; done - -STEP: creating a pod to probe DNS -STEP: submitting the pod to kubernetes -STEP: retrieving the pod -STEP: looking for the results for each expected name from probers -Sep 21 16:04:24.930: INFO: Unable to read wheezy_tcp@dns-test-service.dns-7379.svc.cluster.local from pod dns-7379/dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53: the server could not find the requested resource (get pods dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53) -Sep 21 16:04:24.983: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-7379.svc.cluster.local from pod dns-7379/dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53: the server could not find the requested resource (get pods dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53) -Sep 21 16:04:25.066: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-7379.svc.cluster.local from pod dns-7379/dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53: the server could not find the requested resource (get pods dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53) -Sep 21 16:04:25.570: INFO: Unable to read jessie_udp@dns-test-service.dns-7379.svc.cluster.local from pod dns-7379/dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53: the server could not find the requested resource (get pods dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53) -Sep 21 16:04:25.672: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-7379.svc.cluster.local from pod dns-7379/dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53: the server could not find the requested resource (get pods dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53) -Sep 21 16:04:25.730: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-7379.svc.cluster.local from pod dns-7379/dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53: the server could not find the requested resource (get pods dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53) -Sep 21 16:04:26.968: INFO: Lookups using dns-7379/dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53 failed for: [wheezy_tcp@dns-test-service.dns-7379.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-7379.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-7379.svc.cluster.local jessie_udp@dns-test-service.dns-7379.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-7379.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-7379.svc.cluster.local] - -Sep 21 16:04:33.787: INFO: DNS probes using dns-7379/dns-test-4c7135a5-a888-4a3a-8159-039ddf558f53 succeeded - -STEP: deleting the pod -STEP: deleting the test service -STEP: deleting the test headless service -[AfterEach] [sig-network] DNS +[AfterEach] [k8s.io] Kubelet /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:04:34.714: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "dns-7379" for this suite. - -• [SLOW TEST:15.042 seconds] -[sig-network] DNS -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 - should provide DNS for services [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +Oct 26 16:20:24.760: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubelet-test-7630" for this suite. +•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance]","total":280,"completed":60,"skipped":898,"failed":0} +SSSSSSSSSSSSSSS ------------------------------ -{"msg":"PASSED [sig-network] DNS should provide DNS for services [Conformance]","total":280,"completed":61,"skipped":1082,"failed":0} -SSS +[sig-storage] Projected downwardAPI + should update annotations on modification [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:20:24.796: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename projected +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40 +[It] should update annotations on modification [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating the pod +Oct 26 16:20:27.660: INFO: Successfully updated pod "annotationupdateeeaa07a9-dd5c-40cc-baa1-461a4838a5b7" +[AfterEach] [sig-storage] Projected downwardAPI + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:20:29.728: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-4249" for this suite. +•{"msg":"PASSED [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance]","total":280,"completed":61,"skipped":913,"failed":0} +SSSSSSSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] +[sig-storage] Projected secret + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] EmptyDir volumes +[BeforeEach] [sig-storage] Projected secret /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:04:34.974: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename emptydir +Oct 26 16:20:29.761: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace -[It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] +[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test emptydir 0666 on node default medium -Sep 21 16:04:35.867: INFO: Waiting up to 5m0s for pod "pod-945a2295-6cb4-4c2d-9443-a7ff4b6a76c5" in namespace "emptydir-5226" to be "success or failure" -Sep 21 16:04:35.897: INFO: Pod "pod-945a2295-6cb4-4c2d-9443-a7ff4b6a76c5": Phase="Pending", Reason="", readiness=false. Elapsed: 30.308054ms -Sep 21 16:04:37.947: INFO: Pod "pod-945a2295-6cb4-4c2d-9443-a7ff4b6a76c5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.080068969s -Sep 21 16:04:39.989: INFO: Pod "pod-945a2295-6cb4-4c2d-9443-a7ff4b6a76c5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.122192809s +STEP: Creating projection with secret that has name projected-secret-test-2ec17e57-c03c-4b48-9a02-0a7f88614c58 +STEP: Creating a pod to test consume secrets +Oct 26 16:20:29.971: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-4a48472b-b82c-41d5-8072-5973cf8f4a82" in namespace "projected-7542" to be "success or failure" +Oct 26 16:20:29.987: INFO: Pod "pod-projected-secrets-4a48472b-b82c-41d5-8072-5973cf8f4a82": Phase="Pending", Reason="", readiness=false. Elapsed: 15.903929ms +Oct 26 16:20:32.003: INFO: Pod "pod-projected-secrets-4a48472b-b82c-41d5-8072-5973cf8f4a82": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031964122s +Oct 26 16:20:34.035: INFO: Pod "pod-projected-secrets-4a48472b-b82c-41d5-8072-5973cf8f4a82": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.06362788s STEP: Saw pod success -Sep 21 16:04:39.989: INFO: Pod "pod-945a2295-6cb4-4c2d-9443-a7ff4b6a76c5" satisfied condition "success or failure" -Sep 21 16:04:40.025: INFO: Trying to get logs from node 10.241.51.150 pod pod-945a2295-6cb4-4c2d-9443-a7ff4b6a76c5 container test-container: +Oct 26 16:20:34.035: INFO: Pod "pod-projected-secrets-4a48472b-b82c-41d5-8072-5973cf8f4a82" satisfied condition "success or failure" +Oct 26 16:20:34.050: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-secrets-4a48472b-b82c-41d5-8072-5973cf8f4a82 container projected-secret-volume-test: STEP: delete the pod -Sep 21 16:04:40.864: INFO: Waiting for pod pod-945a2295-6cb4-4c2d-9443-a7ff4b6a76c5 to disappear -Sep 21 16:04:40.890: INFO: Pod pod-945a2295-6cb4-4c2d-9443-a7ff4b6a76c5 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +Oct 26 16:20:34.129: INFO: Waiting for pod pod-projected-secrets-4a48472b-b82c-41d5-8072-5973cf8f4a82 to disappear +Oct 26 16:20:34.145: INFO: Pod pod-projected-secrets-4a48472b-b82c-41d5-8072-5973cf8f4a82 no longer exists +[AfterEach] [sig-storage] Projected secret + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:20:34.145: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-7542" for this suite. +•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":62,"skipped":928,"failed":0} +SSSSSSS +------------------------------ +[sig-apps] Daemon set [Serial] + should retry creating failed daemon pods [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:20:34.183: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename daemonsets +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133 +[It] should retry creating failed daemon pods [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating a simple DaemonSet "daemon-set" +STEP: Check that daemon pods launch on every node of the cluster. +Oct 26 16:20:34.504: INFO: Number of nodes with available pods: 0 +Oct 26 16:20:34.504: INFO: Node 10.72.119.74 is running more than one daemon pod +Oct 26 16:20:35.543: INFO: Number of nodes with available pods: 0 +Oct 26 16:20:35.543: INFO: Node 10.72.119.74 is running more than one daemon pod +Oct 26 16:20:36.536: INFO: Number of nodes with available pods: 1 +Oct 26 16:20:36.536: INFO: Node 10.72.119.76 is running more than one daemon pod +Oct 26 16:20:37.543: INFO: Number of nodes with available pods: 3 +Oct 26 16:20:37.543: INFO: Number of running nodes: 3, number of available pods: 3 +STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived. +Oct 26 16:20:37.648: INFO: Number of nodes with available pods: 2 +Oct 26 16:20:37.648: INFO: Node 10.72.119.76 is running more than one daemon pod +Oct 26 16:20:38.682: INFO: Number of nodes with available pods: 2 +Oct 26 16:20:38.682: INFO: Node 10.72.119.76 is running more than one daemon pod +Oct 26 16:20:39.683: INFO: Number of nodes with available pods: 2 +Oct 26 16:20:39.683: INFO: Node 10.72.119.76 is running more than one daemon pod +Oct 26 16:20:40.687: INFO: Number of nodes with available pods: 3 +Oct 26 16:20:40.687: INFO: Number of running nodes: 3, number of available pods: 3 +STEP: Wait for the failed daemon pod to be completely deleted. +[AfterEach] [sig-apps] Daemon set [Serial] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99 +STEP: Deleting DaemonSet "daemon-set" +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6328, will wait for the garbage collector to delete the pods +Oct 26 16:20:40.820: INFO: Deleting DaemonSet.extensions daemon-set took: 39.003825ms +Oct 26 16:20:40.920: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.306547ms +Oct 26 16:20:49.535: INFO: Number of nodes with available pods: 0 +Oct 26 16:20:49.535: INFO: Number of running nodes: 0, number of available pods: 0 +Oct 26 16:20:49.550: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-6328/daemonsets","resourceVersion":"56107"},"items":null} + +Oct 26 16:20:49.566: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-6328/pods","resourceVersion":"56107"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:04:40.891: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-5226" for this suite. +Oct 26 16:20:49.624: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-6328" for this suite. -• [SLOW TEST:6.009 seconds] -[sig-storage] EmptyDir volumes -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/empty_dir.go:40 - should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] +• [SLOW TEST:15.482 seconds] +[sig-apps] Daemon set [Serial] +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 + should retry creating failed daemon pods [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":62,"skipped":1085,"failed":0} -S +{"msg":"PASSED [sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance]","total":280,"completed":63,"skipped":935,"failed":0} +SSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Garbage collector - should orphan pods created by rc if delete options say so [Conformance] +[k8s.io] Security Context When creating a pod with readOnlyRootFilesystem + should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] Garbage collector +[BeforeEach] [k8s.io] Security Context /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:04:40.985: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename gc +Oct 26 16:20:49.666: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename security-context-test STEP: Waiting for a default service account to be provisioned in namespace -[It] should orphan pods created by rc if delete options say so [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: create the rc -STEP: delete the rc -STEP: wait for the rc to be deleted -STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods -STEP: Gathering metrics -W0921 16:05:21.612452 24 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. -Sep 21 16:05:21.612: INFO: For apiserver_request_total: -For apiserver_request_latency_seconds: -For apiserver_init_events_total: -For garbage_collector_attempt_to_delete_queue_latency: -For garbage_collector_attempt_to_delete_work_duration: -For garbage_collector_attempt_to_orphan_queue_latency: -For garbage_collector_attempt_to_orphan_work_duration: -For garbage_collector_dirty_processing_latency_microseconds: -For garbage_collector_event_processing_latency_microseconds: -For garbage_collector_graph_changes_queue_latency: -For garbage_collector_graph_changes_work_duration: -For garbage_collector_orphan_processing_latency_microseconds: -For namespace_queue_latency: -For namespace_queue_latency_sum: -For namespace_queue_latency_count: -For namespace_retries: -For namespace_work_duration: -For namespace_work_duration_sum: -For namespace_work_duration_count: -For function_duration_seconds: -For errors_total: -For evicted_pods_total: - -[AfterEach] [sig-api-machinery] Garbage collector - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:05:21.612: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "gc-9408" for this suite. - -• [SLOW TEST:40.734 seconds] -[sig-api-machinery] Garbage collector -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - should orphan pods created by rc if delete options say so [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance]","total":280,"completed":63,"skipped":1086,"failed":0} -SS ------------------------------- -[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook - should execute prestop http hook properly [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Container Lifecycle Hook - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 16:05:21.720: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename container-lifecycle-hook -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] when create a pod with lifecycle hook - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64 -STEP: create the container to handle the HTTPGet hook request. -[It] should execute prestop http hook properly [NodeConformance] [Conformance] +[BeforeEach] [k8s.io] Security Context + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39 +[It] should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: create the pod with lifecycle hook -STEP: delete the pod with lifecycle hook -Sep 21 16:05:31.491: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -Sep 21 16:05:31.940: INFO: Pod pod-with-prestop-http-hook still exists -Sep 21 16:05:33.941: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -Sep 21 16:05:33.972: INFO: Pod pod-with-prestop-http-hook still exists -Sep 21 16:05:35.941: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -Sep 21 16:05:35.977: INFO: Pod pod-with-prestop-http-hook still exists -Sep 21 16:05:37.941: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -Sep 21 16:05:37.972: INFO: Pod pod-with-prestop-http-hook still exists -Sep 21 16:05:39.941: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -Sep 21 16:05:39.976: INFO: Pod pod-with-prestop-http-hook no longer exists -STEP: check prestop hook -[AfterEach] [k8s.io] Container Lifecycle Hook +Oct 26 16:20:49.881: INFO: Waiting up to 5m0s for pod "busybox-readonly-false-8b625372-08e1-4af8-9517-4c14be18324c" in namespace "security-context-test-8728" to be "success or failure" +Oct 26 16:20:49.906: INFO: Pod "busybox-readonly-false-8b625372-08e1-4af8-9517-4c14be18324c": Phase="Pending", Reason="", readiness=false. Elapsed: 25.157635ms +Oct 26 16:20:51.928: INFO: Pod "busybox-readonly-false-8b625372-08e1-4af8-9517-4c14be18324c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.047256552s +Oct 26 16:20:53.944: INFO: Pod "busybox-readonly-false-8b625372-08e1-4af8-9517-4c14be18324c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.063278331s +Oct 26 16:20:53.944: INFO: Pod "busybox-readonly-false-8b625372-08e1-4af8-9517-4c14be18324c" satisfied condition "success or failure" +[AfterEach] [k8s.io] Security Context /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:05:40.225: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-lifecycle-hook-9013" for this suite. - -• [SLOW TEST:18.889 seconds] -[k8s.io] Container Lifecycle Hook -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - when create a pod with lifecycle hook - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42 - should execute prestop http hook properly [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance]","total":280,"completed":64,"skipped":1088,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSS +Oct 26 16:20:53.944: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "security-context-test-8728" for this suite. +•{"msg":"PASSED [k8s.io] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]","total":280,"completed":64,"skipped":948,"failed":0} +SSSSSSSSS ------------------------------ -[k8s.io] Pods - should contain environment variables for services [NodeConformance] [Conformance] +[sig-storage] Secrets + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Pods +[BeforeEach] [sig-storage] Secrets /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:05:40.614: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename pods +Oct 26 16:20:53.991: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename secrets STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177 -[It] should contain environment variables for services [NodeConformance] [Conformance] +[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 16:05:44.394: INFO: Waiting up to 5m0s for pod "client-envvars-c6da27a8-10af-4d7a-96db-70cfb91c407f" in namespace "pods-1774" to be "success or failure" -Sep 21 16:05:44.423: INFO: Pod "client-envvars-c6da27a8-10af-4d7a-96db-70cfb91c407f": Phase="Pending", Reason="", readiness=false. Elapsed: 24.263895ms -Sep 21 16:05:46.454: INFO: Pod "client-envvars-c6da27a8-10af-4d7a-96db-70cfb91c407f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.054903691s -Sep 21 16:05:48.505: INFO: Pod "client-envvars-c6da27a8-10af-4d7a-96db-70cfb91c407f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.106332905s +STEP: Creating secret with name secret-test-6236c52b-5a85-4082-a547-b4b807ec2990 +STEP: Creating a pod to test consume secrets +Oct 26 16:20:54.207: INFO: Waiting up to 5m0s for pod "pod-secrets-9ab107d5-a9ce-4d7e-a410-09ae3ef08534" in namespace "secrets-445" to be "success or failure" +Oct 26 16:20:54.223: INFO: Pod "pod-secrets-9ab107d5-a9ce-4d7e-a410-09ae3ef08534": Phase="Pending", Reason="", readiness=false. Elapsed: 16.173888ms +Oct 26 16:20:56.238: INFO: Pod "pod-secrets-9ab107d5-a9ce-4d7e-a410-09ae3ef08534": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031368435s STEP: Saw pod success -Sep 21 16:05:48.505: INFO: Pod "client-envvars-c6da27a8-10af-4d7a-96db-70cfb91c407f" satisfied condition "success or failure" -Sep 21 16:05:48.534: INFO: Trying to get logs from node 10.241.51.147 pod client-envvars-c6da27a8-10af-4d7a-96db-70cfb91c407f container env3cont: +Oct 26 16:20:56.238: INFO: Pod "pod-secrets-9ab107d5-a9ce-4d7e-a410-09ae3ef08534" satisfied condition "success or failure" +Oct 26 16:20:56.259: INFO: Trying to get logs from node 10.72.119.74 pod pod-secrets-9ab107d5-a9ce-4d7e-a410-09ae3ef08534 container secret-volume-test: STEP: delete the pod -Sep 21 16:05:48.770: INFO: Waiting for pod client-envvars-c6da27a8-10af-4d7a-96db-70cfb91c407f to disappear -Sep 21 16:05:48.798: INFO: Pod client-envvars-c6da27a8-10af-4d7a-96db-70cfb91c407f no longer exists -[AfterEach] [k8s.io] Pods +Oct 26 16:20:56.350: INFO: Waiting for pod pod-secrets-9ab107d5-a9ce-4d7e-a410-09ae3ef08534 to disappear +Oct 26 16:20:56.366: INFO: Pod pod-secrets-9ab107d5-a9ce-4d7e-a410-09ae3ef08534 no longer exists +[AfterEach] [sig-storage] Secrets /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:05:48.798: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-1774" for this suite. - -• [SLOW TEST:8.297 seconds] -[k8s.io] Pods -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - should contain environment variables for services [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [k8s.io] Pods should contain environment variables for services [NodeConformance] [Conformance]","total":280,"completed":65,"skipped":1115,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +Oct 26 16:20:56.366: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-445" for this suite. +•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":65,"skipped":957,"failed":0} +SSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] ConfigMap - should be consumable via environment variable [NodeConformance] [Conformance] +[sig-storage] Projected secret + optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-node] ConfigMap +[BeforeEach] [sig-storage] Projected secret /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:05:48.914: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename configmap +Oct 26 16:20:56.410: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename projected STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable via environment variable [NodeConformance] [Conformance] +[It] optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating configMap configmap-3053/configmap-test-db6ab22e-8a43-49dc-8a3a-89326cb13e39 -STEP: Creating a pod to test consume configMaps -Sep 21 16:05:49.586: INFO: Waiting up to 5m0s for pod "pod-configmaps-582bbefe-5636-4109-8a70-393d0be55865" in namespace "configmap-3053" to be "success or failure" -Sep 21 16:05:49.823: INFO: Pod "pod-configmaps-582bbefe-5636-4109-8a70-393d0be55865": Phase="Pending", Reason="", readiness=false. Elapsed: 237.195755ms -Sep 21 16:05:51.859: INFO: Pod "pod-configmaps-582bbefe-5636-4109-8a70-393d0be55865": Phase="Pending", Reason="", readiness=false. Elapsed: 2.272792862s -Sep 21 16:05:53.905: INFO: Pod "pod-configmaps-582bbefe-5636-4109-8a70-393d0be55865": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.319302987s -STEP: Saw pod success -Sep 21 16:05:53.905: INFO: Pod "pod-configmaps-582bbefe-5636-4109-8a70-393d0be55865" satisfied condition "success or failure" -Sep 21 16:05:53.939: INFO: Trying to get logs from node 10.241.51.150 pod pod-configmaps-582bbefe-5636-4109-8a70-393d0be55865 container env-test: -STEP: delete the pod -Sep 21 16:05:54.369: INFO: Waiting for pod pod-configmaps-582bbefe-5636-4109-8a70-393d0be55865 to disappear -Sep 21 16:05:54.412: INFO: Pod pod-configmaps-582bbefe-5636-4109-8a70-393d0be55865 no longer exists -[AfterEach] [sig-node] ConfigMap +Oct 26 16:20:56.633: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node +STEP: Creating secret with name s-test-opt-del-9f22693b-5a58-4516-8200-e617d3bfb5c2 +STEP: Creating secret with name s-test-opt-upd-eb53387e-7450-4da4-a29e-edc2b614d978 +STEP: Creating the pod +STEP: Deleting secret s-test-opt-del-9f22693b-5a58-4516-8200-e617d3bfb5c2 +STEP: Updating secret s-test-opt-upd-eb53387e-7450-4da4-a29e-edc2b614d978 +STEP: Creating secret with name s-test-opt-create-ee377c16-36da-4d19-8ae7-e952727b6a90 +STEP: waiting to observe update in volume +[AfterEach] [sig-storage] Projected secret /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:05:54.412: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-3053" for this suite. +Oct 26 16:22:18.564: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-7686" for this suite. -• [SLOW TEST:5.637 seconds] -[sig-node] ConfigMap -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31 - should be consumable via environment variable [NodeConformance] [Conformance] +• [SLOW TEST:82.187 seconds] +[sig-storage] Projected secret +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 + optional updates should be reflected in volume [NodeConformance] [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance]","total":280,"completed":66,"skipped":1183,"failed":0} -SSSSS +{"msg":"PASSED [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":66,"skipped":975,"failed":0} +SSS ------------------------------ -[sig-scheduling] SchedulerPredicates [Serial] - validates that NodeSelector is respected if matching [Conformance] +[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + custom resource defaulting for requests and from storage works [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] +[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:05:54.552: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename sched-pred +Oct 26 16:22:18.598: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename custom-resource-definition STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86 -Sep 21 16:05:54.874: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready -Sep 21 16:05:55.103: INFO: Waiting for terminating namespaces to be deleted... -Sep 21 16:05:55.140: INFO: -Logging pods the kubelet thinks is on node 10.241.51.147 before test -Sep 21 16:05:55.298: INFO: dns-default-f6vtl from openshift-dns started at 2020-09-21 14:11:49 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container dns ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container dns-node-resolver ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: alertmanager-main-2 from openshift-monitoring started at 2020-09-21 14:13:42 +0000 UTC (3 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container alertmanager ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container alertmanager-proxy ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container config-reloader ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: thanos-querier-8f4c5c746-nh7wt from openshift-monitoring started at 2020-09-21 14:14:55 +0000 UTC (4 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container oauth-proxy ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container prom-label-proxy ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container thanos-querier ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: console-8549dd6c4f-dkvld from openshift-console started at 2020-09-21 14:18:04 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container console ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: sonobuoy from sonobuoy started at 2020-09-21 15:39:44 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container kube-sonobuoy ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: node-exporter-46mwx from openshift-monitoring started at 2020-09-21 14:11:49 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container node-exporter ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: node-ca-6r9nd from openshift-image-registry started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container node-ca ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: tuned-6qvw2 from openshift-cluster-node-tuning-operator started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container tuned ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: calico-typha-6fc9ff8b5f-9pmb4 from calico-system started at 2020-09-21 14:13:14 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container calico-typha ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-shgq6 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container sonobuoy-worker ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container systemd-logs ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: ibmcloud-block-storage-driver-8r6lc from kube-system started at 2020-09-21 14:12:04 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container ibmcloud-block-storage-driver-container ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: test-k8s-e2e-pvg-master-verification from default started at 2020-09-21 14:17:11 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: ibm-keepalived-watcher-h4dbn from kube-system started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container keepalived-watcher ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: calico-node-t2n87 from calico-system started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container calico-node ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: prometheus-operator-8d96bcc68-b2bz9 from openshift-monitoring started at 2020-09-21 14:18:25 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container prometheus-operator ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: packageserver-5ffd66994-7kjtm from openshift-operator-lifecycle-manager started at 2020-09-21 14:19:19 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container packageserver ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: openshift-kube-proxy-tvbvw from openshift-kube-proxy started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container kube-proxy ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: router-default-84dfcdc9b5-pt6zq from openshift-ingress started at 2020-09-21 14:12:50 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container router ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: vpn-64d8d4987c-c89mq from kube-system started at 2020-09-21 14:19:34 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container vpn ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: prometheus-k8s-1 from openshift-monitoring started at 2020-09-21 14:14:54 +0000 UTC (7 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container prom-label-proxy ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container prometheus ready: true, restart count 1 -Sep 21 16:05:55.298: INFO: Container prometheus-config-reloader ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container prometheus-proxy ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container rules-configmap-reloader ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container thanos-sidecar ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: multus-7xt8r from openshift-multus started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container kube-multus ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: multus-admission-controller-l9hmr from openshift-multus started at 2020-09-21 14:12:49 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container multus-admission-controller ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: sonobuoy-e2e-job-36e1517c33ed4f81 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container e2e ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container sonobuoy-worker ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: server-envvars-564be65f-1b7a-42da-80a8-c766df47e514 from pods-1774 started at 2020-09-21 16:05:41 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container srv ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: ibm-master-proxy-static-10.241.51.147 from kube-system started at 2020-09-21 14:11:47 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container ibm-master-proxy-static ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: Container pause ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-t6sqd from ibm-system started at 2020-09-21 14:12:51 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.298: INFO: Container ibm-cloud-provider-ip-169-48-211-85 ready: true, restart count 0 -Sep 21 16:05:55.298: INFO: -Logging pods the kubelet thinks is on node 10.241.51.149 before test -Sep 21 16:05:55.793: INFO: kube-storage-version-migrator-operator-86d8fc6986-vl4fl from openshift-kube-storage-version-migrator-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.793: INFO: Container kube-storage-version-migrator-operator ready: true, restart count 1 -Sep 21 16:05:55.793: INFO: network-operator-f456cfdbb-4m6pj from openshift-network-operator started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.793: INFO: Container network-operator ready: true, restart count 0 -Sep 21 16:05:55.793: INFO: cluster-monitoring-operator-75f4ddd85f-579q6 from openshift-monitoring started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.793: INFO: Container cluster-monitoring-operator ready: true, restart count 0 -Sep 21 16:05:55.793: INFO: prometheus-adapter-5f78ddc679-9hj7w from openshift-monitoring started at 2020-09-21 14:08:28 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.793: INFO: Container prometheus-adapter ready: true, restart count 0 -Sep 21 16:05:55.793: INFO: calico-kube-controllers-79d75767dd-xkvb6 from calico-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.793: INFO: Container calico-kube-controllers ready: true, restart count 0 -Sep 21 16:05:55.793: INFO: multus-admission-controller-gkchr from openshift-multus started at 2020-09-21 14:07:34 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.793: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.793: INFO: Container multus-admission-controller ready: true, restart count 0 -Sep 21 16:05:55.793: INFO: community-operators-6566687cd-smnd2 from openshift-marketplace started at 2020-09-21 14:09:35 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.793: INFO: Container community-operators ready: true, restart count 0 -Sep 21 16:05:55.793: INFO: ibm-storage-watcher-7956f954f-26xx6 from kube-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.793: INFO: Container ibm-storage-watcher-container ready: true, restart count 0 -Sep 21 16:05:55.793: INFO: service-ca-bd8866b9b-qzggv from openshift-service-ca started at 2020-09-21 14:07:50 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.793: INFO: Container service-ca-controller ready: true, restart count 0 -Sep 21 16:05:55.793: INFO: ibmcloud-block-storage-driver-pldbm from kube-system started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.793: INFO: Container ibmcloud-block-storage-driver-container ready: true, restart count 0 -Sep 21 16:05:55.793: INFO: dns-operator-7886f4f4f4-rk7dv from openshift-dns-operator started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.793: INFO: Container dns-operator ready: true, restart count 0 -Sep 21 16:05:55.793: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: ingress-operator-57546dfcd9-2sg5v from openshift-ingress-operator started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container ingress-operator ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: catalog-operator-69d4f5c59f-lqjgz from openshift-operator-lifecycle-manager started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container catalog-operator ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-lgbbw from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container sonobuoy-worker ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: Container systemd-logs ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: openshift-kube-proxy-9svk2 from openshift-kube-proxy started at 2020-09-21 14:06:38 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container kube-proxy ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: service-ca-operator-9f84dfdb5-wdhvg from openshift-service-ca-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container operator ready: true, restart count 1 -Sep 21 16:05:55.794: INFO: router-default-84dfcdc9b5-tk98j from openshift-ingress started at 2020-09-21 14:09:06 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container router ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: node-exporter-jnrzm from openshift-monitoring started at 2020-09-21 14:08:24 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: Container node-exporter ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: dns-default-958xc from openshift-dns started at 2020-09-21 14:09:05 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container dns ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: Container dns-node-resolver ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: alertmanager-main-0 from openshift-monitoring started at 2020-09-21 14:14:12 +0000 UTC (3 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container alertmanager ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: Container alertmanager-proxy ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: Container config-reloader ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: downloads-7989988989-9rhh4 from openshift-console started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container download-server ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: tuned-gpfhj from openshift-cluster-node-tuning-operator started at 2020-09-21 14:08:21 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container tuned ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: downloads-7989988989-cjvjw from openshift-console started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container download-server ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-44pnn from ibm-system started at 2020-09-21 14:12:56 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container ibm-cloud-provider-ip-169-48-211-85 ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: ibm-keepalived-watcher-r4tn7 from kube-system started at 2020-09-21 14:06:04 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container keepalived-watcher ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: marketplace-operator-56c6694c8f-x9rcq from openshift-marketplace started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container marketplace-operator ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: console-operator-58fd84b95c-dc6vq from openshift-console-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container console-operator ready: true, restart count 1 -Sep 21 16:05:55.794: INFO: migrator-6c94c8c775-vxsnp from openshift-kube-storage-version-migrator started at 2020-09-21 14:07:55 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container migrator ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: tigera-operator-679798d94d-hdlvj from tigera-operator started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container tigera-operator ready: true, restart count 2 -Sep 21 16:05:55.794: INFO: cluster-image-registry-operator-695bf78ffc-95l5m from openshift-image-registry started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.794: INFO: Container cluster-image-registry-operator ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: Container cluster-image-registry-operator-watch ready: true, restart count 0 -Sep 21 16:05:55.794: INFO: cluster-storage-operator-5bcf8d9f8f-dls4n from openshift-cluster-storage-operator started at 2020-09-21 14:07:36 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.795: INFO: Container cluster-storage-operator ready: true, restart count 0 -Sep 21 16:05:55.795: INFO: image-registry-775dd777d7-5qtdd from openshift-image-registry started at 2020-09-21 14:11:01 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.795: INFO: Container registry ready: true, restart count 0 -Sep 21 16:05:55.795: INFO: ibm-master-proxy-static-10.241.51.149 from kube-system started at 2020-09-21 14:06:02 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.795: INFO: Container ibm-master-proxy-static ready: true, restart count 0 -Sep 21 16:05:55.795: INFO: Container pause ready: true, restart count 0 -Sep 21 16:05:55.795: INFO: registry-pvc-permissions-rlfxm from openshift-image-registry started at 2020-09-21 14:11:01 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.795: INFO: Container pvc-permissions ready: false, restart count 0 -Sep 21 16:05:55.795: INFO: cluster-node-tuning-operator-bdbf98554-22kv8 from openshift-cluster-node-tuning-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container cluster-node-tuning-operator ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: redhat-operators-54dcdd9697-pcwgz from openshift-marketplace started at 2020-09-21 14:09:34 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container redhat-operators ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: olm-operator-b5f47bf98-8d8cw from openshift-operator-lifecycle-manager started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container olm-operator ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: node-ca-bxmsf from openshift-image-registry started at 2020-09-21 14:09:05 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container node-ca ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: certified-operators-5f495d9886-xjbvn from openshift-marketplace started at 2020-09-21 14:09:34 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container certified-operators ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: calico-node-9gqht from calico-system started at 2020-09-21 14:07:18 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container calico-node ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: openshift-service-catalog-controller-manager-operator-b9ccczlcs from openshift-service-catalog-controller-manager-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container operator ready: true, restart count 1 -Sep 21 16:05:55.796: INFO: cluster-samples-operator-5cfb985bf-n7nxd from openshift-cluster-samples-operator started at 2020-09-21 14:09:54 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container cluster-samples-operator ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: Container cluster-samples-operator-watch ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: ibm-file-plugin-66fdcfc4c9-tm6qk from kube-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container ibm-file-plugin-container ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: openshift-state-metrics-678b6c786f-698gn from openshift-monitoring started at 2020-09-21 14:08:22 +0000 UTC (3 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container kube-rbac-proxy-main ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: Container kube-rbac-proxy-self ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: Container openshift-state-metrics ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: redhat-marketplace-7b9bcdf96b-p4khw from openshift-marketplace started at 2020-09-21 14:09:34 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container redhat-marketplace ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: openshift-service-catalog-apiserver-operator-6988647bc8-t9z7w from openshift-service-catalog-apiserver-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container operator ready: true, restart count 1 -Sep 21 16:05:55.796: INFO: ibmcloud-block-storage-plugin-68d5c65db9-7q8bv from kube-system started at 2020-09-21 14:07:36 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container ibmcloud-block-storage-plugin-container ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: kube-state-metrics-7455b6c8-455w7 from openshift-monitoring started at 2020-09-21 14:08:22 +0000 UTC (3 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container kube-rbac-proxy-main ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: Container kube-rbac-proxy-self ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: Container kube-state-metrics ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: prometheus-adapter-5f78ddc679-vb486 from openshift-monitoring started at 2020-09-21 14:08:28 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container prometheus-adapter ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: multus-hprgb from openshift-multus started at 2020-09-21 14:06:32 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container kube-multus ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: calico-typha-6fc9ff8b5f-gzhgw from calico-system started at 2020-09-21 14:07:18 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.796: INFO: Container calico-typha ready: true, restart count 0 -Sep 21 16:05:55.796: INFO: -Logging pods the kubelet thinks is on node 10.241.51.150 before test -Sep 21 16:05:55.929: INFO: node-ca-ql97q from openshift-image-registry started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.930: INFO: Container node-ca ready: true, restart count 0 -Sep 21 16:05:55.930: INFO: dns-default-z8v9j from openshift-dns started at 2020-09-21 14:12:21 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.930: INFO: Container dns ready: true, restart count 0 -Sep 21 16:05:55.930: INFO: Container dns-node-resolver ready: true, restart count 0 -Sep 21 16:05:55.930: INFO: tuned-5n2xm from openshift-cluster-node-tuning-operator started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.930: INFO: Container tuned ready: true, restart count 0 -Sep 21 16:05:55.930: INFO: calico-typha-6fc9ff8b5f-cbhcz from calico-system started at 2020-09-21 14:13:14 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.930: INFO: Container calico-typha ready: true, restart count 0 -Sep 21 16:05:55.930: INFO: alertmanager-main-1 from openshift-monitoring started at 2020-09-21 14:14:02 +0000 UTC (3 container statuses recorded) -Sep 21 16:05:55.930: INFO: Container alertmanager ready: true, restart count 0 -Sep 21 16:05:55.930: INFO: Container alertmanager-proxy ready: true, restart count 0 -Sep 21 16:05:55.930: INFO: Container config-reloader ready: true, restart count 0 -Sep 21 16:05:55.930: INFO: packageserver-5ffd66994-csrwm from openshift-operator-lifecycle-manager started at 2020-09-21 14:19:37 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.930: INFO: Container packageserver ready: true, restart count 0 -Sep 21 16:05:55.930: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-xzz27 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.930: INFO: Container sonobuoy-worker ready: true, restart count 0 -Sep 21 16:05:55.930: INFO: Container systemd-logs ready: true, restart count 0 -Sep 21 16:05:55.930: INFO: ibm-keepalived-watcher-bntgc from kube-system started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.930: INFO: Container keepalived-watcher ready: true, restart count 0 -Sep 21 16:05:55.930: INFO: multus-n5qc8 from openshift-multus started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.930: INFO: Container kube-multus ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: openshift-kube-proxy-9bv2v from openshift-kube-proxy started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.931: INFO: Container kube-proxy ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: multus-admission-controller-snglx from openshift-multus started at 2020-09-21 14:13:31 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.931: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: Container multus-admission-controller ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: prometheus-k8s-0 from openshift-monitoring started at 2020-09-21 14:15:10 +0000 UTC (7 container statuses recorded) -Sep 21 16:05:55.931: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: Container prom-label-proxy ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: Container prometheus ready: true, restart count 1 -Sep 21 16:05:55.931: INFO: Container prometheus-config-reloader ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: Container prometheus-proxy ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: Container rules-configmap-reloader ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: Container thanos-sidecar ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: console-8549dd6c4f-b5jnn from openshift-console started at 2020-09-21 14:18:13 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.931: INFO: Container console ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: pod-handle-http-request from container-lifecycle-hook-9013 started at 2020-09-21 16:05:22 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.931: INFO: Container pod-handle-http-request ready: false, restart count 0 -Sep 21 16:05:55.931: INFO: ibm-master-proxy-static-10.241.51.150 from kube-system started at 2020-09-21 14:12:19 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.931: INFO: Container ibm-master-proxy-static ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: Container pause ready: true, restart count 0 -Sep 21 16:05:55.931: INFO: node-exporter-flb5l from openshift-monitoring started at 2020-09-21 14:12:21 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.931: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.932: INFO: Container node-exporter ready: true, restart count 0 -Sep 21 16:05:55.932: INFO: calico-node-dn4jv from calico-system started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.932: INFO: Container calico-node ready: true, restart count 0 -Sep 21 16:05:55.932: INFO: telemeter-client-68bdb7f795-nnhfl from openshift-monitoring started at 2020-09-21 14:13:34 +0000 UTC (3 container statuses recorded) -Sep 21 16:05:55.932: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.932: INFO: Container reload ready: true, restart count 0 -Sep 21 16:05:55.932: INFO: Container telemeter-client ready: true, restart count 0 -Sep 21 16:05:55.932: INFO: ibmcloud-block-storage-driver-jf9s4 from kube-system started at 2020-09-21 14:12:26 +0000 UTC (1 container statuses recorded) -Sep 21 16:05:55.932: INFO: Container ibmcloud-block-storage-driver-container ready: true, restart count 0 -Sep 21 16:05:55.932: INFO: grafana-c745ff78c-9pgfx from openshift-monitoring started at 2020-09-21 14:13:42 +0000 UTC (2 container statuses recorded) -Sep 21 16:05:55.932: INFO: Container grafana ready: true, restart count 0 -Sep 21 16:05:55.932: INFO: Container grafana-proxy ready: true, restart count 0 -Sep 21 16:05:55.932: INFO: thanos-querier-8f4c5c746-v982g from openshift-monitoring started at 2020-09-21 14:14:42 +0000 UTC (4 container statuses recorded) -Sep 21 16:05:55.932: INFO: Container kube-rbac-proxy ready: true, restart count 0 -Sep 21 16:05:55.932: INFO: Container oauth-proxy ready: true, restart count 0 -Sep 21 16:05:55.932: INFO: Container prom-label-proxy ready: true, restart count 0 -Sep 21 16:05:55.933: INFO: Container thanos-querier ready: true, restart count 0 -[It] validates that NodeSelector is respected if matching [Conformance] +[It] custom resource defaulting for requests and from storage works [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Trying to launch a pod without a label to get a node which can launch it. -STEP: Explicitly delete pod here to free the resource it takes. -STEP: Trying to apply a random label on the found node. -STEP: verifying the node has the label kubernetes.io/e2e-60395c76-f53c-4697-8f63-d409712e6860 42 -STEP: Trying to relaunch the pod, now with labels. -STEP: removing the label kubernetes.io/e2e-60395c76-f53c-4697-8f63-d409712e6860 off the node 10.241.51.147 -STEP: verifying the node doesn't have the label kubernetes.io/e2e-60395c76-f53c-4697-8f63-d409712e6860 -[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] +Oct 26 16:22:18.749: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:06:02.851: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sched-pred-3825" for this suite. -[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77 - -• [SLOW TEST:8.375 seconds] -[sig-scheduling] SchedulerPredicates [Serial] -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 - validates that NodeSelector is respected if matching [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching [Conformance]","total":280,"completed":67,"skipped":1188,"failed":0} -SSSSSSSSSS +Oct 26 16:22:20.142: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "custom-resource-definition-7754" for this suite. +•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works [Conformance]","total":280,"completed":67,"skipped":978,"failed":0} +SSSSSSSSSSSS ------------------------------ [sig-api-machinery] ResourceQuota - should verify ResourceQuota with terminating scopes. [Conformance] + should create a ResourceQuota and capture the life of a service. [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 [BeforeEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:06:02.928: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:22:20.183: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename resourcequota STEP: Waiting for a default service account to be provisioned in namespace -[It] should verify ResourceQuota with terminating scopes. [Conformance] +[It] should create a ResourceQuota and capture the life of a service. [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a ResourceQuota with terminating scope -STEP: Ensuring ResourceQuota status is calculated -STEP: Creating a ResourceQuota with not terminating scope -STEP: Ensuring ResourceQuota status is calculated -STEP: Creating a long running pod -STEP: Ensuring resource quota with not terminating scope captures the pod usage -STEP: Ensuring resource quota with terminating scope ignored the pod usage -STEP: Deleting the pod -STEP: Ensuring resource quota status released the pod usage -STEP: Creating a terminating pod -STEP: Ensuring resource quota with terminating scope captures the pod usage -STEP: Ensuring resource quota with not terminating scope ignored the pod usage -STEP: Deleting the pod -STEP: Ensuring resource quota status released the pod usage +STEP: Counting existing ResourceQuota +STEP: Creating a ResourceQuota +STEP: Ensuring resource quota status is calculated +STEP: Creating a Service +STEP: Ensuring resource quota status captures service creation +STEP: Deleting a Service +STEP: Ensuring resource quota status released usage [AfterEach] [sig-api-machinery] ResourceQuota /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:06:20.541: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-3259" for this suite. +Oct 26 16:22:31.531: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-1283" for this suite. -• [SLOW TEST:17.684 seconds] +• [SLOW TEST:11.400 seconds] [sig-api-machinery] ResourceQuota /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - should verify ResourceQuota with terminating scopes. [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance]","total":280,"completed":68,"skipped":1198,"failed":0} -SSSSSSS ------------------------------- -[sig-scheduling] NoExecuteTaintManager Single Pod [Serial] - removing taint cancels eviction [Disruptive] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-scheduling] NoExecuteTaintManager Single Pod [Serial] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 16:06:20.611: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename taint-single-pod -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-scheduling] NoExecuteTaintManager Single Pod [Serial] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/taints.go:163 -Sep 21 16:06:20.919: INFO: Waiting up to 1m0s for all nodes to be ready -Sep 21 16:07:21.040: INFO: Waiting for terminating namespaces to be deleted... -[It] removing taint cancels eviction [Disruptive] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 16:07:21.064: INFO: Starting informer... -STEP: Starting pod... -Sep 21 16:07:21.379: INFO: Pod is running on 10.241.51.147. Tainting Node -STEP: Trying to apply a taint on the Node -STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute -STEP: Waiting short time to make sure Pod is queued for deletion -Sep 21 16:07:21.628: INFO: Pod wasn't evicted. Proceeding -Sep 21 16:07:21.634: INFO: Removing taint from Node -STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute -STEP: Waiting some time to make sure that toleration time passed. -Sep 21 16:08:36.758: INFO: Pod wasn't evicted. Test successful -[AfterEach] [sig-scheduling] NoExecuteTaintManager Single Pod [Serial] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:08:36.758: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "taint-single-pod-4371" for this suite. - -• [SLOW TEST:136.281 seconds] -[sig-scheduling] NoExecuteTaintManager Single Pod [Serial] -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 - removing taint cancels eviction [Disruptive] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-scheduling] NoExecuteTaintManager Single Pod [Serial] removing taint cancels eviction [Disruptive] [Conformance]","total":280,"completed":69,"skipped":1205,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch - watch on custom resource definition objects [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 16:08:36.894: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename crd-watch -STEP: Waiting for a default service account to be provisioned in namespace -[It] watch on custom resource definition objects [Conformance] + should create a ResourceQuota and capture the life of a service. [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 16:08:37.163: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Creating first CR -Sep 21 16:08:42.733: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-09-21T16:08:42Z generation:1 name:name1 resourceVersion:57412 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:13e450fa-207f-419f-8fe8-38da4741ec98] num:map[num1:9223372036854775807 num2:1000000]]} -STEP: Creating second CR -Sep 21 16:08:52.764: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-09-21T16:08:52Z generation:1 name:name2 resourceVersion:57480 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:954a3178-24f3-4dca-ba3a-23d4161a4ce1] num:map[num1:9223372036854775807 num2:1000000]]} -STEP: Modifying first CR -Sep 21 16:09:02.798: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-09-21T16:08:42Z generation:2 name:name1 resourceVersion:57526 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:13e450fa-207f-419f-8fe8-38da4741ec98] num:map[num1:9223372036854775807 num2:1000000]]} -STEP: Modifying second CR -Sep 21 16:09:12.827: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-09-21T16:08:52Z generation:2 name:name2 resourceVersion:57573 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:954a3178-24f3-4dca-ba3a-23d4161a4ce1] num:map[num1:9223372036854775807 num2:1000000]]} -STEP: Deleting first CR -Sep 21 16:09:22.905: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-09-21T16:08:42Z generation:2 name:name1 resourceVersion:57611 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:13e450fa-207f-419f-8fe8-38da4741ec98] num:map[num1:9223372036854775807 num2:1000000]]} -STEP: Deleting second CR -Sep 21 16:09:33.016: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-09-21T16:08:52Z generation:2 name:name2 resourceVersion:57650 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:954a3178-24f3-4dca-ba3a-23d4161a4ce1] num:map[num1:9223372036854775807 num2:1000000]]} -[AfterEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:09:43.596: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-watch-2620" for this suite. - -• [SLOW TEST:66.785 seconds] -[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - CustomResourceDefinition Watch - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_watch.go:41 - watch on custom resource definition objects [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance]","total":280,"completed":70,"skipped":1229,"failed":0} -SS +{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance]","total":280,"completed":68,"skipped":990,"failed":0} +SSSSSSSS ------------------------------ [sig-cli] Kubectl client Update Demo should create and stop a replication controller [Conformance] @@ -3428,7 +3243,7 @@ SS [BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:09:43.679: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 +Oct 26 16:22:31.586: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client @@ -3438,69 +3253,62 @@ STEP: Waiting for a default service account to be provisioned in namespace [It] should create and stop a replication controller [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 STEP: creating a replication controller -Sep 21 16:09:44.858: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-6503' -Sep 21 16:09:46.474: INFO: stderr: "" -Sep 21 16:09:46.474: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" +Oct 26 16:22:31.766: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-6736' +Oct 26 16:22:32.280: INFO: stderr: "" +Oct 26 16:22:32.280: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" STEP: waiting for all containers in name=update-demo pods to come up. -Sep 21 16:09:46.474: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6503' -Sep 21 16:09:46.681: INFO: stderr: "" -Sep 21 16:09:46.681: INFO: stdout: "update-demo-nautilus-8gfw6 update-demo-nautilus-h52bp " -Sep 21 16:09:46.681: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8gfw6 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6503' -Sep 21 16:09:46.788: INFO: stderr: "" -Sep 21 16:09:46.788: INFO: stdout: "" -Sep 21 16:09:46.788: INFO: update-demo-nautilus-8gfw6 is created but not running -Sep 21 16:09:51.789: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6503' -Sep 21 16:09:51.933: INFO: stderr: "" -Sep 21 16:09:51.933: INFO: stdout: "update-demo-nautilus-8gfw6 update-demo-nautilus-h52bp " -Sep 21 16:09:51.933: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8gfw6 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6503' -Sep 21 16:09:52.096: INFO: stderr: "" -Sep 21 16:09:52.096: INFO: stdout: "" -Sep 21 16:09:52.096: INFO: update-demo-nautilus-8gfw6 is created but not running -Sep 21 16:09:57.097: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6503' -Sep 21 16:09:57.229: INFO: stderr: "" -Sep 21 16:09:57.229: INFO: stdout: "update-demo-nautilus-8gfw6 update-demo-nautilus-h52bp " -Sep 21 16:09:57.229: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8gfw6 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6503' -Sep 21 16:09:57.353: INFO: stderr: "" -Sep 21 16:09:57.353: INFO: stdout: "true" -Sep 21 16:09:57.353: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8gfw6 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6503' -Sep 21 16:09:57.482: INFO: stderr: "" -Sep 21 16:09:57.482: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" -Sep 21 16:09:57.482: INFO: validating pod update-demo-nautilus-8gfw6 -Sep 21 16:09:57.569: INFO: got data: { +Oct 26 16:22:32.280: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6736' +Oct 26 16:22:32.424: INFO: stderr: "" +Oct 26 16:22:32.424: INFO: stdout: "update-demo-nautilus-rf7hx update-demo-nautilus-wzcm9 " +Oct 26 16:22:32.424: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-rf7hx -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6736' +Oct 26 16:22:32.553: INFO: stderr: "" +Oct 26 16:22:32.553: INFO: stdout: "" +Oct 26 16:22:32.553: INFO: update-demo-nautilus-rf7hx is created but not running +Oct 26 16:22:37.553: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-6736' +Oct 26 16:22:37.694: INFO: stderr: "" +Oct 26 16:22:37.694: INFO: stdout: "update-demo-nautilus-rf7hx update-demo-nautilus-wzcm9 " +Oct 26 16:22:37.694: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-rf7hx -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6736' +Oct 26 16:22:37.843: INFO: stderr: "" +Oct 26 16:22:37.843: INFO: stdout: "true" +Oct 26 16:22:37.843: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-rf7hx -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6736' +Oct 26 16:22:37.987: INFO: stderr: "" +Oct 26 16:22:37.987: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Oct 26 16:22:37.987: INFO: validating pod update-demo-nautilus-rf7hx +Oct 26 16:22:38.020: INFO: got data: { "image": "nautilus.jpg" } -Sep 21 16:09:57.569: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -Sep 21 16:09:57.569: INFO: update-demo-nautilus-8gfw6 is verified up and running -Sep 21 16:09:57.569: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-h52bp -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6503' -Sep 21 16:09:57.696: INFO: stderr: "" -Sep 21 16:09:57.696: INFO: stdout: "true" -Sep 21 16:09:57.697: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-h52bp -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6503' -Sep 21 16:09:57.842: INFO: stderr: "" -Sep 21 16:09:57.842: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" -Sep 21 16:09:57.842: INFO: validating pod update-demo-nautilus-h52bp -Sep 21 16:09:57.898: INFO: got data: { +Oct 26 16:22:38.020: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Oct 26 16:22:38.020: INFO: update-demo-nautilus-rf7hx is verified up and running +Oct 26 16:22:38.020: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-wzcm9 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-6736' +Oct 26 16:22:38.155: INFO: stderr: "" +Oct 26 16:22:38.155: INFO: stdout: "true" +Oct 26 16:22:38.155: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-wzcm9 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-6736' +Oct 26 16:22:38.280: INFO: stderr: "" +Oct 26 16:22:38.280: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0" +Oct 26 16:22:38.280: INFO: validating pod update-demo-nautilus-wzcm9 +Oct 26 16:22:38.313: INFO: got data: { "image": "nautilus.jpg" } -Sep 21 16:09:57.898: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -Sep 21 16:09:57.898: INFO: update-demo-nautilus-h52bp is verified up and running +Oct 26 16:22:38.313: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Oct 26 16:22:38.313: INFO: update-demo-nautilus-wzcm9 is verified up and running STEP: using delete to clean up resources -Sep 21 16:09:57.898: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete --grace-period=0 --force -f - --namespace=kubectl-6503' -Sep 21 16:09:58.064: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Sep 21 16:09:58.064: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" -Sep 21 16:09:58.064: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-6503' -Sep 21 16:09:58.318: INFO: stderr: "No resources found in kubectl-6503 namespace.\n" -Sep 21 16:09:58.318: INFO: stdout: "" -Sep 21 16:09:58.318: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -l name=update-demo --namespace=kubectl-6503 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' -Sep 21 16:09:58.467: INFO: stderr: "" -Sep 21 16:09:58.467: INFO: stdout: "" +Oct 26 16:22:38.313: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete --grace-period=0 --force -f - --namespace=kubectl-6736' +Oct 26 16:22:38.468: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Oct 26 16:22:38.468: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" +Oct 26 16:22:38.468: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-6736' +Oct 26 16:22:38.649: INFO: stderr: "No resources found in kubectl-6736 namespace.\n" +Oct 26 16:22:38.649: INFO: stdout: "" +Oct 26 16:22:38.649: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -l name=update-demo --namespace=kubectl-6736 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' +Oct 26 16:22:38.802: INFO: stderr: "" +Oct 26 16:22:38.802: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:09:58.468: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-6503" for this suite. +Oct 26 16:22:38.803: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-6736" for this suite. -• [SLOW TEST:14.879 seconds] +• [SLOW TEST:7.256 seconds] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 Update Demo @@ -3508,1107 +3316,1680 @@ STEP: Destroying namespace "kubectl-6503" for this suite. should create and stop a replication controller [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-cli] Kubectl client Update Demo should create and stop a replication controller [Conformance]","total":280,"completed":71,"skipped":1231,"failed":0} -SSS ------------------------------- -[sig-storage] Projected configMap - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] +{"msg":"PASSED [sig-cli] Kubectl client Update Demo should create and stop a replication controller [Conformance]","total":280,"completed":69,"skipped":998,"failed":0} +[sig-api-machinery] Garbage collector + should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] Projected configMap +[BeforeEach] [sig-api-machinery] Garbage collector /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:09:58.558: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename projected +Oct 26 16:22:38.843: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename gc STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating configMap with name projected-configmap-test-volume-6608ec54-64e3-408c-a112-31eb9e00e546 -STEP: Creating a pod to test consume configMaps -Sep 21 16:09:59.178: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-be79045f-df9c-45bb-8d4a-63a0cb3f9928" in namespace "projected-814" to be "success or failure" -Sep 21 16:09:59.272: INFO: Pod "pod-projected-configmaps-be79045f-df9c-45bb-8d4a-63a0cb3f9928": Phase="Pending", Reason="", readiness=false. Elapsed: 93.837882ms -Sep 21 16:10:01.516: INFO: Pod "pod-projected-configmaps-be79045f-df9c-45bb-8d4a-63a0cb3f9928": Phase="Pending", Reason="", readiness=false. Elapsed: 2.337993407s -Sep 21 16:10:03.552: INFO: Pod "pod-projected-configmaps-be79045f-df9c-45bb-8d4a-63a0cb3f9928": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.374004056s -STEP: Saw pod success -Sep 21 16:10:03.552: INFO: Pod "pod-projected-configmaps-be79045f-df9c-45bb-8d4a-63a0cb3f9928" satisfied condition "success or failure" -Sep 21 16:10:03.591: INFO: Trying to get logs from node 10.241.51.147 pod pod-projected-configmaps-be79045f-df9c-45bb-8d4a-63a0cb3f9928 container projected-configmap-volume-test: -STEP: delete the pod -Sep 21 16:10:04.006: INFO: Waiting for pod pod-projected-configmaps-be79045f-df9c-45bb-8d4a-63a0cb3f9928 to disappear -Sep 21 16:10:04.242: INFO: Pod pod-projected-configmaps-be79045f-df9c-45bb-8d4a-63a0cb3f9928 no longer exists -[AfterEach] [sig-storage] Projected configMap - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:10:04.242: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-814" for this suite. - -• [SLOW TEST:6.296 seconds] -[sig-storage] Projected configMap -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34 - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":72,"skipped":1234,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-apps] Job - should adopt matching orphans and release non-matching pods [Conformance] +[It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-apps] Job +STEP: create the rc1 +STEP: create the rc2 +STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well +STEP: delete the rc simpletest-rc-to-be-deleted +STEP: wait for the rc to be deleted +STEP: Gathering metrics +W1026 16:22:49.289062 26 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +Oct 26 16:22:49.289: INFO: For apiserver_request_total: +For apiserver_request_latency_seconds: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: + +[AfterEach] [sig-api-machinery] Garbage collector + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:22:49.289: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-7169" for this suite. + +• [SLOW TEST:10.495 seconds] +[sig-api-machinery] Garbage collector +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 + should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +------------------------------ +{"msg":"PASSED [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]","total":280,"completed":70,"skipped":998,"failed":0} +SSSSS +------------------------------ +[sig-network] Services + should be able to change the type from ClusterIP to ExternalName [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-network] Services /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:10:04.855: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename job +Oct 26 16:22:49.339: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename services STEP: Waiting for a default service account to be provisioned in namespace -[It] should adopt matching orphans and release non-matching pods [Conformance] +[BeforeEach] [sig-network] Services + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139 +[It] should be able to change the type from ClusterIP to ExternalName [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a job -STEP: Ensuring active pods == parallelism -STEP: Orphaning one of the Job's Pods -Sep 21 16:10:10.106: INFO: Successfully updated pod "adopt-release-97cmr" -STEP: Checking that the Job readopts the Pod -Sep 21 16:10:10.106: INFO: Waiting up to 15m0s for pod "adopt-release-97cmr" in namespace "job-2177" to be "adopted" -Sep 21 16:10:10.142: INFO: Pod "adopt-release-97cmr": Phase="Running", Reason="", readiness=true. Elapsed: 36.30461ms -Sep 21 16:10:10.142: INFO: Pod "adopt-release-97cmr" satisfied condition "adopted" -STEP: Removing the labels from the Job's Pod -Sep 21 16:10:10.961: INFO: Successfully updated pod "adopt-release-97cmr" -STEP: Checking that the Job releases the Pod -Sep 21 16:10:10.961: INFO: Waiting up to 15m0s for pod "adopt-release-97cmr" in namespace "job-2177" to be "released" -Sep 21 16:10:10.995: INFO: Pod "adopt-release-97cmr": Phase="Running", Reason="", readiness=true. Elapsed: 33.748129ms -Sep 21 16:10:10.995: INFO: Pod "adopt-release-97cmr" satisfied condition "released" -[AfterEach] [sig-apps] Job +STEP: creating a service clusterip-service with the type=ClusterIP in namespace services-607 +STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service +STEP: creating service externalsvc in namespace services-607 +STEP: creating replication controller externalsvc in namespace services-607 +I1026 16:22:49.566953 26 runners.go:189] Created replication controller with name: externalsvc, namespace: services-607, replica count: 2 +I1026 16:22:52.617774 26 runners.go:189] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +STEP: changing the ClusterIP service to type=ExternalName +Oct 26 16:22:52.681: INFO: Creating new exec pod +Oct 26 16:22:54.769: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-607 execpodhhk6v -- /bin/sh -x -c nslookup clusterip-service' +Oct 26 16:22:55.171: INFO: stderr: "+ nslookup clusterip-service\n" +Oct 26 16:22:55.171: INFO: stdout: "Server:\t\t172.21.0.10\nAddress:\t172.21.0.10#53\n\nclusterip-service.services-607.svc.cluster.local\tcanonical name = externalsvc.services-607.svc.cluster.local.\nName:\texternalsvc.services-607.svc.cluster.local\nAddress: 172.21.87.8\n\n" +STEP: deleting ReplicationController externalsvc in namespace services-607, will wait for the garbage collector to delete the pods +Oct 26 16:22:55.259: INFO: Deleting ReplicationController externalsvc took: 24.289288ms +Oct 26 16:22:55.360: INFO: Terminating ReplicationController externalsvc pods took: 101.317139ms +Oct 26 16:23:09.638: INFO: Cleaning up the ClusterIP to ExternalName test service +[AfterEach] [sig-network] Services /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:10:10.995: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "job-2177" for this suite. +Oct 26 16:23:09.694: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-607" for this suite. +[AfterEach] [sig-network] Services + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143 -• [SLOW TEST:6.221 seconds] -[sig-apps] Job -/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 - should adopt matching orphans and release non-matching pods [Conformance] +• [SLOW TEST:20.397 seconds] +[sig-network] Services +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 + should be able to change the type from ClusterIP to ExternalName [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance]","total":280,"completed":73,"skipped":1257,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance]","total":280,"completed":71,"skipped":1003,"failed":0} +SS ------------------------------ -[sig-apps] ReplicationController - should surface a failure condition on a common issue like exceeded quota [Conformance] +[sig-cli] Kubectl client Kubectl label + should update the label on a resource [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-apps] ReplicationController +[BeforeEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:10:11.082: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename replication-controller +Oct 26 16:23:09.738: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename kubectl STEP: Waiting for a default service account to be provisioned in namespace -[It] should surface a failure condition on a common issue like exceeded quota [Conformance] +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278 +[BeforeEach] Kubectl label + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1382 +STEP: creating the pod +Oct 26 16:23:09.862: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-4476' +Oct 26 16:23:10.227: INFO: stderr: "" +Oct 26 16:23:10.227: INFO: stdout: "pod/pause created\n" +Oct 26 16:23:10.227: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] +Oct 26 16:23:10.227: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-4476" to be "running and ready" +Oct 26 16:23:10.249: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 22.271353ms +Oct 26 16:23:12.264: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 2.037019072s +Oct 26 16:23:12.264: INFO: Pod "pause" satisfied condition "running and ready" +Oct 26 16:23:12.264: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] +[It] should update the label on a resource [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 16:10:11.429: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace -STEP: Creating rc "condition-test" that asks for more than the allowed pod quota -STEP: Checking rc "condition-test" has the desired failure condition set -STEP: Scaling down rc "condition-test" to satisfy pod quota -Sep 21 16:10:12.242: INFO: Updating replication controller "condition-test" -STEP: Checking rc "condition-test" has no failure condition set -[AfterEach] [sig-apps] ReplicationController +STEP: adding the label testing-label with value testing-label-value to a pod +Oct 26 16:23:12.264: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 label pods pause testing-label=testing-label-value --namespace=kubectl-4476' +Oct 26 16:23:12.469: INFO: stderr: "" +Oct 26 16:23:12.469: INFO: stdout: "pod/pause labeled\n" +STEP: verifying the pod has the label testing-label with the value testing-label-value +Oct 26 16:23:12.469: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pod pause -L testing-label --namespace=kubectl-4476' +Oct 26 16:23:12.603: INFO: stderr: "" +Oct 26 16:23:12.603: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 2s testing-label-value\n" +STEP: removing the label testing-label of a pod +Oct 26 16:23:12.603: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 label pods pause testing-label- --namespace=kubectl-4476' +Oct 26 16:23:12.787: INFO: stderr: "" +Oct 26 16:23:12.787: INFO: stdout: "pod/pause labeled\n" +STEP: verifying the pod doesn't have the label testing-label +Oct 26 16:23:12.787: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pod pause -L testing-label --namespace=kubectl-4476' +Oct 26 16:23:12.931: INFO: stderr: "" +Oct 26 16:23:12.932: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 2s \n" +[AfterEach] Kubectl label + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1389 +STEP: using delete to clean up resources +Oct 26 16:23:12.932: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete --grace-period=0 --force -f - --namespace=kubectl-4476' +Oct 26 16:23:13.124: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Oct 26 16:23:13.124: INFO: stdout: "pod \"pause\" force deleted\n" +Oct 26 16:23:13.124: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get rc,svc -l name=pause --no-headers --namespace=kubectl-4476' +Oct 26 16:23:13.282: INFO: stderr: "No resources found in kubectl-4476 namespace.\n" +Oct 26 16:23:13.282: INFO: stdout: "" +Oct 26 16:23:13.282: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -l name=pause --namespace=kubectl-4476 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' +Oct 26 16:23:13.427: INFO: stderr: "" +Oct 26 16:23:13.427: INFO: stdout: "" +[AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:10:12.270: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "replication-controller-5064" for this suite. -•{"msg":"PASSED [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance]","total":280,"completed":74,"skipped":1308,"failed":0} +Oct 26 16:23:13.427: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-4476" for this suite. +•{"msg":"PASSED [sig-cli] Kubectl client Kubectl label should update the label on a resource [Conformance]","total":280,"completed":72,"skipped":1005,"failed":0} SSSSSSSSSS ------------------------------ -[sig-api-machinery] Watchers - should be able to restart watching from the last resource version observed by the previous watch [Conformance] +[sig-auth] ServiceAccounts + should allow opting out of API token automount [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] Watchers +[BeforeEach] [sig-auth] ServiceAccounts /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Sep 21 16:10:12.381: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename watch +Oct 26 16:23:13.463: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename svcaccounts STEP: Waiting for a default service account to be provisioned in namespace -[It] should be able to restart watching from the last resource version observed by the previous watch [Conformance] +[It] should allow opting out of API token automount [Conformance] /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: creating a watch on configmaps -STEP: creating a new configmap -STEP: modifying the configmap once -STEP: closing the watch once it receives two notifications -Sep 21 16:10:12.667: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-4312 /api/v1/namespaces/watch-4312/configmaps/e2e-watch-test-watch-closed 0c51cf58-e775-4453-afea-f4725c9091ff 58228 0 2020-09-21 16:10:12 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] []},Data:map[string]string{},BinaryData:map[string][]byte{},} -Sep 21 16:10:12.667: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-4312 /api/v1/namespaces/watch-4312/configmaps/e2e-watch-test-watch-closed 0c51cf58-e775-4453-afea-f4725c9091ff 58235 0 2020-09-21 16:10:12 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},} -STEP: modifying the configmap a second time, while the watch is closed -STEP: creating a new watch on configmaps from the last resource version observed by the first watch -STEP: deleting the configmap -STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed -Sep 21 16:10:13.016: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-4312 /api/v1/namespaces/watch-4312/configmaps/e2e-watch-test-watch-closed 0c51cf58-e775-4453-afea-f4725c9091ff 58240 0 2020-09-21 16:10:12 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} -Sep 21 16:10:13.016: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-4312 /api/v1/namespaces/watch-4312/configmaps/e2e-watch-test-watch-closed 0c51cf58-e775-4453-afea-f4725c9091ff 58244 0 2020-09-21 16:10:12 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} -[AfterEach] [sig-api-machinery] Watchers +STEP: getting the auto-created API token +Oct 26 16:23:14.253: INFO: created pod pod-service-account-defaultsa +Oct 26 16:23:14.253: INFO: pod pod-service-account-defaultsa service account token volume mount: true +Oct 26 16:23:14.291: INFO: created pod pod-service-account-mountsa +Oct 26 16:23:14.291: INFO: pod pod-service-account-mountsa service account token volume mount: true +Oct 26 16:23:14.330: INFO: created pod pod-service-account-nomountsa +Oct 26 16:23:14.330: INFO: pod pod-service-account-nomountsa service account token volume mount: false +Oct 26 16:23:14.371: INFO: created pod pod-service-account-defaultsa-mountspec +Oct 26 16:23:14.371: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true +Oct 26 16:23:14.413: INFO: created pod pod-service-account-mountsa-mountspec +Oct 26 16:23:14.413: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true +Oct 26 16:23:14.450: INFO: created pod pod-service-account-nomountsa-mountspec +Oct 26 16:23:14.450: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true +Oct 26 16:23:14.490: INFO: created pod pod-service-account-defaultsa-nomountspec +Oct 26 16:23:14.490: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false +Oct 26 16:23:14.533: INFO: created pod pod-service-account-mountsa-nomountspec +Oct 26 16:23:14.533: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false +Oct 26 16:23:14.568: INFO: created pod pod-service-account-nomountsa-nomountspec +Oct 26 16:23:14.568: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false +[AfterEach] [sig-auth] ServiceAccounts + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:23:14.568: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svcaccounts-5881" for this suite. +•{"msg":"PASSED [sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance]","total":280,"completed":73,"skipped":1015,"failed":0} +S +------------------------------ +[k8s.io] Kubelet when scheduling a busybox command in a pod + should print the output to logs [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [k8s.io] Kubelet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:23:14.606: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename kubelet-test +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [k8s.io] Kubelet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37 +[It] should print the output to logs [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[AfterEach] [k8s.io] Kubelet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:23:17.947: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubelet-test-387" for this suite. +•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance]","total":280,"completed":74,"skipped":1016,"failed":0} +SSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + should have a working scale subresource [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:23:17.993: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename statefulset +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64 +[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79 +STEP: Creating service test in namespace statefulset-3583 +[It] should have a working scale subresource [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating statefulset ss in namespace statefulset-3583 +Oct 26 16:23:18.227: INFO: Found 0 stateful pods, waiting for 1 +Oct 26 16:23:28.245: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +STEP: getting scale subresource +STEP: updating a scale subresource +STEP: verifying the statefulset Spec.Replicas was modified +[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 +Oct 26 16:23:28.352: INFO: Deleting all statefulset in ns statefulset-3583 +Oct 26 16:23:28.370: INFO: Scaling statefulset ss to 0 +Oct 26 16:23:48.440: INFO: Waiting for statefulset status.replicas updated to 0 +Oct 26 16:23:48.457: INFO: Deleting statefulset ss +[AfterEach] [sig-apps] StatefulSet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:23:48.543: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-3583" for this suite. + +• [SLOW TEST:30.588 seconds] +[sig-apps] StatefulSet +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 + [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 + should have a working scale subresource [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +------------------------------ +{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance]","total":280,"completed":75,"skipped":1035,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-api-machinery] ResourceQuota + should create a ResourceQuota and capture the life of a pod. [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-api-machinery] ResourceQuota + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:23:48.583: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename resourcequota +STEP: Waiting for a default service account to be provisioned in namespace +[It] should create a ResourceQuota and capture the life of a pod. [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Counting existing ResourceQuota +STEP: Creating a ResourceQuota +STEP: Ensuring resource quota status is calculated +STEP: Creating a Pod that fits quota +STEP: Ensuring ResourceQuota status captures the pod usage +STEP: Not allowing a pod to be created that exceeds remaining quota +STEP: Not allowing a pod to be created that exceeds remaining quota(validation on extended resources) +STEP: Ensuring a pod cannot update its resource requirements +STEP: Ensuring attempts to update pod resource requirements did not change quota usage +STEP: Deleting the pod +STEP: Ensuring resource quota status released the pod usage +[AfterEach] [sig-api-machinery] ResourceQuota + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:24:02.070: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-1092" for this suite. + +• [SLOW TEST:13.527 seconds] +[sig-api-machinery] ResourceQuota +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 + should create a ResourceQuota and capture the life of a pod. [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +------------------------------ +{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance]","total":280,"completed":76,"skipped":1064,"failed":0} +SSSSSSSSSSSSSSS +------------------------------ +[sig-storage] Secrets + should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-storage] Secrets + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:24:02.111: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename secrets +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating secret with name secret-test-map-ad7496dc-f0a0-43ca-a08f-ced80a59675d +STEP: Creating a pod to test consume secrets +Oct 26 16:24:02.336: INFO: Waiting up to 5m0s for pod "pod-secrets-be1092b0-2ed8-4043-af0f-b433e04724b7" in namespace "secrets-6459" to be "success or failure" +Oct 26 16:24:02.352: INFO: Pod "pod-secrets-be1092b0-2ed8-4043-af0f-b433e04724b7": Phase="Pending", Reason="", readiness=false. Elapsed: 15.802342ms +Oct 26 16:24:04.370: INFO: Pod "pod-secrets-be1092b0-2ed8-4043-af0f-b433e04724b7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.033564684s +STEP: Saw pod success +Oct 26 16:24:04.370: INFO: Pod "pod-secrets-be1092b0-2ed8-4043-af0f-b433e04724b7" satisfied condition "success or failure" +Oct 26 16:24:04.386: INFO: Trying to get logs from node 10.72.119.74 pod pod-secrets-be1092b0-2ed8-4043-af0f-b433e04724b7 container secret-volume-test: +STEP: delete the pod +Oct 26 16:24:04.496: INFO: Waiting for pod pod-secrets-be1092b0-2ed8-4043-af0f-b433e04724b7 to disappear +Oct 26 16:24:04.511: INFO: Pod pod-secrets-be1092b0-2ed8-4043-af0f-b433e04724b7 no longer exists +[AfterEach] [sig-storage] Secrets + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:24:04.511: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-6459" for this suite. +•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":77,"skipped":1079,"failed":0} +SSSSSSSSS +------------------------------ +[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + Should recreate evicted statefulset [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:24:04.545: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename statefulset +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-apps] StatefulSet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64 +[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79 +STEP: Creating service test in namespace statefulset-1750 +[It] Should recreate evicted statefulset [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Looking for a node to schedule stateful set and pod +STEP: Creating pod with conflicting port in namespace statefulset-1750 +STEP: Creating statefulset with conflicting port in namespace statefulset-1750 +STEP: Waiting until pod test-pod will start running in namespace statefulset-1750 +STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-1750 +Oct 26 16:24:08.895: INFO: Observed stateful pod in namespace: statefulset-1750, name: ss-0, uid: 7ade9757-8d37-4f1b-8ad1-9e95bc4f6e77, status phase: Pending. Waiting for statefulset controller to delete. +Oct 26 16:24:08.935: INFO: Observed stateful pod in namespace: statefulset-1750, name: ss-0, uid: 7ade9757-8d37-4f1b-8ad1-9e95bc4f6e77, status phase: Failed. Waiting for statefulset controller to delete. +Oct 26 16:24:08.957: INFO: Observed stateful pod in namespace: statefulset-1750, name: ss-0, uid: 7ade9757-8d37-4f1b-8ad1-9e95bc4f6e77, status phase: Failed. Waiting for statefulset controller to delete. +Oct 26 16:24:08.975: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-1750 +STEP: Removing pod with conflicting port in namespace statefulset-1750 +STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-1750 and will be in running state +[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90 +Oct 26 16:24:13.070: INFO: Deleting all statefulset in ns statefulset-1750 +Oct 26 16:24:13.084: INFO: Scaling statefulset ss to 0 +Oct 26 16:24:23.149: INFO: Waiting for statefulset status.replicas updated to 0 +Oct 26 16:24:23.166: INFO: Deleting statefulset ss +[AfterEach] [sig-apps] StatefulSet + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:24:23.248: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-1750" for this suite. + +• [SLOW TEST:18.743 seconds] +[sig-apps] StatefulSet +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 + [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 + Should recreate evicted statefulset [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +------------------------------ +{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance]","total":280,"completed":78,"skipped":1088,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] ConfigMap + updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-storage] ConfigMap + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:24:23.289: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename configmap +STEP: Waiting for a default service account to be provisioned in namespace +[It] updates should be reflected in volume [NodeConformance] [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +Oct 26 16:24:23.447: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node +STEP: Creating configMap with name configmap-test-upd-1d9e0a8c-45f5-48ec-a1a5-4070afe9891f +STEP: Creating the pod +STEP: Updating configmap configmap-test-upd-1d9e0a8c-45f5-48ec-a1a5-4070afe9891f +STEP: waiting to observe update in volume +[AfterEach] [sig-storage] ConfigMap + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:24:27.688: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-8007" for this suite. +•{"msg":"PASSED [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":79,"skipped":1138,"failed":0} +SS +------------------------------ +[sig-network] Services + should serve a basic endpoint from pods [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-network] Services + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:24:27.723: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename services +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-network] Services + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139 +[It] should serve a basic endpoint from pods [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: creating service endpoint-test2 in namespace services-8287 +STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8287 to expose endpoints map[] +Oct 26 16:24:27.938: INFO: successfully validated that service endpoint-test2 in namespace services-8287 exposes endpoints map[] (19.756177ms elapsed) +STEP: Creating pod pod1 in namespace services-8287 +STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8287 to expose endpoints map[pod1:[80]] +Oct 26 16:24:31.136: INFO: successfully validated that service endpoint-test2 in namespace services-8287 exposes endpoints map[pod1:[80]] (3.150391287s elapsed) +STEP: Creating pod pod2 in namespace services-8287 +STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8287 to expose endpoints map[pod1:[80] pod2:[80]] +Oct 26 16:24:33.318: INFO: successfully validated that service endpoint-test2 in namespace services-8287 exposes endpoints map[pod1:[80] pod2:[80]] (2.14697484s elapsed) +STEP: Deleting pod pod1 in namespace services-8287 +STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8287 to expose endpoints map[pod2:[80]] +Oct 26 16:24:33.389: INFO: successfully validated that service endpoint-test2 in namespace services-8287 exposes endpoints map[pod2:[80]] (43.126417ms elapsed) +STEP: Deleting pod pod2 in namespace services-8287 +STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-8287 to expose endpoints map[] +Oct 26 16:24:33.441: INFO: successfully validated that service endpoint-test2 in namespace services-8287 exposes endpoints map[] (14.682623ms elapsed) +[AfterEach] [sig-network] Services + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:24:33.511: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-8287" for this suite. +[AfterEach] [sig-network] Services + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143 + +• [SLOW TEST:5.842 seconds] +[sig-network] Services +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23 + should serve a basic endpoint from pods [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +------------------------------ +{"msg":"PASSED [sig-network] Services should serve a basic endpoint from pods [Conformance]","total":280,"completed":80,"skipped":1140,"failed":0} +SSSSSSSSSSSSSSS +------------------------------ +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + should mutate pod and apply defaults after mutation [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:24:33.565: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename webhook +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 +STEP: Setting up server cert +STEP: Create role binding to let webhook read extension-apiserver-authentication +STEP: Deploying the webhook pod +STEP: Wait for the deployment to be ready +Oct 26 16:24:34.604: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +Oct 26 16:24:36.681: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326274, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326274, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326274, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326274, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service +STEP: Verifying the service has paired with the endpoint +Oct 26 16:24:39.733: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should mutate pod and apply defaults after mutation [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Registering the mutating pod webhook via the AdmissionRegistration API +STEP: create a pod that should be updated by the webhook +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Oct 26 16:24:39.992: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-2284" for this suite. +STEP: Destroying namespace "webhook-2284-markers" for this suite. +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 + +• [SLOW TEST:6.689 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 + should mutate pod and apply defaults after mutation [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +------------------------------ +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance]","total":280,"completed":81,"skipped":1155,"failed":0} +SSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-cli] Kubectl client Guestbook application + should create and stop a working application [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Oct 26 16:24:40.254: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394 +STEP: Building a namespace api object, basename kubectl +STEP: Waiting for a default service account to be provisioned in namespace +[BeforeEach] [sig-cli] Kubectl client + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278 +[It] should create and stop a working application [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: creating all guestbook components +Oct 26 16:24:40.419: INFO: apiVersion: v1 +kind: Service +metadata: + name: agnhost-slave + labels: + app: agnhost + role: slave + tier: backend +spec: + ports: + - port: 6379 + selector: + app: agnhost + role: slave + tier: backend + +Oct 26 16:24:40.419: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-7751' +Oct 26 16:24:41.418: INFO: stderr: "" +Oct 26 16:24:41.418: INFO: stdout: "service/agnhost-slave created\n" +Oct 26 16:24:41.418: INFO: apiVersion: v1 +kind: Service +metadata: + name: agnhost-master + labels: + app: agnhost + role: master + tier: backend +spec: + ports: + - port: 6379 + targetPort: 6379 + selector: + app: agnhost + role: master + tier: backend + +Oct 26 16:24:41.418: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-7751' +Oct 26 16:24:41.925: INFO: stderr: "" +Oct 26 16:24:41.925: INFO: stdout: "service/agnhost-master created\n" +Oct 26 16:24:41.925: INFO: apiVersion: v1 +kind: Service +metadata: + name: frontend + labels: + app: guestbook + tier: frontend +spec: + # if your cluster supports it, uncomment the following to automatically create + # an external load-balanced IP for the frontend service. + # type: LoadBalancer + ports: + - port: 80 + selector: + app: guestbook + tier: frontend + +Oct 26 16:24:41.925: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-7751' +Oct 26 16:24:42.528: INFO: stderr: "" +Oct 26 16:24:42.528: INFO: stdout: "service/frontend created\n" +Oct 26 16:24:42.528: INFO: apiVersion: apps/v1 +kind: Deployment +metadata: + name: frontend +spec: + replicas: 3 + selector: + matchLabels: + app: guestbook + tier: frontend + template: + metadata: + labels: + app: guestbook + tier: frontend + spec: + containers: + - name: guestbook-frontend + image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8 + args: [ "guestbook", "--backend-port", "6379" ] + resources: + requests: + cpu: 100m + memory: 100Mi + ports: + - containerPort: 80 + +Oct 26 16:24:42.528: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-7751' +Oct 26 16:24:42.900: INFO: stderr: "" +Oct 26 16:24:42.900: INFO: stdout: "deployment.apps/frontend created\n" +Oct 26 16:24:42.901: INFO: apiVersion: apps/v1 +kind: Deployment +metadata: + name: agnhost-master +spec: + replicas: 1 + selector: + matchLabels: + app: agnhost + role: master + tier: backend + template: + metadata: + labels: + app: agnhost + role: master + tier: backend + spec: + containers: + - name: master + image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8 + args: [ "guestbook", "--http-port", "6379" ] + resources: + requests: + cpu: 100m + memory: 100Mi + ports: + - containerPort: 6379 + +Oct 26 16:24:42.901: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-7751' +Oct 26 16:24:43.280: INFO: stderr: "" +Oct 26 16:24:43.280: INFO: stdout: "deployment.apps/agnhost-master created\n" +Oct 26 16:24:43.281: INFO: apiVersion: apps/v1 +kind: Deployment +metadata: + name: agnhost-slave +spec: + replicas: 2 + selector: + matchLabels: + app: agnhost + role: slave + tier: backend + template: + metadata: + labels: + app: agnhost + role: slave + tier: backend + spec: + containers: + - name: slave + image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8 + args: [ "guestbook", "--slaveof", "agnhost-master", "--http-port", "6379" ] + resources: + requests: + cpu: 100m + memory: 100Mi + ports: + - containerPort: 6379 + +Oct 26 16:24:43.281: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-7751' +Oct 26 16:24:43.673: INFO: stderr: "" +Oct 26 16:24:43.674: INFO: stdout: "deployment.apps/agnhost-slave created\n" +STEP: validating guestbook app +Oct 26 16:24:43.674: INFO: Waiting for all frontend pods to be Running. +Oct 26 16:24:48.724: INFO: Waiting for frontend to serve content. +Oct 26 16:24:48.772: INFO: Trying to add a new entry to the guestbook. +Oct 26 16:24:48.831: INFO: Verifying that added entry can be retrieved. +Oct 26 16:24:48.878: INFO: Failed to get response from guestbook. err: , response: {"data":""} +STEP: using delete to clean up resources +Oct 26 16:24:53.922: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete --grace-period=0 --force -f - --namespace=kubectl-7751' +Oct 26 16:24:54.132: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Oct 26 16:24:54.132: INFO: stdout: "service \"agnhost-slave\" force deleted\n" +STEP: using delete to clean up resources +Oct 26 16:24:54.133: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete --grace-period=0 --force -f - --namespace=kubectl-7751' +Oct 26 16:24:54.360: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Oct 26 16:24:54.360: INFO: stdout: "service \"agnhost-master\" force deleted\n" +STEP: using delete to clean up resources +Oct 26 16:24:54.360: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete --grace-period=0 --force -f - --namespace=kubectl-7751' +Oct 26 16:24:54.593: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Oct 26 16:24:54.593: INFO: stdout: "service \"frontend\" force deleted\n" +STEP: using delete to clean up resources +Oct 26 16:24:54.593: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete --grace-period=0 --force -f - --namespace=kubectl-7751' +Oct 26 16:24:54.773: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Oct 26 16:24:54.773: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" +STEP: using delete to clean up resources +Oct 26 16:24:54.773: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete --grace-period=0 --force -f - --namespace=kubectl-7751' +Oct 26 16:24:54.953: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Oct 26 16:24:54.953: INFO: stdout: "deployment.apps \"agnhost-master\" force deleted\n" +STEP: using delete to clean up resources +Oct 26 16:24:54.953: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete --grace-period=0 --force -f - --namespace=kubectl-7751' +Oct 26 16:24:55.140: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Oct 26 16:24:55.140: INFO: stdout: "deployment.apps \"agnhost-slave\" force deleted\n" +[AfterEach] [sig-cli] Kubectl client /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Sep 21 16:10:13.016: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "watch-4312" for this suite. -•{"msg":"PASSED [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance]","total":280,"completed":75,"skipped":1318,"failed":0} -SSSSS +Oct 26 16:24:55.140: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-7751" for this suite. + +• [SLOW TEST:14.927 seconds] +[sig-cli] Kubectl client +/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 + Guestbook application + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:386 + should create and stop a working application [Conformance] + /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -[sig-network] Proxy version v1 - should proxy logs on node using proxy subresource [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] version v1 - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Sep 21 16:10:13.121: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458 -STEP: Building a namespace api object, basename proxy -STEP: Waiting for a default service account to be provisioned in namespace -[It] should proxy logs on node using proxy subresource [Conformance] - /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Sep 21 16:10:13.707: INFO: (0) /api/v1/nodes/10.241.51.147/proxy/logs/:
-anaconda/
-at/
-at-no-rotat... (200; 139.370406ms)
-Sep 21 16:10:13.778: INFO: (1) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 71.190541ms)
-Sep 21 16:10:13.864: INFO: (2) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 86.154697ms)
-Sep 21 16:10:13.957: INFO: (3) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 92.549541ms)
-Sep 21 16:10:14.045: INFO: (4) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 88.430108ms)
-Sep 21 16:10:14.093: INFO: (5) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 48.135822ms)
-Sep 21 16:10:14.155: INFO: (6) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 61.371874ms)
-Sep 21 16:10:14.204: INFO: (7) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 49.062365ms)
-Sep 21 16:10:14.276: INFO: (8) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 71.239885ms)
-Sep 21 16:10:14.324: INFO: (9) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 48.387511ms)
-Sep 21 16:10:14.498: INFO: (10) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 173.538471ms)
-Sep 21 16:10:14.578: INFO: (11) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 79.844865ms)
-Sep 21 16:10:14.896: INFO: (12) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 318.327198ms)
-Sep 21 16:10:15.082: INFO: (13) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 186.038679ms)
-Sep 21 16:10:15.128: INFO: (14) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 44.964348ms)
-Sep 21 16:10:15.206: INFO: (15) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 78.758878ms)
-Sep 21 16:10:15.263: INFO: (16) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 56.587029ms)
-Sep 21 16:10:15.311: INFO: (17) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 47.441295ms)
-Sep 21 16:10:15.356: INFO: (18) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 44.892814ms)
-Sep 21 16:10:15.582: INFO: (19) /api/v1/nodes/10.241.51.147/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 226.216392ms)
-[AfterEach] version v1
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:10:15.582: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "proxy-390" for this suite.
-•{"msg":"PASSED [sig-network] Proxy version v1 should proxy logs on node using proxy subresource  [Conformance]","total":280,"completed":76,"skipped":1323,"failed":0}
-SSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-cli] Kubectl client Guestbook application should create and stop a working application  [Conformance]","total":280,"completed":82,"skipped":1177,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-storage] Downward API volume 
-  should provide container's cpu request [NodeConformance] [Conformance]
+  should provide container's cpu limit [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:10:15.948: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 16:24:55.181: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename downward-api
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should provide container's cpu request [NodeConformance] [Conformance]
+[It] should provide container's cpu limit [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:10:16.626: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e6c5ae7a-e843-430c-8758-e29673896bee" in namespace "downward-api-8813" to be "success or failure"
-Sep 21 16:10:16.676: INFO: Pod "downwardapi-volume-e6c5ae7a-e843-430c-8758-e29673896bee": Phase="Pending", Reason="", readiness=false. Elapsed: 43.425677ms
-Sep 21 16:10:18.707: INFO: Pod "downwardapi-volume-e6c5ae7a-e843-430c-8758-e29673896bee": Phase="Pending", Reason="", readiness=false. Elapsed: 2.074172832s
-Sep 21 16:10:20.951: INFO: Pod "downwardapi-volume-e6c5ae7a-e843-430c-8758-e29673896bee": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.318308206s
+Oct 26 16:24:55.368: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9d7f984c-9147-4c78-a9fb-0962b82f871b" in namespace "downward-api-3585" to be "success or failure"
+Oct 26 16:24:55.389: INFO: Pod "downwardapi-volume-9d7f984c-9147-4c78-a9fb-0962b82f871b": Phase="Pending", Reason="", readiness=false. Elapsed: 21.159686ms
+Oct 26 16:24:57.415: INFO: Pod "downwardapi-volume-9d7f984c-9147-4c78-a9fb-0962b82f871b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.046696772s
+Oct 26 16:24:59.432: INFO: Pod "downwardapi-volume-9d7f984c-9147-4c78-a9fb-0962b82f871b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.064060741s
 STEP: Saw pod success
-Sep 21 16:10:20.951: INFO: Pod "downwardapi-volume-e6c5ae7a-e843-430c-8758-e29673896bee" satisfied condition "success or failure"
-Sep 21 16:10:20.981: INFO: Trying to get logs from node 10.241.51.150 pod downwardapi-volume-e6c5ae7a-e843-430c-8758-e29673896bee container client-container: 
+Oct 26 16:24:59.432: INFO: Pod "downwardapi-volume-9d7f984c-9147-4c78-a9fb-0962b82f871b" satisfied condition "success or failure"
+Oct 26 16:24:59.450: INFO: Trying to get logs from node 10.72.119.82 pod downwardapi-volume-9d7f984c-9147-4c78-a9fb-0962b82f871b container client-container: 
 STEP: delete the pod
-Sep 21 16:10:21.734: INFO: Waiting for pod downwardapi-volume-e6c5ae7a-e843-430c-8758-e29673896bee to disappear
-Sep 21 16:10:21.764: INFO: Pod downwardapi-volume-e6c5ae7a-e843-430c-8758-e29673896bee no longer exists
+Oct 26 16:24:59.561: INFO: Waiting for pod downwardapi-volume-9d7f984c-9147-4c78-a9fb-0962b82f871b to disappear
+Oct 26 16:24:59.576: INFO: Pod downwardapi-volume-9d7f984c-9147-4c78-a9fb-0962b82f871b no longer exists
 [AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:10:21.764: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-8813" for this suite.
-
-• [SLOW TEST:5.911 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:35
-  should provide container's cpu request [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance]","total":280,"completed":77,"skipped":1340,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:24:59.576: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-3585" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance]","total":280,"completed":83,"skipped":1206,"failed":0}
+SSSSS
 ------------------------------
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with projected pod [LinuxOnly] [Conformance]
+[sig-storage] Downward API volume 
+  should provide podname only [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Subpath
+[BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:10:21.861: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename subpath
+Oct 26 16:24:59.624: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename downward-api
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
-STEP: Setting up data
-[It] should support subpaths with projected pod [LinuxOnly] [Conformance]
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should provide podname only [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod pod-subpath-test-projected-qkht
-STEP: Creating a pod to test atomic-volume-subpath
-Sep 21 16:10:22.671: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-qkht" in namespace "subpath-427" to be "success or failure"
-Sep 21 16:10:22.704: INFO: Pod "pod-subpath-test-projected-qkht": Phase="Pending", Reason="", readiness=false. Elapsed: 33.490633ms
-Sep 21 16:10:24.734: INFO: Pod "pod-subpath-test-projected-qkht": Phase="Pending", Reason="", readiness=false. Elapsed: 2.063558616s
-Sep 21 16:10:26.761: INFO: Pod "pod-subpath-test-projected-qkht": Phase="Running", Reason="", readiness=true. Elapsed: 4.090461259s
-Sep 21 16:10:28.786: INFO: Pod "pod-subpath-test-projected-qkht": Phase="Running", Reason="", readiness=true. Elapsed: 6.115560703s
-Sep 21 16:10:30.812: INFO: Pod "pod-subpath-test-projected-qkht": Phase="Running", Reason="", readiness=true. Elapsed: 8.141652937s
-Sep 21 16:10:32.842: INFO: Pod "pod-subpath-test-projected-qkht": Phase="Running", Reason="", readiness=true. Elapsed: 10.171564312s
-Sep 21 16:10:34.869: INFO: Pod "pod-subpath-test-projected-qkht": Phase="Running", Reason="", readiness=true. Elapsed: 12.198415682s
-Sep 21 16:10:36.896: INFO: Pod "pod-subpath-test-projected-qkht": Phase="Running", Reason="", readiness=true. Elapsed: 14.225699437s
-Sep 21 16:10:38.927: INFO: Pod "pod-subpath-test-projected-qkht": Phase="Running", Reason="", readiness=true. Elapsed: 16.256765647s
-Sep 21 16:10:41.180: INFO: Pod "pod-subpath-test-projected-qkht": Phase="Running", Reason="", readiness=true. Elapsed: 18.509666625s
-Sep 21 16:10:43.217: INFO: Pod "pod-subpath-test-projected-qkht": Phase="Running", Reason="", readiness=true. Elapsed: 20.546696769s
-Sep 21 16:10:45.258: INFO: Pod "pod-subpath-test-projected-qkht": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.586925948s
+STEP: Creating a pod to test downward API volume plugin
+Oct 26 16:24:59.832: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fa2b066a-b388-48c2-9054-b4758064d0b2" in namespace "downward-api-1328" to be "success or failure"
+Oct 26 16:24:59.868: INFO: Pod "downwardapi-volume-fa2b066a-b388-48c2-9054-b4758064d0b2": Phase="Pending", Reason="", readiness=false. Elapsed: 35.981934ms
+Oct 26 16:25:01.885: INFO: Pod "downwardapi-volume-fa2b066a-b388-48c2-9054-b4758064d0b2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.053134707s
 STEP: Saw pod success
-Sep 21 16:10:45.258: INFO: Pod "pod-subpath-test-projected-qkht" satisfied condition "success or failure"
-Sep 21 16:10:45.288: INFO: Trying to get logs from node 10.241.51.150 pod pod-subpath-test-projected-qkht container test-container-subpath-projected-qkht: 
+Oct 26 16:25:01.886: INFO: Pod "downwardapi-volume-fa2b066a-b388-48c2-9054-b4758064d0b2" satisfied condition "success or failure"
+Oct 26 16:25:01.905: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-fa2b066a-b388-48c2-9054-b4758064d0b2 container client-container: 
 STEP: delete the pod
-Sep 21 16:10:45.731: INFO: Waiting for pod pod-subpath-test-projected-qkht to disappear
-Sep 21 16:10:45.764: INFO: Pod pod-subpath-test-projected-qkht no longer exists
-STEP: Deleting pod pod-subpath-test-projected-qkht
-Sep 21 16:10:45.764: INFO: Deleting pod "pod-subpath-test-projected-qkht" in namespace "subpath-427"
-[AfterEach] [sig-storage] Subpath
+Oct 26 16:25:02.015: INFO: Waiting for pod downwardapi-volume-fa2b066a-b388-48c2-9054-b4758064d0b2 to disappear
+Oct 26 16:25:02.033: INFO: Pod downwardapi-volume-fa2b066a-b388-48c2-9054-b4758064d0b2 no longer exists
+[AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:10:45.795: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "subpath-427" for this suite.
-
-• [SLOW TEST:24.002 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
-  Atomic writer volumes
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
-    should support subpaths with projected pod [LinuxOnly] [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Oct 26 16:25:02.033: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-1328" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance]","total":280,"completed":84,"skipped":1211,"failed":0}
+SSSSSSSSSS
 ------------------------------
-{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance]","total":280,"completed":78,"skipped":1363,"failed":0}
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  listing validating webhooks should work [Conformance]
+[k8s.io] [sig-node] Pods Extended [k8s.io] Delete Grace Period 
+  should be submitted and removed [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+[BeforeEach] [k8s.io] [sig-node] Pods Extended
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:10:45.864: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename webhook
+Oct 26 16:25:02.087: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename pods
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Sep 21 16:10:47.166: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301447, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301447, loc:(*time.Location)(0x7db5bc0)}}, Reason:"NewReplicaSetCreated", Message:"Created new replica set \"sample-webhook-deployment-5f65f8c764\""}, v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301447, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301447, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:10:49.205: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301447, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301447, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301447, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301447, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Sep 21 16:10:52.367: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] listing validating webhooks should work [Conformance]
+[BeforeEach] [k8s.io] Delete Grace Period
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:46
+[It] should be submitted and removed [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Listing all of the created validation webhooks
-STEP: Creating a configMap that does not comply to the validation webhook rules
-STEP: Deleting the collection of validation webhooks
-STEP: Creating a configMap that does not comply to the validation webhook rules
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+STEP: creating the pod
+STEP: setting up selector
+STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+Oct 26 16:25:06.392: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-871182394 proxy -p 0'
+STEP: deleting the pod gracefully
+STEP: verifying the kubelet observed the termination notice
+Oct 26 16:25:21.624: INFO: no pod exists with the name we were looking for, assuming the termination request was observed and completed
+[AfterEach] [k8s.io] [sig-node] Pods Extended
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:10:54.159: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-5838" for this suite.
-STEP: Destroying namespace "webhook-5838-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+Oct 26 16:25:21.643: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-3285" for this suite.
 
-• [SLOW TEST:8.917 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  listing validating webhooks should work [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:19.598 seconds]
+[k8s.io] [sig-node] Pods Extended
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  [k8s.io] Delete Grace Period
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+    should be submitted and removed [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance]","total":280,"completed":79,"skipped":1363,"failed":0}
-SSSSSSSSSSSSSS
+{"msg":"PASSED [k8s.io] [sig-node] Pods Extended [k8s.io] Delete Grace Period should be submitted and removed [Conformance]","total":280,"completed":85,"skipped":1221,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] Daemon set [Serial] 
-  should retry creating failed daemon pods [Conformance]
+[sig-storage] EmptyDir volumes 
+  should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Daemon set [Serial]
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:10:54.781: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename daemonsets
+Oct 26 16:25:21.688: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
-[It] should retry creating failed daemon pods [Conformance]
+[It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a simple DaemonSet "daemon-set"
-STEP: Check that daemon pods launch on every node of the cluster.
-Sep 21 16:10:55.703: INFO: Number of nodes with available pods: 0
-Sep 21 16:10:55.703: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:10:56.785: INFO: Number of nodes with available pods: 0
-Sep 21 16:10:56.785: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:10:57.771: INFO: Number of nodes with available pods: 1
-Sep 21 16:10:57.771: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:10:58.772: INFO: Number of nodes with available pods: 3
-Sep 21 16:10:58.772: INFO: Number of running nodes: 3, number of available pods: 3
-STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived.
-Sep 21 16:10:58.946: INFO: Number of nodes with available pods: 2
-Sep 21 16:10:58.946: INFO: Node 10.241.51.149 is running more than one daemon pod
-Sep 21 16:11:00.019: INFO: Number of nodes with available pods: 2
-Sep 21 16:11:00.019: INFO: Node 10.241.51.149 is running more than one daemon pod
-Sep 21 16:11:01.033: INFO: Number of nodes with available pods: 2
-Sep 21 16:11:01.033: INFO: Node 10.241.51.149 is running more than one daemon pod
-Sep 21 16:11:02.011: INFO: Number of nodes with available pods: 3
-Sep 21 16:11:02.011: INFO: Number of running nodes: 3, number of available pods: 3
-STEP: Wait for the failed daemon pod to be completely deleted.
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6060, will wait for the garbage collector to delete the pods
-Sep 21 16:11:02.189: INFO: Deleting DaemonSet.extensions daemon-set took: 62.430032ms
-Sep 21 16:11:02.290: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.253225ms
-Sep 21 16:11:12.416: INFO: Number of nodes with available pods: 0
-Sep 21 16:11:12.416: INFO: Number of running nodes: 0, number of available pods: 0
-Sep 21 16:11:12.578: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-6060/daemonsets","resourceVersion":"59172"},"items":null}
-
-Sep 21 16:11:12.853: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-6060/pods","resourceVersion":"59172"},"items":null}
-
-[AfterEach] [sig-apps] Daemon set [Serial]
+STEP: Creating a pod to test emptydir 0777 on tmpfs
+Oct 26 16:25:21.915: INFO: Waiting up to 5m0s for pod "pod-d0c926d2-86ce-455c-adfe-897b6d8525c2" in namespace "emptydir-1440" to be "success or failure"
+Oct 26 16:25:21.931: INFO: Pod "pod-d0c926d2-86ce-455c-adfe-897b6d8525c2": Phase="Pending", Reason="", readiness=false. Elapsed: 15.176547ms
+Oct 26 16:25:23.945: INFO: Pod "pod-d0c926d2-86ce-455c-adfe-897b6d8525c2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030083385s
+Oct 26 16:25:25.961: INFO: Pod "pod-d0c926d2-86ce-455c-adfe-897b6d8525c2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046125052s
+STEP: Saw pod success
+Oct 26 16:25:25.962: INFO: Pod "pod-d0c926d2-86ce-455c-adfe-897b6d8525c2" satisfied condition "success or failure"
+Oct 26 16:25:25.976: INFO: Trying to get logs from node 10.72.119.74 pod pod-d0c926d2-86ce-455c-adfe-897b6d8525c2 container test-container: 
+STEP: delete the pod
+Oct 26 16:25:26.057: INFO: Waiting for pod pod-d0c926d2-86ce-455c-adfe-897b6d8525c2 to disappear
+Oct 26 16:25:26.071: INFO: Pod pod-d0c926d2-86ce-455c-adfe-897b6d8525c2 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:11:13.026: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "daemonsets-6060" for this suite.
-
-• [SLOW TEST:18.319 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should retry creating failed daemon pods [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance]","total":280,"completed":80,"skipped":1377,"failed":0}
-SS
+Oct 26 16:25:26.071: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-1440" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":86,"skipped":1252,"failed":0}
+SSSSS
 ------------------------------
 [sig-api-machinery] ResourceQuota 
-  should create a ResourceQuota and capture the life of a service. [Conformance]
+  should create a ResourceQuota and capture the life of a replication controller. [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] ResourceQuota
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:11:13.100: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 16:25:26.110: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename resourcequota
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should create a ResourceQuota and capture the life of a service. [Conformance]
+[It] should create a ResourceQuota and capture the life of a replication controller. [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Counting existing ResourceQuota
 STEP: Creating a ResourceQuota
 STEP: Ensuring resource quota status is calculated
-STEP: Creating a Service
-STEP: Ensuring resource quota status captures service creation
-STEP: Deleting a Service
+STEP: Creating a ReplicationController
+STEP: Ensuring resource quota status captures replication controller creation
+STEP: Deleting a ReplicationController
 STEP: Ensuring resource quota status released usage
 [AfterEach] [sig-api-machinery] ResourceQuota
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:11:24.946: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-899" for this suite.
+Oct 26 16:25:37.437: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-4915" for this suite.
 
-• [SLOW TEST:11.931 seconds]
+• [SLOW TEST:11.388 seconds]
 [sig-api-machinery] ResourceQuota
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should create a ResourceQuota and capture the life of a service. [Conformance]
+  should create a ResourceQuota and capture the life of a replication controller. [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance]","total":280,"completed":81,"skipped":1379,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance]","total":280,"completed":87,"skipped":1257,"failed":0}
+SSSSSSSSSSSSS
 ------------------------------
 [k8s.io] Pods 
-  should support remote command execution over websockets [NodeConformance] [Conformance]
+  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [k8s.io] Pods
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:11:25.035: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 16:25:37.499: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename pods
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [k8s.io] Pods
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
-[It] should support remote command execution over websockets [NodeConformance] [Conformance]
+[It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:11:25.311: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
 STEP: creating the pod
 STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+STEP: updating the pod
+Oct 26 16:25:42.331: INFO: Successfully updated pod "pod-update-activedeadlineseconds-74a2e837-c84c-4b29-b76f-55076a2c4d0d"
+Oct 26 16:25:42.331: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-74a2e837-c84c-4b29-b76f-55076a2c4d0d" in namespace "pods-4286" to be "terminated due to deadline exceeded"
+Oct 26 16:25:42.345: INFO: Pod "pod-update-activedeadlineseconds-74a2e837-c84c-4b29-b76f-55076a2c4d0d": Phase="Running", Reason="", readiness=true. Elapsed: 13.834617ms
+Oct 26 16:25:44.360: INFO: Pod "pod-update-activedeadlineseconds-74a2e837-c84c-4b29-b76f-55076a2c4d0d": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 2.029578098s
+Oct 26 16:25:44.360: INFO: Pod "pod-update-activedeadlineseconds-74a2e837-c84c-4b29-b76f-55076a2c4d0d" satisfied condition "terminated due to deadline exceeded"
 [AfterEach] [k8s.io] Pods
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:11:29.877: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-3264" for this suite.
-•{"msg":"PASSED [k8s.io] Pods should support remote command execution over websockets [NodeConformance] [Conformance]","total":280,"completed":82,"skipped":1428,"failed":0}
-SSSSSSSS
-------------------------------
-[sig-api-machinery] ResourceQuota 
-  should create a ResourceQuota and capture the life of a replica set. [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:11:29.955: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename resourcequota
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should create a ResourceQuota and capture the life of a replica set. [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Counting existing ResourceQuota
-STEP: Creating a ResourceQuota
-STEP: Ensuring resource quota status is calculated
-STEP: Creating a ReplicaSet
-STEP: Ensuring resource quota status captures replicaset creation
-STEP: Deleting a ReplicaSet
-STEP: Ensuring resource quota status released usage
-[AfterEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:11:41.559: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-1373" for this suite.
-
-• [SLOW TEST:11.750 seconds]
-[sig-api-machinery] ResourceQuota
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should create a ResourceQuota and capture the life of a replica set. [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance]","total":280,"completed":83,"skipped":1436,"failed":0}
-SSSSSSSSSSSSSSSSSS
-------------------------------
-[k8s.io] Probing container 
-  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:11:41.706: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-probe
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:12:42.184: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-943" for this suite.
+Oct 26 16:25:44.361: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-4286" for this suite.
 
-• [SLOW TEST:60.566 seconds]
-[k8s.io] Probing container
+• [SLOW TEST:6.909 seconds]
+[k8s.io] Pods
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]","total":280,"completed":84,"skipped":1454,"failed":0}
-SSSSSS
+{"msg":"PASSED [k8s.io] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]","total":280,"completed":88,"skipped":1270,"failed":0}
+SSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should include webhook resources in discovery documents [Conformance]
+[k8s.io] Docker Containers 
+  should be able to override the image's default command and arguments [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+[BeforeEach] [k8s.io] Docker Containers
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:12:42.273: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename webhook
+Oct 26 16:25:44.411: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename containers
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Sep 21 16:12:43.830: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Sep 21 16:12:45.913: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301563, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301563, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301563, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301563, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:12:47.968: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301563, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301563, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301563, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301563, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Sep 21 16:12:51.049: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should include webhook resources in discovery documents [Conformance]
+[It] should be able to override the image's default command and arguments [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: fetching the /apis discovery document
-STEP: finding the admissionregistration.k8s.io API group in the /apis discovery document
-STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis discovery document
-STEP: fetching the /apis/admissionregistration.k8s.io discovery document
-STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis/admissionregistration.k8s.io discovery document
-STEP: fetching the /apis/admissionregistration.k8s.io/v1 discovery document
-STEP: finding mutatingwebhookconfigurations and validatingwebhookconfigurations resources in the /apis/admissionregistration.k8s.io/v1 discovery document
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+STEP: Creating a pod to test override all
+Oct 26 16:25:45.625: INFO: Waiting up to 5m0s for pod "client-containers-df4769b2-6a21-44cd-9b62-c713c2024731" in namespace "containers-5959" to be "success or failure"
+Oct 26 16:25:45.642: INFO: Pod "client-containers-df4769b2-6a21-44cd-9b62-c713c2024731": Phase="Pending", Reason="", readiness=false. Elapsed: 16.119313ms
+Oct 26 16:25:47.659: INFO: Pod "client-containers-df4769b2-6a21-44cd-9b62-c713c2024731": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.033594373s
+STEP: Saw pod success
+Oct 26 16:25:47.659: INFO: Pod "client-containers-df4769b2-6a21-44cd-9b62-c713c2024731" satisfied condition "success or failure"
+Oct 26 16:25:47.674: INFO: Trying to get logs from node 10.72.119.74 pod client-containers-df4769b2-6a21-44cd-9b62-c713c2024731 container test-container: 
+STEP: delete the pod
+Oct 26 16:25:47.765: INFO: Waiting for pod client-containers-df4769b2-6a21-44cd-9b62-c713c2024731 to disappear
+Oct 26 16:25:47.779: INFO: Pod client-containers-df4769b2-6a21-44cd-9b62-c713c2024731 no longer exists
+[AfterEach] [k8s.io] Docker Containers
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:12:51.090: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-4243" for this suite.
-STEP: Destroying namespace "webhook-4243-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
-
-• [SLOW TEST:9.474 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should include webhook resources in discovery documents [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance]","total":280,"completed":85,"skipped":1460,"failed":0}
+Oct 26 16:25:47.779: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "containers-5959" for this suite.
+•{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance]","total":280,"completed":89,"skipped":1284,"failed":0}
 SSSSSS
 ------------------------------
-[sig-apps] Deployment 
-  deployment should delete old replica sets [Conformance]
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates that NodeSelector is respected if not matching  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Deployment
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:12:51.747: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename deployment
+Oct 26 16:25:47.813: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename sched-pred
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
-[It] deployment should delete old replica sets [Conformance]
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
+Oct 26 16:25:47.963: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Oct 26 16:25:48.016: INFO: Waiting for terminating namespaces to be deleted...
+Oct 26 16:25:48.035: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.74 before test
+Oct 26 16:25:48.119: INFO: ibm-keepalived-watcher-49jbs from kube-system started at 2020-10-26 14:34:28 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.120: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 16:25:48.120: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-nbr6w from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.120: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 16:25:48.120: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 16:25:48.120: INFO: service-ca-7d9db9c5b8-4w4cj from openshift-service-ca started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.120: INFO: 	Container service-ca-controller ready: true, restart count 0
+Oct 26 16:25:48.120: INFO: openshift-kube-proxy-7d26g from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.120: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 16:25:48.120: INFO: tuned-p4cpc from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.120: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 16:25:48.120: INFO: node-ca-4ppdl from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.120: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 16:25:48.120: INFO: tigera-operator-5c45c894b4-z9bnx from tigera-operator started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.120: INFO: 	Container tigera-operator ready: true, restart count 0
+Oct 26 16:25:48.120: INFO: ibm-master-proxy-static-10.72.119.74 from kube-system started at 2020-10-26 14:34:26 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.121: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 16:25:48.121: INFO: 	Container pause ready: true, restart count 0
+Oct 26 16:25:48.121: INFO: multus-cc44t from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.121: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 16:25:48.121: INFO: alertmanager-main-0 from openshift-monitoring started at 2020-10-26 16:09:57 +0000 UTC (3 container statuses recorded)
+Oct 26 16:25:48.121: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 16:25:48.121: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 16:25:48.121: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 16:25:48.121: INFO: multus-admission-controller-dnt9r from openshift-multus started at 2020-10-26 16:10:16 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.121: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.121: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 16:25:48.121: INFO: redhat-operators-5b446b45df-fgb5b from openshift-marketplace started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.121: INFO: 	Container redhat-operators ready: true, restart count 0
+Oct 26 16:25:48.121: INFO: packageserver-5f76c64df7-jg7sj from openshift-operator-lifecycle-manager started at 2020-10-26 16:10:08 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.121: INFO: 	Container packageserver ready: true, restart count 0
+Oct 26 16:25:48.121: INFO: vpn-5b455f6f88-64q7j from kube-system started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.122: INFO: 	Container vpn ready: true, restart count 0
+Oct 26 16:25:48.122: INFO: pod-update-activedeadlineseconds-74a2e837-c84c-4b29-b76f-55076a2c4d0d from pods-4286 started at 2020-10-26 16:25:37 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.122: INFO: 	Container nginx ready: false, restart count 0
+Oct 26 16:25:48.122: INFO: calico-typha-7fd49c779d-46t4v from calico-system started at 2020-10-26 14:37:36 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.122: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 16:25:48.122: INFO: sonobuoy from sonobuoy started at 2020-10-26 15:59:53 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.122: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Oct 26 16:25:48.122: INFO: sonobuoy-e2e-job-c6d919d638514ce4 from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.122: INFO: 	Container e2e ready: true, restart count 0
+Oct 26 16:25:48.122: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 16:25:48.122: INFO: thanos-querier-84d7d58bc-c27ss from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (4 container statuses recorded)
+Oct 26 16:25:48.122: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.122: INFO: 	Container oauth-proxy ready: true, restart count 0
+Oct 26 16:25:48.122: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 16:25:48.122: INFO: 	Container thanos-querier ready: true, restart count 0
+Oct 26 16:25:48.123: INFO: grafana-f999495bf-t4tcv from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.123: INFO: 	Container grafana ready: true, restart count 0
+Oct 26 16:25:48.123: INFO: 	Container grafana-proxy ready: true, restart count 0
+Oct 26 16:25:48.123: INFO: telemeter-client-d5d9896c4-jpd8z from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (3 container statuses recorded)
+Oct 26 16:25:48.123: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.123: INFO: 	Container reload ready: true, restart count 0
+Oct 26 16:25:48.123: INFO: 	Container telemeter-client ready: true, restart count 0
+Oct 26 16:25:48.123: INFO: ibmcloud-block-storage-driver-ws85j from kube-system started at 2020-10-26 14:34:32 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.123: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 16:25:48.123: INFO: node-exporter-ls6pf from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.123: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.123: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 16:25:48.123: INFO: dns-default-s9kvh from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.123: INFO: 	Container dns ready: true, restart count 0
+Oct 26 16:25:48.123: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 16:25:48.123: INFO: prometheus-adapter-674456c5d6-p6bj5 from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.124: INFO: 	Container prometheus-adapter ready: true, restart count 0
+Oct 26 16:25:48.124: INFO: console-5dd98d6f6f-zxw5x from openshift-console started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.124: INFO: 	Container console ready: true, restart count 0
+Oct 26 16:25:48.124: INFO: calico-node-xlzf4 from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.124: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 16:25:48.124: INFO: cluster-samples-operator-848994bd88-rw4rq from openshift-cluster-samples-operator started at 2020-10-26 16:09:41 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.124: INFO: 	Container cluster-samples-operator ready: true, restart count 0
+Oct 26 16:25:48.124: INFO: 	Container cluster-samples-operator-watch ready: true, restart count 0
+Oct 26 16:25:48.124: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.76 before test
+Oct 26 16:25:48.285: INFO: node-ca-bpxm5 from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.285: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 16:25:48.285: INFO: ibm-cloud-provider-ip-158-176-78-170-5cdb49c95f-76fjw from ibm-system started at 2020-10-26 14:42:32 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.285: INFO: 	Container ibm-cloud-provider-ip-158-176-78-170 ready: true, restart count 0
+Oct 26 16:25:48.285: INFO: ingress-operator-65878dcc8d-wdnrl from openshift-ingress-operator started at 2020-10-26 14:35:49 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.285: INFO: 	Container ingress-operator ready: true, restart count 0
+Oct 26 16:25:48.285: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.285: INFO: downloads-74f6b6dcb6-r9ksv from openshift-console started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.285: INFO: 	Container download-server ready: true, restart count 0
+Oct 26 16:25:48.286: INFO: multus-admission-controller-5j529 from openshift-multus started at 2020-10-26 14:35:50 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.286: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.286: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 16:25:48.286: INFO: node-exporter-zmwjz from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.286: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.286: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 16:25:48.286: INFO: calico-kube-controllers-cdc7f7f8f-jqklm from calico-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.286: INFO: 	Container calico-kube-controllers ready: true, restart count 0
+Oct 26 16:25:48.286: INFO: downloads-74f6b6dcb6-vqzf9 from openshift-console started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.286: INFO: 	Container download-server ready: true, restart count 0
+Oct 26 16:25:48.286: INFO: cluster-node-tuning-operator-575b98978c-zz7s9 from openshift-cluster-node-tuning-operator started at 2020-10-26 14:35:51 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.287: INFO: 	Container cluster-node-tuning-operator ready: true, restart count 0
+Oct 26 16:25:48.287: INFO: calico-node-dhd7p from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.287: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 16:25:48.287: INFO: kube-storage-version-migrator-operator-86d479ffb5-9cmml from openshift-kube-storage-version-migrator-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.287: INFO: 	Container kube-storage-version-migrator-operator ready: true, restart count 1
+Oct 26 16:25:48.287: INFO: catalog-operator-5db6d5849c-z2lgx from openshift-operator-lifecycle-manager started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.287: INFO: 	Container catalog-operator ready: true, restart count 0
+Oct 26 16:25:48.287: INFO: ibm-keepalived-watcher-qwg48 from kube-system started at 2020-10-26 14:34:30 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.287: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 16:25:48.287: INFO: ibmcloud-block-storage-plugin-5f486d686f-6brjv from kube-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.287: INFO: 	Container ibmcloud-block-storage-plugin-container ready: true, restart count 0
+Oct 26 16:25:48.287: INFO: ibm-master-proxy-static-10.72.119.76 from kube-system started at 2020-10-26 14:34:28 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.287: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 16:25:48.287: INFO: 	Container pause ready: true, restart count 0
+Oct 26 16:25:48.287: INFO: cluster-image-registry-operator-8444cfb976-gdjk6 from openshift-image-registry started at 2020-10-26 14:35:50 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.287: INFO: 	Container cluster-image-registry-operator ready: true, restart count 0
+Oct 26 16:25:48.288: INFO: 	Container cluster-image-registry-operator-watch ready: true, restart count 0
+Oct 26 16:25:48.288: INFO: prometheus-k8s-0 from openshift-monitoring started at 2020-10-26 14:45:39 +0000 UTC (7 container statuses recorded)
+Oct 26 16:25:48.288: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.288: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 16:25:48.288: INFO: 	Container prometheus ready: true, restart count 1
+Oct 26 16:25:48.288: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
+Oct 26 16:25:48.288: INFO: 	Container prometheus-proxy ready: true, restart count 0
+Oct 26 16:25:48.288: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
+Oct 26 16:25:48.288: INFO: 	Container thanos-sidecar ready: true, restart count 0
+Oct 26 16:25:48.288: INFO: marketplace-operator-5df598b96b-6vr64 from openshift-marketplace started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.288: INFO: 	Container marketplace-operator ready: true, restart count 0
+Oct 26 16:25:48.288: INFO: console-operator-5fc674d8d4-fjctd from openshift-console-operator started at 2020-10-26 14:35:51 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.288: INFO: 	Container console-operator ready: true, restart count 1
+Oct 26 16:25:48.288: INFO: tuned-s6flg from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.288: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 16:25:48.288: INFO: network-operator-86b67c6f7f-pcq2w from openshift-network-operator started at 2020-10-26 14:34:37 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.289: INFO: 	Container network-operator ready: true, restart count 0
+Oct 26 16:25:48.289: INFO: dns-operator-58d86f779c-wbm98 from openshift-dns-operator started at 2020-10-26 14:35:49 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.289: INFO: 	Container dns-operator ready: true, restart count 0
+Oct 26 16:25:48.289: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.289: INFO: cluster-storage-operator-9b57569c9-gpqd2 from openshift-cluster-storage-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.289: INFO: 	Container cluster-storage-operator ready: true, restart count 0
+Oct 26 16:25:48.289: INFO: service-ca-operator-656f67898-svtt9 from openshift-service-ca-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.289: INFO: 	Container operator ready: true, restart count 1
+Oct 26 16:25:48.289: INFO: openshift-kube-proxy-74ss4 from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.289: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 16:25:48.289: INFO: ibm-file-plugin-944b59c94-k7mrm from kube-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.289: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
+Oct 26 16:25:48.289: INFO: calico-typha-7fd49c779d-766qb from calico-system started at 2020-10-26 14:37:36 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.289: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 16:25:48.289: INFO: dns-default-z8pn9 from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.289: INFO: 	Container dns ready: true, restart count 0
+Oct 26 16:25:48.290: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 16:25:48.290: INFO: router-default-84766ff4fd-cftxw from openshift-ingress started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.290: INFO: 	Container router ready: true, restart count 0
+Oct 26 16:25:48.290: INFO: ibmcloud-block-storage-driver-b724k from kube-system started at 2020-10-26 14:34:34 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.290: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 16:25:48.290: INFO: openshift-service-catalog-apiserver-operator-7cfbcc49dc-hrn66 from openshift-service-catalog-apiserver-operator started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.290: INFO: 	Container operator ready: true, restart count 1
+Oct 26 16:25:48.290: INFO: alertmanager-main-2 from openshift-monitoring started at 2020-10-26 14:44:28 +0000 UTC (3 container statuses recorded)
+Oct 26 16:25:48.290: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 16:25:48.290: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 16:25:48.290: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 16:25:48.290: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-2scb4 from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.290: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 16:25:48.290: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 16:25:48.290: INFO: cluster-monitoring-operator-69649564f8-r4h9b from openshift-monitoring started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.290: INFO: 	Container cluster-monitoring-operator ready: true, restart count 0
+Oct 26 16:25:48.290: INFO: multus-lbpcn from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.291: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 16:25:48.291: INFO: openshift-service-catalog-controller-manager-operator-d49bz89wp from openshift-service-catalog-controller-manager-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.291: INFO: 	Container operator ready: true, restart count 1
+Oct 26 16:25:48.291: INFO: ibm-storage-watcher-77d6cc685c-bvg8q from kube-system started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.291: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
+Oct 26 16:25:48.291: INFO: olm-operator-7cd6cbf579-zww2l from openshift-operator-lifecycle-manager started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.291: INFO: 	Container olm-operator ready: true, restart count 0
+Oct 26 16:25:48.291: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.82 before test
+Oct 26 16:25:48.399: INFO: multus-admission-controller-9rv7d from openshift-multus started at 2020-10-26 14:36:02 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.399: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.399: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 16:25:48.399: INFO: ibm-cloud-provider-ip-158-176-78-170-5cdb49c95f-89x4j from ibm-system started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.399: INFO: 	Container ibm-cloud-provider-ip-158-176-78-170 ready: true, restart count 0
+Oct 26 16:25:48.399: INFO: console-5dd98d6f6f-xvz7c from openshift-console started at 2020-10-26 14:45:54 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.399: INFO: 	Container console ready: true, restart count 0
+Oct 26 16:25:48.399: INFO: packageserver-5f76c64df7-6nrvc from openshift-operator-lifecycle-manager started at 2020-10-26 16:10:11 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.399: INFO: 	Container packageserver ready: true, restart count 0
+Oct 26 16:25:48.400: INFO: calico-node-c7wjs from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.400: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 16:25:48.400: INFO: dns-default-ggg6q from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.400: INFO: 	Container dns ready: true, restart count 0
+Oct 26 16:25:48.400: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 16:25:48.400: INFO: prometheus-k8s-1 from openshift-monitoring started at 2020-10-26 14:45:29 +0000 UTC (7 container statuses recorded)
+Oct 26 16:25:48.400: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.400: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 16:25:48.400: INFO: 	Container prometheus ready: true, restart count 1
+Oct 26 16:25:48.400: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
+Oct 26 16:25:48.400: INFO: 	Container prometheus-proxy ready: true, restart count 0
+Oct 26 16:25:48.400: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
+Oct 26 16:25:48.400: INFO: 	Container thanos-sidecar ready: true, restart count 0
+Oct 26 16:25:48.400: INFO: kube-state-metrics-7bfb6f7864-vmgz9 from openshift-monitoring started at 2020-10-26 14:36:32 +0000 UTC (3 container statuses recorded)
+Oct 26 16:25:48.400: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
+Oct 26 16:25:48.401: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
+Oct 26 16:25:48.401: INFO: 	Container kube-state-metrics ready: true, restart count 0
+Oct 26 16:25:48.401: INFO: thanos-querier-84d7d58bc-mpnzq from openshift-monitoring started at 2020-10-26 14:45:18 +0000 UTC (4 container statuses recorded)
+Oct 26 16:25:48.401: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.401: INFO: 	Container oauth-proxy ready: true, restart count 0
+Oct 26 16:25:48.401: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 16:25:48.401: INFO: 	Container thanos-querier ready: true, restart count 0
+Oct 26 16:25:48.401: INFO: community-operators-f4df99f64-twf5k from openshift-marketplace started at 2020-10-26 14:38:39 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.401: INFO: 	Container community-operators ready: true, restart count 0
+Oct 26 16:25:48.401: INFO: ibm-master-proxy-static-10.72.119.82 from kube-system started at 2020-10-26 14:34:29 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.401: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 16:25:48.401: INFO: 	Container pause ready: true, restart count 0
+Oct 26 16:25:48.401: INFO: ibmcloud-block-storage-driver-sxlj2 from kube-system started at 2020-10-26 14:34:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.401: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 16:25:48.402: INFO: multus-hwkgh from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.402: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 16:25:48.402: INFO: node-ca-z564l from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.402: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 16:25:48.402: INFO: redhat-marketplace-77d9b69f69-qb5s9 from openshift-marketplace started at 2020-10-26 14:38:39 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.402: INFO: 	Container redhat-marketplace ready: true, restart count 0
+Oct 26 16:25:48.402: INFO: alertmanager-main-1 from openshift-monitoring started at 2020-10-26 14:44:39 +0000 UTC (3 container statuses recorded)
+Oct 26 16:25:48.402: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 16:25:48.402: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 16:25:48.402: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 16:25:48.402: INFO: ibm-keepalived-watcher-98n85 from kube-system started at 2020-10-26 14:34:32 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.402: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 16:25:48.402: INFO: tuned-h578f from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.403: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 16:25:48.403: INFO: certified-operators-69bb9f65cc-9mhpt from openshift-marketplace started at 2020-10-26 14:38:38 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.403: INFO: 	Container certified-operators ready: true, restart count 0
+Oct 26 16:25:48.403: INFO: calico-typha-7fd49c779d-nmzbk from calico-system started at 2020-10-26 14:35:39 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.403: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 16:25:48.403: INFO: test-k8s-e2e-pvg-master-verification from default started at 2020-10-26 14:37:46 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.403: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
+Oct 26 16:25:48.403: INFO: registry-pvc-permissions-m9wwv from openshift-image-registry started at 2020-10-26 14:40:09 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.403: INFO: 	Container pvc-permissions ready: false, restart count 0
+Oct 26 16:25:48.403: INFO: migrator-686fc6cc66-9wpj9 from openshift-kube-storage-version-migrator started at 2020-10-26 14:36:53 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.403: INFO: 	Container migrator ready: true, restart count 0
+Oct 26 16:25:48.403: INFO: prometheus-adapter-674456c5d6-kj77x from openshift-monitoring started at 2020-10-26 14:44:24 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.403: INFO: 	Container prometheus-adapter ready: true, restart count 0
+Oct 26 16:25:48.403: INFO: image-registry-6db5c967cf-d4ljt from openshift-image-registry started at 2020-10-26 14:40:09 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.403: INFO: 	Container registry ready: true, restart count 0
+Oct 26 16:25:48.403: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-x2lkl from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.404: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 16:25:48.404: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 16:25:48.404: INFO: openshift-kube-proxy-z9sjt from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.404: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 16:25:48.404: INFO: openshift-state-metrics-d4d5d6f57-lszmv from openshift-monitoring started at 2020-10-26 14:36:34 +0000 UTC (3 container statuses recorded)
+Oct 26 16:25:48.404: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
+Oct 26 16:25:48.404: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
+Oct 26 16:25:48.404: INFO: 	Container openshift-state-metrics ready: true, restart count 0
+Oct 26 16:25:48.404: INFO: node-exporter-4lfg7 from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:48.404: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:48.404: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 16:25:48.404: INFO: prometheus-operator-55f946f5d-tpnd8 from openshift-monitoring started at 2020-10-26 14:44:11 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.404: INFO: 	Container prometheus-operator ready: true, restart count 0
+Oct 26 16:25:48.404: INFO: router-default-84766ff4fd-7dgrd from openshift-ingress started at 2020-10-26 14:38:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:48.405: INFO: 	Container router ready: true, restart count 0
+[It] validates that NodeSelector is respected if not matching  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:12:52.222: INFO: Pod name cleanup-pod: Found 1 pods out of 1
-STEP: ensuring each pod is running
-Sep 21 16:12:56.281: INFO: Creating deployment test-cleanup-deployment
-STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
-Sep 21 16:13:00.520: INFO: Deployment "test-cleanup-deployment":
-&Deployment{ObjectMeta:{test-cleanup-deployment  deployment-6309 /apis/apps/v1/namespaces/deployment-6309/deployments/test-cleanup-deployment 3da362a4-3c75-4efe-9865-cd452ee4edac 60164 1 2020-09-21 16:12:56 +0000 UTC   map[name:cleanup-pod] map[deployment.kubernetes.io/revision:1] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:cleanup-pod] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc001f7d0f8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-09-21 16:12:56 +0000 UTC,LastTransitionTime:2020-09-21 16:12:56 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-cleanup-deployment-55ffc6b7b6" has successfully progressed.,LastUpdateTime:2020-09-21 16:12:58 +0000 UTC,LastTransitionTime:2020-09-21 16:12:56 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},}
-
-Sep 21 16:13:00.542: INFO: New ReplicaSet "test-cleanup-deployment-55ffc6b7b6" of Deployment "test-cleanup-deployment":
-&ReplicaSet{ObjectMeta:{test-cleanup-deployment-55ffc6b7b6  deployment-6309 /apis/apps/v1/namespaces/deployment-6309/replicasets/test-cleanup-deployment-55ffc6b7b6 605dd543-27fd-4f0c-aa2c-8425610cbb18 60154 1 2020-09-21 16:12:56 +0000 UTC   map[name:cleanup-pod pod-template-hash:55ffc6b7b6] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-cleanup-deployment 3da362a4-3c75-4efe-9865-cd452ee4edac 0xc0036ac2c7 0xc0036ac2c8}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod-template-hash: 55ffc6b7b6,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:cleanup-pod pod-template-hash:55ffc6b7b6] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0036ac338  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},}
-Sep 21 16:13:00.569: INFO: Pod "test-cleanup-deployment-55ffc6b7b6-v9dfb" is available:
-&Pod{ObjectMeta:{test-cleanup-deployment-55ffc6b7b6-v9dfb test-cleanup-deployment-55ffc6b7b6- deployment-6309 /api/v1/namespaces/deployment-6309/pods/test-cleanup-deployment-55ffc6b7b6-v9dfb c366725d-ecb5-4b39-b82b-b563574aa65b 60153 0 2020-09-21 16:12:56 +0000 UTC   map[name:cleanup-pod pod-template-hash:55ffc6b7b6] map[cni.projectcalico.org/podIP:172.30.237.101/32 cni.projectcalico.org/podIPs:172.30.237.101/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.101"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet test-cleanup-deployment-55ffc6b7b6 605dd543-27fd-4f0c-aa2c-8425610cbb18 0xc0036ac6e7 0xc0036ac6e8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-pdppp,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-pdppp,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-pdppp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-2zzfc,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:12:56 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:12:58 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:12:58 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:12:56 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:172.30.237.101,StartTime:2020-09-21 16:12:56 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:12:58 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:gcr.io/kubernetes-e2e-test-images/agnhost@sha256:02b9e10ea1f7439a03d9bc3ee16d984172ca7eac1818792ad4d721c5c8f72ff4,ContainerID:cri-o://aed2e91b55ecbfb6671ea75dd85f051b5bd9479441d47db46fb028cfe140219a,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.237.101,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-[AfterEach] [sig-apps] Deployment
+STEP: Trying to schedule Pod with nonempty NodeSelector.
+STEP: Considering event: 
+Type = [Warning], Name = [restricted-pod.1641978707fa76d2], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match node selector.]
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:13:00.569: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "deployment-6309" for this suite.
-
-• [SLOW TEST:8.908 seconds]
-[sig-apps] Deployment
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  deployment should delete old replica sets [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] Deployment deployment should delete old replica sets [Conformance]","total":280,"completed":86,"skipped":1466,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:25:49.549: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-9780" for this suite.
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
+•{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching  [Conformance]","total":280,"completed":90,"skipped":1290,"failed":0}
+SS
 ------------------------------
-[sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial] 
-  evicts pods with minTolerationSeconds [Disruptive] [Conformance]
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:13:00.656: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename taint-multiple-pods
+Oct 26 16:25:49.587: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename sched-pred
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/taints.go:344
-Sep 21 16:13:00.881: INFO: Waiting up to 1m0s for all nodes to be ready
-Sep 21 16:14:01.029: INFO: Waiting for terminating namespaces to be deleted...
-[It] evicts pods with minTolerationSeconds [Disruptive] [Conformance]
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
+Oct 26 16:25:49.752: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Oct 26 16:25:49.799: INFO: Waiting for terminating namespaces to be deleted...
+Oct 26 16:25:49.813: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.74 before test
+Oct 26 16:25:49.875: INFO: service-ca-7d9db9c5b8-4w4cj from openshift-service-ca started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container service-ca-controller ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: ibm-keepalived-watcher-49jbs from kube-system started at 2020-10-26 14:34:28 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-nbr6w from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: openshift-kube-proxy-7d26g from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: tuned-p4cpc from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: node-ca-4ppdl from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: redhat-operators-5b446b45df-fgb5b from openshift-marketplace started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container redhat-operators ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: packageserver-5f76c64df7-jg7sj from openshift-operator-lifecycle-manager started at 2020-10-26 16:10:08 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container packageserver ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: vpn-5b455f6f88-64q7j from kube-system started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container vpn ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: tigera-operator-5c45c894b4-z9bnx from tigera-operator started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container tigera-operator ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: ibm-master-proxy-static-10.72.119.74 from kube-system started at 2020-10-26 14:34:26 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: 	Container pause ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: multus-cc44t from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: alertmanager-main-0 from openshift-monitoring started at 2020-10-26 16:09:57 +0000 UTC (3 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: multus-admission-controller-dnt9r from openshift-multus started at 2020-10-26 16:10:16 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: pod-update-activedeadlineseconds-74a2e837-c84c-4b29-b76f-55076a2c4d0d from pods-4286 started at 2020-10-26 16:25:37 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container nginx ready: false, restart count 0
+Oct 26 16:25:49.875: INFO: thanos-querier-84d7d58bc-c27ss from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (4 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: 	Container oauth-proxy ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: 	Container thanos-querier ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: grafana-f999495bf-t4tcv from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container grafana ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: 	Container grafana-proxy ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: telemeter-client-d5d9896c4-jpd8z from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (3 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: 	Container reload ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: 	Container telemeter-client ready: true, restart count 0
+Oct 26 16:25:49.875: INFO: calico-typha-7fd49c779d-46t4v from calico-system started at 2020-10-26 14:37:36 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.875: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: sonobuoy from sonobuoy started at 2020-10-26 15:59:53 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.876: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: sonobuoy-e2e-job-c6d919d638514ce4 from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.876: INFO: 	Container e2e ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: prometheus-adapter-674456c5d6-p6bj5 from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.876: INFO: 	Container prometheus-adapter ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: console-5dd98d6f6f-zxw5x from openshift-console started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.876: INFO: 	Container console ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: ibmcloud-block-storage-driver-ws85j from kube-system started at 2020-10-26 14:34:32 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.876: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: node-exporter-ls6pf from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.876: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: dns-default-s9kvh from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.876: INFO: 	Container dns ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: cluster-samples-operator-848994bd88-rw4rq from openshift-cluster-samples-operator started at 2020-10-26 16:09:41 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.876: INFO: 	Container cluster-samples-operator ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: 	Container cluster-samples-operator-watch ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: calico-node-xlzf4 from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.876: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 16:25:49.876: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.76 before test
+Oct 26 16:25:49.990: INFO: ibm-keepalived-watcher-qwg48 from kube-system started at 2020-10-26 14:34:30 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.990: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 16:25:49.990: INFO: ibmcloud-block-storage-plugin-5f486d686f-6brjv from kube-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.990: INFO: 	Container ibmcloud-block-storage-plugin-container ready: true, restart count 0
+Oct 26 16:25:49.990: INFO: ibm-master-proxy-static-10.72.119.76 from kube-system started at 2020-10-26 14:34:28 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.990: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 16:25:49.990: INFO: 	Container pause ready: true, restart count 0
+Oct 26 16:25:49.990: INFO: cluster-image-registry-operator-8444cfb976-gdjk6 from openshift-image-registry started at 2020-10-26 14:35:50 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.990: INFO: 	Container cluster-image-registry-operator ready: true, restart count 0
+Oct 26 16:25:49.990: INFO: 	Container cluster-image-registry-operator-watch ready: true, restart count 0
+Oct 26 16:25:49.990: INFO: prometheus-k8s-0 from openshift-monitoring started at 2020-10-26 14:45:39 +0000 UTC (7 container statuses recorded)
+Oct 26 16:25:49.991: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:49.991: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 16:25:49.991: INFO: 	Container prometheus ready: true, restart count 1
+Oct 26 16:25:49.991: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
+Oct 26 16:25:49.991: INFO: 	Container prometheus-proxy ready: true, restart count 0
+Oct 26 16:25:49.991: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
+Oct 26 16:25:49.991: INFO: 	Container thanos-sidecar ready: true, restart count 0
+Oct 26 16:25:49.991: INFO: tuned-s6flg from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.991: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 16:25:49.991: INFO: network-operator-86b67c6f7f-pcq2w from openshift-network-operator started at 2020-10-26 14:34:37 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.991: INFO: 	Container network-operator ready: true, restart count 0
+Oct 26 16:25:49.991: INFO: dns-operator-58d86f779c-wbm98 from openshift-dns-operator started at 2020-10-26 14:35:49 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.991: INFO: 	Container dns-operator ready: true, restart count 0
+Oct 26 16:25:49.991: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:49.991: INFO: cluster-storage-operator-9b57569c9-gpqd2 from openshift-cluster-storage-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.991: INFO: 	Container cluster-storage-operator ready: true, restart count 0
+Oct 26 16:25:49.992: INFO: service-ca-operator-656f67898-svtt9 from openshift-service-ca-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.992: INFO: 	Container operator ready: true, restart count 1
+Oct 26 16:25:49.992: INFO: marketplace-operator-5df598b96b-6vr64 from openshift-marketplace started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.992: INFO: 	Container marketplace-operator ready: true, restart count 0
+Oct 26 16:25:49.992: INFO: console-operator-5fc674d8d4-fjctd from openshift-console-operator started at 2020-10-26 14:35:51 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.992: INFO: 	Container console-operator ready: true, restart count 1
+Oct 26 16:25:49.992: INFO: openshift-kube-proxy-74ss4 from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.992: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 16:25:49.992: INFO: ibm-file-plugin-944b59c94-k7mrm from kube-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.992: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
+Oct 26 16:25:49.992: INFO: calico-typha-7fd49c779d-766qb from calico-system started at 2020-10-26 14:37:36 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.992: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 16:25:49.992: INFO: dns-default-z8pn9 from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.992: INFO: 	Container dns ready: true, restart count 0
+Oct 26 16:25:49.992: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 16:25:49.992: INFO: ibmcloud-block-storage-driver-b724k from kube-system started at 2020-10-26 14:34:34 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.993: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 16:25:49.993: INFO: openshift-service-catalog-apiserver-operator-7cfbcc49dc-hrn66 from openshift-service-catalog-apiserver-operator started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.993: INFO: 	Container operator ready: true, restart count 1
+Oct 26 16:25:49.993: INFO: alertmanager-main-2 from openshift-monitoring started at 2020-10-26 14:44:28 +0000 UTC (3 container statuses recorded)
+Oct 26 16:25:49.993: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 16:25:49.993: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 16:25:49.993: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 16:25:49.993: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-2scb4 from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.993: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 16:25:49.993: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 16:25:49.993: INFO: router-default-84766ff4fd-cftxw from openshift-ingress started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.993: INFO: 	Container router ready: true, restart count 0
+Oct 26 16:25:49.993: INFO: multus-lbpcn from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.993: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 16:25:49.994: INFO: openshift-service-catalog-controller-manager-operator-d49bz89wp from openshift-service-catalog-controller-manager-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.994: INFO: 	Container operator ready: true, restart count 1
+Oct 26 16:25:49.994: INFO: ibm-storage-watcher-77d6cc685c-bvg8q from kube-system started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.994: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
+Oct 26 16:25:49.994: INFO: olm-operator-7cd6cbf579-zww2l from openshift-operator-lifecycle-manager started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.994: INFO: 	Container olm-operator ready: true, restart count 0
+Oct 26 16:25:49.994: INFO: cluster-monitoring-operator-69649564f8-r4h9b from openshift-monitoring started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.994: INFO: 	Container cluster-monitoring-operator ready: true, restart count 0
+Oct 26 16:25:49.994: INFO: ibm-cloud-provider-ip-158-176-78-170-5cdb49c95f-76fjw from ibm-system started at 2020-10-26 14:42:32 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.994: INFO: 	Container ibm-cloud-provider-ip-158-176-78-170 ready: true, restart count 0
+Oct 26 16:25:49.994: INFO: ingress-operator-65878dcc8d-wdnrl from openshift-ingress-operator started at 2020-10-26 14:35:49 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.994: INFO: 	Container ingress-operator ready: true, restart count 0
+Oct 26 16:25:49.994: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:49.994: INFO: downloads-74f6b6dcb6-r9ksv from openshift-console started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.994: INFO: 	Container download-server ready: true, restart count 0
+Oct 26 16:25:49.995: INFO: multus-admission-controller-5j529 from openshift-multus started at 2020-10-26 14:35:50 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.995: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:49.995: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 16:25:49.995: INFO: node-exporter-zmwjz from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:49.995: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:49.995: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 16:25:49.995: INFO: node-ca-bpxm5 from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.995: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 16:25:49.995: INFO: cluster-node-tuning-operator-575b98978c-zz7s9 from openshift-cluster-node-tuning-operator started at 2020-10-26 14:35:51 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.995: INFO: 	Container cluster-node-tuning-operator ready: true, restart count 0
+Oct 26 16:25:49.995: INFO: calico-node-dhd7p from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.995: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 16:25:49.995: INFO: kube-storage-version-migrator-operator-86d479ffb5-9cmml from openshift-kube-storage-version-migrator-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.995: INFO: 	Container kube-storage-version-migrator-operator ready: true, restart count 1
+Oct 26 16:25:49.995: INFO: catalog-operator-5db6d5849c-z2lgx from openshift-operator-lifecycle-manager started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.995: INFO: 	Container catalog-operator ready: true, restart count 0
+Oct 26 16:25:49.996: INFO: calico-kube-controllers-cdc7f7f8f-jqklm from calico-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.996: INFO: 	Container calico-kube-controllers ready: true, restart count 0
+Oct 26 16:25:49.996: INFO: downloads-74f6b6dcb6-vqzf9 from openshift-console started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:49.996: INFO: 	Container download-server ready: true, restart count 0
+Oct 26 16:25:49.996: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.82 before test
+Oct 26 16:25:50.060: INFO: ibm-cloud-provider-ip-158-176-78-170-5cdb49c95f-89x4j from ibm-system started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.060: INFO: 	Container ibm-cloud-provider-ip-158-176-78-170 ready: true, restart count 0
+Oct 26 16:25:50.060: INFO: multus-admission-controller-9rv7d from openshift-multus started at 2020-10-26 14:36:02 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:50.060: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:50.060: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 16:25:50.060: INFO: dns-default-ggg6q from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:50.060: INFO: 	Container dns ready: true, restart count 0
+Oct 26 16:25:50.060: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 16:25:50.060: INFO: prometheus-k8s-1 from openshift-monitoring started at 2020-10-26 14:45:29 +0000 UTC (7 container statuses recorded)
+Oct 26 16:25:50.060: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:50.060: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 16:25:50.060: INFO: 	Container prometheus ready: true, restart count 1
+Oct 26 16:25:50.061: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: 	Container prometheus-proxy ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: 	Container thanos-sidecar ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: console-5dd98d6f6f-xvz7c from openshift-console started at 2020-10-26 14:45:54 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.061: INFO: 	Container console ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: packageserver-5f76c64df7-6nrvc from openshift-operator-lifecycle-manager started at 2020-10-26 16:10:11 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.061: INFO: 	Container packageserver ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: calico-node-c7wjs from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.061: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: ibmcloud-block-storage-driver-sxlj2 from kube-system started at 2020-10-26 14:34:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.061: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: multus-hwkgh from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.061: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: kube-state-metrics-7bfb6f7864-vmgz9 from openshift-monitoring started at 2020-10-26 14:36:32 +0000 UTC (3 container statuses recorded)
+Oct 26 16:25:50.061: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: 	Container kube-state-metrics ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: thanos-querier-84d7d58bc-mpnzq from openshift-monitoring started at 2020-10-26 14:45:18 +0000 UTC (4 container statuses recorded)
+Oct 26 16:25:50.061: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:50.061: INFO: 	Container oauth-proxy ready: true, restart count 0
+Oct 26 16:25:50.062: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 16:25:50.062: INFO: 	Container thanos-querier ready: true, restart count 0
+Oct 26 16:25:50.062: INFO: community-operators-f4df99f64-twf5k from openshift-marketplace started at 2020-10-26 14:38:39 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.062: INFO: 	Container community-operators ready: true, restart count 0
+Oct 26 16:25:50.062: INFO: ibm-master-proxy-static-10.72.119.82 from kube-system started at 2020-10-26 14:34:29 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:50.062: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 16:25:50.062: INFO: 	Container pause ready: true, restart count 0
+Oct 26 16:25:50.062: INFO: node-ca-z564l from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.062: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 16:25:50.062: INFO: tuned-h578f from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.062: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 16:25:50.062: INFO: certified-operators-69bb9f65cc-9mhpt from openshift-marketplace started at 2020-10-26 14:38:38 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.062: INFO: 	Container certified-operators ready: true, restart count 0
+Oct 26 16:25:50.062: INFO: redhat-marketplace-77d9b69f69-qb5s9 from openshift-marketplace started at 2020-10-26 14:38:39 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.062: INFO: 	Container redhat-marketplace ready: true, restart count 0
+Oct 26 16:25:50.062: INFO: alertmanager-main-1 from openshift-monitoring started at 2020-10-26 14:44:39 +0000 UTC (3 container statuses recorded)
+Oct 26 16:25:50.062: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 16:25:50.062: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 16:25:50.062: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 16:25:50.063: INFO: ibm-keepalived-watcher-98n85 from kube-system started at 2020-10-26 14:34:32 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.063: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 16:25:50.063: INFO: test-k8s-e2e-pvg-master-verification from default started at 2020-10-26 14:37:46 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.063: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
+Oct 26 16:25:50.063: INFO: registry-pvc-permissions-m9wwv from openshift-image-registry started at 2020-10-26 14:40:09 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.063: INFO: 	Container pvc-permissions ready: false, restart count 0
+Oct 26 16:25:50.063: INFO: calico-typha-7fd49c779d-nmzbk from calico-system started at 2020-10-26 14:35:39 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.063: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 16:25:50.063: INFO: openshift-state-metrics-d4d5d6f57-lszmv from openshift-monitoring started at 2020-10-26 14:36:34 +0000 UTC (3 container statuses recorded)
+Oct 26 16:25:50.063: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
+Oct 26 16:25:50.063: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
+Oct 26 16:25:50.063: INFO: 	Container openshift-state-metrics ready: true, restart count 0
+Oct 26 16:25:50.063: INFO: node-exporter-4lfg7 from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:50.063: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:25:50.063: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 16:25:50.063: INFO: migrator-686fc6cc66-9wpj9 from openshift-kube-storage-version-migrator started at 2020-10-26 14:36:53 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.063: INFO: 	Container migrator ready: true, restart count 0
+Oct 26 16:25:50.063: INFO: prometheus-adapter-674456c5d6-kj77x from openshift-monitoring started at 2020-10-26 14:44:24 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.063: INFO: 	Container prometheus-adapter ready: true, restart count 0
+Oct 26 16:25:50.063: INFO: image-registry-6db5c967cf-d4ljt from openshift-image-registry started at 2020-10-26 14:40:09 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.064: INFO: 	Container registry ready: true, restart count 0
+Oct 26 16:25:50.064: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-x2lkl from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 16:25:50.064: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 16:25:50.064: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 16:25:50.064: INFO: openshift-kube-proxy-z9sjt from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.064: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 16:25:50.064: INFO: router-default-84766ff4fd-7dgrd from openshift-ingress started at 2020-10-26 14:38:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.064: INFO: 	Container router ready: true, restart count 0
+Oct 26 16:25:50.064: INFO: prometheus-operator-55f946f5d-tpnd8 from openshift-monitoring started at 2020-10-26 14:44:11 +0000 UTC (1 container statuses recorded)
+Oct 26 16:25:50.064: INFO: 	Container prometheus-operator ready: true, restart count 0
+[It] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:14:01.053: INFO: Starting informer...
-STEP: Starting pods...
-Sep 21 16:14:01.398: INFO: Pod1 is running on 10.241.51.147. Tainting Node
-Sep 21 16:14:04.872: INFO: Pod2 is running on 10.241.51.147. Tainting Node
-STEP: Trying to apply a taint on the Node
-STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute
-STEP: Waiting for Pod1 and Pod2 to be deleted
-Sep 21 16:14:17.436: INFO: Noticed Pod "taint-eviction-b1" gets evicted.
-Sep 21 16:14:37.346: INFO: Noticed Pod "taint-eviction-b2" gets evicted.
-STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute
-[AfterEach] [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
+STEP: Trying to launch a pod without a label to get a node which can launch it.
+STEP: Explicitly delete pod here to free the resource it takes.
+STEP: Trying to apply a random label on the found node.
+STEP: verifying the node has the label kubernetes.io/e2e-5b6ff1b8-0aa1-4df5-8596-f13f6836547b 95
+STEP: Trying to create a pod(pod4) with hostport 54322 and hostIP 0.0.0.0(empty string here) and expect scheduled
+STEP: Trying to create another pod(pod5) with hostport 54322 but hostIP 127.0.0.1 on the node which pod4 resides and expect not scheduled
+STEP: removing the label kubernetes.io/e2e-5b6ff1b8-0aa1-4df5-8596-f13f6836547b off the node 10.72.119.74
+STEP: verifying the node doesn't have the label kubernetes.io/e2e-5b6ff1b8-0aa1-4df5-8596-f13f6836547b
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:14:37.451: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "taint-multiple-pods-2154" for this suite.
+Oct 26 16:30:54.508: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-6207" for this suite.
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
 
-• [SLOW TEST:96.893 seconds]
-[sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
+• [SLOW TEST:304.955 seconds]
+[sig-scheduling] SchedulerPredicates [Serial]
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
-  evicts pods with minTolerationSeconds [Disruptive] [Conformance]
+  validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial] evicts pods with minTolerationSeconds [Disruptive] [Conformance]","total":280,"completed":87,"skipped":1531,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]","total":280,"completed":91,"skipped":1292,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] Deployment 
-  deployment should support proportional scaling [Conformance]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  works for CRD preserving unknown fields in an embedded object [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Deployment
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:14:37.549: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename deployment
+Oct 26 16:30:54.549: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename crd-publish-openapi
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
-[It] deployment should support proportional scaling [Conformance]
+[It] works for CRD preserving unknown fields in an embedded object [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:14:37.876: INFO: Creating deployment "webserver-deployment"
-Sep 21 16:14:37.927: INFO: Waiting for observed generation 1
-Sep 21 16:14:40.076: INFO: Waiting for all required pods to come up
-Sep 21 16:14:40.148: INFO: Pod name httpd: Found 10 pods out of 10
-STEP: ensuring each pod is running
-Sep 21 16:14:42.263: INFO: Waiting for deployment "webserver-deployment" to complete
-Sep 21 16:14:42.304: INFO: Updating deployment "webserver-deployment" with a non-existent image
-Sep 21 16:14:42.444: INFO: Updating deployment webserver-deployment
-Sep 21 16:14:42.444: INFO: Waiting for observed generation 2
-Sep 21 16:14:44.501: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8
-Sep 21 16:14:44.527: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8
-Sep 21 16:14:44.556: INFO: Waiting for the first rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas
-Sep 21 16:14:44.792: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0
-Sep 21 16:14:44.792: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5
-Sep 21 16:14:44.821: INFO: Waiting for the second rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas
-Sep 21 16:14:44.861: INFO: Verifying that deployment "webserver-deployment" has minimum required number of available replicas
-Sep 21 16:14:44.861: INFO: Scaling up the deployment "webserver-deployment" from 10 to 30
-Sep 21 16:14:44.909: INFO: Updating deployment webserver-deployment
-Sep 21 16:14:44.909: INFO: Waiting for the replicasets of deployment "webserver-deployment" to have desired number of replicas
-Sep 21 16:14:45.040: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20
-Sep 21 16:14:45.732: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
-Sep 21 16:14:47.834: INFO: Deployment "webserver-deployment":
-&Deployment{ObjectMeta:{webserver-deployment  deployment-1153 /apis/apps/v1/namespaces/deployment-1153/deployments/webserver-deployment d1085b70-7d1a-4384-bea2-4abf5bf26fe4 61834 3 2020-09-21 16:14:37 +0000 UTC   map[name:httpd] map[deployment.kubernetes.io/revision:2] [] []  []},Spec:DeploymentSpec{Replicas:*30,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:httpd] map[] [] []  []} {[] [] [{httpd webserver:404 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc002233128  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:33,UpdatedReplicas:13,AvailableReplicas:11,UnavailableReplicas:22,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2020-09-21 16:14:44 +0000 UTC,LastTransitionTime:2020-09-21 16:14:44 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "webserver-deployment-c7997dcc8" is progressing.,LastUpdateTime:2020-09-21 16:14:47 +0000 UTC,LastTransitionTime:2020-09-21 16:14:37 +0000 UTC,},},ReadyReplicas:11,CollisionCount:nil,},}
-
-Sep 21 16:14:47.889: INFO: New ReplicaSet "webserver-deployment-c7997dcc8" of Deployment "webserver-deployment":
-&ReplicaSet{ObjectMeta:{webserver-deployment-c7997dcc8  deployment-1153 /apis/apps/v1/namespaces/deployment-1153/replicasets/webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 61667 3 2020-09-21 16:14:42 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment webserver-deployment d1085b70-7d1a-4384-bea2-4abf5bf26fe4 0xc002051f17 0xc002051f18}] []  []},Spec:ReplicaSetSpec{Replicas:*13,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: c7997dcc8,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [] []  []} {[] [] [{httpd webserver:404 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc002051f88  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:13,FullyLabeledReplicas:13,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
-Sep 21 16:14:47.889: INFO: All old ReplicaSets of Deployment "webserver-deployment":
-Sep 21 16:14:47.889: INFO: &ReplicaSet{ObjectMeta:{webserver-deployment-595b5b9587  deployment-1153 /apis/apps/v1/namespaces/deployment-1153/replicasets/webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 61842 3 2020-09-21 16:14:37 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment webserver-deployment d1085b70-7d1a-4384-bea2-4abf5bf26fe4 0xc002051e57 0xc002051e58}] []  []},Spec:ReplicaSetSpec{Replicas:*20,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 595b5b9587,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc002051eb8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:14,AvailableReplicas:14,Conditions:[]ReplicaSetCondition{},},}
-Sep 21 16:14:47.981: INFO: Pod "webserver-deployment-595b5b9587-4bh5v" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-4bh5v webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-4bh5v 54bd0d17-10e8-4f65-b2fa-31a345fbe9ae 61839 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.81.60/32 cni.projectcalico.org/podIPs:172.30.81.60/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.81.60"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc005278497 0xc005278498}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:172.30.81.60,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:47 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://7572e893a3a717fcc5c221859bee43d08c7741289c63de20fd893e3eb8362567,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.81.60,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.982: INFO: Pod "webserver-deployment-595b5b9587-52589" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-52589 webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-52589 82ee33e8-7bbb-451c-ab43-68a1d27f52cd 61825 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.81.16/32 cni.projectcalico.org/podIPs:172.30.81.16/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.81.16"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc005278657 0xc005278658}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.982: INFO: Pod "webserver-deployment-595b5b9587-7vtrz" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-7vtrz webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-7vtrz 760ec87d-fb9a-4bed-ae26-a333d47d6d72 61376 0 2020-09-21 16:14:37 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.237.110/32 cni.projectcalico.org/podIPs:172.30.237.110/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.110"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc0052787d7 0xc0052787d8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:172.30.237.110,StartTime:2020-09-21 16:14:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:39 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://3c768ad6dee2f4de8abb9132d432bc3e045d28bbdbff5efe8451c1b965bd4b5b,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.237.110,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.982: INFO: Pod "webserver-deployment-595b5b9587-8q2mg" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-8q2mg webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-8q2mg 7264078b-e9a6-4f86-a1ef-34e9ccb919a0 61378 0 2020-09-21 16:14:38 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.81.5/32 cni.projectcalico.org/podIPs:172.30.81.5/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.81.5"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc005278997 0xc005278998}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:172.30.81.5,StartTime:2020-09-21 16:14:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:40 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://65895faffbf7007b9c4171887581e50e5e5961e2dbd5403f7ac87a0302d8fccc,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.81.5,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.983: INFO: Pod "webserver-deployment-595b5b9587-9d2jb" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-9d2jb webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-9d2jb c0960f6b-d6a5-481f-8e45-1cedcbf541a3 61371 0 2020-09-21 16:14:38 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.139.54/32 cni.projectcalico.org/podIPs:172.30.139.54/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.139.54"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc005278b50 0xc005278b51}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.149,PodIP:172.30.139.54,StartTime:2020-09-21 16:14:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:39 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://bb46ac0462ffb6465e653d48e5f26455d65c8b19455f1a624e5ee259366b70a6,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.139.54,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.983: INFO: Pod "webserver-deployment-595b5b9587-dxlbg" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-dxlbg webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-dxlbg bf254e07-5268-4b72-b9ba-1f6e408d8a5b 61848 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.237.123/32 cni.projectcalico.org/podIPs:172.30.237.123/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.123"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc005278ce7 0xc005278ce8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.983: INFO: Pod "webserver-deployment-595b5b9587-fcdz5" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-fcdz5 webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-fcdz5 a232ae01-1e7f-4a75-a87c-2319076d2369 61397 0 2020-09-21 16:14:38 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.81.59/32 cni.projectcalico.org/podIPs:172.30.81.59/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.81.59"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc005278e87 0xc005278e88}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:172.30.81.59,StartTime:2020-09-21 16:14:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:40 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://82c8b54916b0945e5a1a9b3ceb2f74be71447c224c383ddea62e6d04db85eb78,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.81.59,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.984: INFO: Pod "webserver-deployment-595b5b9587-gc8p4" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-gc8p4 webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-gc8p4 e9f0950d-e42d-494c-86d8-0b1f796bac98 61844 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.237.119/32 cni.projectcalico.org/podIPs:172.30.237.119/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.119"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc005279027 0xc005279028}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:172.30.237.119,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:47 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://38b9a9aaf4c31f02ddc63a4519bf8f4dc7ec00490d88b9de872fd0438d407d76,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.237.119,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.984: INFO: Pod "webserver-deployment-595b5b9587-hjlk2" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-hjlk2 webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-hjlk2 b0ea0f96-518f-42e7-ab97-d561d9b84f9c 61831 0 2020-09-21 16:14:44 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.237.117/32 cni.projectcalico.org/podIPs:172.30.237.117/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.117"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc0052792a7 0xc0052792a8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:172.30.237.117,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:46 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://71de54808cf96a9b7d3385e40c876df776a7cce96a0afad0e57b1e5ade9bcf5e,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.237.117,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.984: INFO: Pod "webserver-deployment-595b5b9587-hwqgl" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-hwqgl webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-hwqgl d7b4f316-57d2-492a-813a-fba1213855b1 61812 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.139.53/32 cni.projectcalico.org/podIPs:172.30.139.53/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.139.53"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc005279527 0xc005279528}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.149,PodIP:172.30.139.53,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:47 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://c0c4981d124d4893ae9cdf41a837f44cafc3c19e3606607621182a6af8a2face,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.139.53,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.985: INFO: Pod "webserver-deployment-595b5b9587-jp9zz" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-jp9zz webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-jp9zz a290b8b7-6010-4f59-be83-b63553cece26 61838 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.237.118/32 cni.projectcalico.org/podIPs:172.30.237.118/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.118"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc0052796c7 0xc0052796c8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:172.30.237.118,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:47 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://b88b7ce84c44c069b8aeade0953988a53f1bdda4d6dcf98f7f576a9ead0eff64,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.237.118,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.985: INFO: Pod "webserver-deployment-595b5b9587-js9xr" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-js9xr webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-js9xr c6b2b546-fec6-4cab-acab-4a94039bdbb8 61389 0 2020-09-21 16:14:38 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.81.57/32 cni.projectcalico.org/podIPs:172.30.81.57/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.81.57"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc0052798a7 0xc0052798a8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:172.30.81.57,StartTime:2020-09-21 16:14:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:39 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://9903068b3cde1a155872c83d7f1f3108487d5b607f389336c6e1acc543a72334,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.81.57,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.985: INFO: Pod "webserver-deployment-595b5b9587-kg8n2" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-kg8n2 webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-kg8n2 b1d7932c-6fa9-47f0-9e5a-d6c8bf7c4917 61791 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.81.13/32 cni.projectcalico.org/podIPs:172.30.81.13/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.81.13"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc005279a77 0xc005279a78}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.985: INFO: Pod "webserver-deployment-595b5b9587-kvj2c" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-kvj2c webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-kvj2c f26a4ddb-c670-4a3b-8edc-30a31dd30126 61806 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.139.56/32 cni.projectcalico.org/podIPs:172.30.139.56/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.139.56"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc005279c17 0xc005279c18}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.149,PodIP:172.30.139.56,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:46 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://0f68212f39f277524a031d5b6612d6e18550cb122bc8a082ee78dde95b21bc02,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.139.56,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.986: INFO: Pod "webserver-deployment-595b5b9587-kwc8c" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-kwc8c webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-kwc8c ef31330f-6f8f-49d0-b51c-4b1742022c01 61777 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.81.7/32 cni.projectcalico.org/podIPs:172.30.81.7/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.81.7"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc005279dd7 0xc005279dd8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.986: INFO: Pod "webserver-deployment-595b5b9587-lplbh" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-lplbh webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-lplbh 1c4932cf-d95f-417d-938a-85b512bf34e9 61386 0 2020-09-21 16:14:38 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.237.112/32 cni.projectcalico.org/podIPs:172.30.237.112/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.112"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc005279f57 0xc005279f58}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:172.30.237.112,StartTime:2020-09-21 16:14:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:39 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://9b490c11187ebf2042a4563210c6a1b7cabc0a0bd08843ea27caa2ea8189a407,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.237.112,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.986: INFO: Pod "webserver-deployment-595b5b9587-m52v4" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-m52v4 webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-m52v4 62b9b476-9fa6-4903-9ec6-e6a867d90c22 61396 0 2020-09-21 16:14:38 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.237.113/32 cni.projectcalico.org/podIPs:172.30.237.113/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.113"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc00373e0f7 0xc00373e0f8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:172.30.237.113,StartTime:2020-09-21 16:14:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:40 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://6c6be22d657be91d18edfa934e26cd8f5c66685d1174395488de27683e22126e,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.237.113,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.987: INFO: Pod "webserver-deployment-595b5b9587-p7w8d" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-p7w8d webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-p7w8d dc2d2353-bd9b-4a30-bda7-11d71cb1d66e 61406 0 2020-09-21 16:14:38 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.81.58/32 cni.projectcalico.org/podIPs:172.30.81.58/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.81.58"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc00373e2b7 0xc00373e2b8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:38 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:172.30.81.58,StartTime:2020-09-21 16:14:38 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:39 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://ec5b7a19a6b43f8374fa253b7f4abb558a673ac110cef64bb329f7d199b0e21e,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.81.58,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.987: INFO: Pod "webserver-deployment-595b5b9587-smg9k" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-smg9k webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-smg9k 14005f84-c432-4802-a0af-a0c3a7f17cd6 61817 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.237.121/32 cni.projectcalico.org/podIPs:172.30.237.121/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.121"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc00373e467 0xc00373e468}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.987: INFO: Pod "webserver-deployment-595b5b9587-v589z" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-v589z webserver-deployment-595b5b9587- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-595b5b9587-v589z 02c1d347-c5b2-4ae0-9b59-b6c91a882813 61827 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.30.81.61/32 cni.projectcalico.org/podIPs:172.30.81.61/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.81.61"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 a63d2beb-769b-4826-99e8-4c18ecc10d6a 0xc00373e607 0xc00373e608}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:172.30.81.61,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:14:47 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4,ContainerID:cri-o://133876f63dfd788be55063ce36b51cf31e4393d19387f851032810f05e792e85,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.81.61,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.987: INFO: Pod "webserver-deployment-c7997dcc8-22vbx" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-22vbx webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-22vbx 609e2a5a-1e6f-43dc-9672-201d222a8510 61792 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.237.122/32 cni.projectcalico.org/podIPs:172.30.237.122/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.122"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373e7c7 0xc00373e7c8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.988: INFO: Pod "webserver-deployment-c7997dcc8-42wc6" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-42wc6 webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-42wc6 54532ae7-0253-4d85-a23e-a96bebfe2262 61778 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.237.111/32 cni.projectcalico.org/podIPs:172.30.237.111/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.111"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373e967 0xc00373e968}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.988: INFO: Pod "webserver-deployment-c7997dcc8-7zsh2" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-7zsh2 webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-7zsh2 11672c3a-b668-4484-82b8-b50b4fda7b2e 61840 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.81.15/32 cni.projectcalico.org/podIPs:172.30.81.15/32 openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373eb27 0xc00373eb28}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.988: INFO: Pod "webserver-deployment-c7997dcc8-9xgfn" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-9xgfn webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-9xgfn ae6750db-40ef-49e1-bc1c-892d0bdab018 61644 0 2020-09-21 16:14:42 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.237.115/32 cni.projectcalico.org/podIPs:172.30.237.115/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.115"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373ecc7 0xc00373ecc8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:172.30.237.115,StartTime:2020-09-21 16:14:42 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = Error reading manifest 404 in docker.io/library/webserver: errors:
-denied: requested access to the resource is denied
-unauthorized: authentication required
-,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.237.115,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.989: INFO: Pod "webserver-deployment-c7997dcc8-ck8bm" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-ck8bm webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-ck8bm 9a3c7f97-9119-4c29-92bb-756888870a80 61663 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373eea7 0xc00373eea8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.989: INFO: Pod "webserver-deployment-c7997dcc8-gxwkg" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-gxwkg webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-gxwkg 178e08d0-557a-4ccc-9d49-463db10f3295 61758 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.139.58/32 cni.projectcalico.org/podIPs:172.30.139.58/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.139.58"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373f067 0xc00373f068}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.149,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.989: INFO: Pod "webserver-deployment-c7997dcc8-l2qst" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-l2qst webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-l2qst b0ad8769-e559-4375-b418-836e35572cae 61713 0 2020-09-21 16:14:42 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.237.116/32 cni.projectcalico.org/podIPs:172.30.237.116/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.116"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373f217 0xc00373f218}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:172.30.237.116,StartTime:2020-09-21 16:14:42 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = Error reading manifest 404 in docker.io/library/webserver: errors:
-denied: requested access to the resource is denied
-unauthorized: authentication required
-,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.237.116,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.989: INFO: Pod "webserver-deployment-c7997dcc8-m5rgh" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-m5rgh webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-m5rgh 64f150f8-1510-4f5f-94fe-ac4859700214 61786 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.139.57/32 cni.projectcalico.org/podIPs:172.30.139.57/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.139.57"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373f417 0xc00373f418}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.149,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.990: INFO: Pod "webserver-deployment-c7997dcc8-mld84" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-mld84 webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-mld84 36f9ee44-9a70-4fbb-8188-94e6c3d7df8e 61802 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.81.14/32 cni.projectcalico.org/podIPs:172.30.81.14/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.81.14"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373f5d7 0xc00373f5d8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:nil,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.990: INFO: Pod "webserver-deployment-c7997dcc8-nd5kg" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-nd5kg webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-nd5kg 4c2c889c-d083-4f63-ae50-d648695b9a9e 61636 0 2020-09-21 16:14:42 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.81.62/32 cni.projectcalico.org/podIPs:172.30.81.62/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.81.62"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373f7b7 0xc00373f7b8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:172.30.81.62,StartTime:2020-09-21 16:14:42 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = Error reading manifest 404 in docker.io/library/webserver: errors:
-denied: requested access to the resource is denied
-unauthorized: authentication required
-,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.81.62,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.990: INFO: Pod "webserver-deployment-c7997dcc8-nk6fs" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-nk6fs webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-nk6fs d9715ee0-dd55-40ce-a0bb-7eafeae8b278 61618 0 2020-09-21 16:14:42 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.139.55/32 cni.projectcalico.org/podIPs:172.30.139.55/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.139.55"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373f9a7 0xc00373f9a8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.149,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:42 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.149,PodIP:172.30.139.55,StartTime:2020-09-21 16:14:42 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = Error reading manifest 404 in docker.io/library/webserver: errors:
-denied: requested access to the resource is denied
-unauthorized: authentication required
-,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.139.55,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.990: INFO: Pod "webserver-deployment-c7997dcc8-nkvsx" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-nkvsx webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-nkvsx 7b3f3662-14ce-4335-91ac-1197a25f36d7 61707 0 2020-09-21 16:14:43 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.30.81.63/32 cni.projectcalico.org/podIPs:172.30.81.63/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.81.63"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373fb97 0xc00373fb98}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:43 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:43 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:43 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:43 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:172.30.81.63,StartTime:2020-09-21 16:14:43 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = Error reading manifest 404 in docker.io/library/webserver: errors:
-denied: requested access to the resource is denied
-unauthorized: authentication required
-,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.81.63,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Sep 21 16:14:47.991: INFO: Pod "webserver-deployment-c7997dcc8-tbwrk" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-tbwrk webserver-deployment-c7997dcc8- deployment-1153 /api/v1/namespaces/deployment-1153/pods/webserver-deployment-c7997dcc8-tbwrk ec072525-c689-49e9-ba13-d95e277f0f1f 61657 0 2020-09-21 16:14:45 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[openshift.io/scc:privileged] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 f72d1c93-388f-43ea-ab74-fa9e848afc09 0xc00373fd67 0xc00373fd68}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kc29x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kc29x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kc29x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.150,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-m6p5p,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.150,PodIP:,StartTime:2020-09-21 16:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-[AfterEach] [sig-apps] Deployment
+Oct 26 16:30:54.673: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: client-side validation (kubectl create and apply) allows request with any unknown properties
+Oct 26 16:31:03.215: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-821 create -f -'
+Oct 26 16:31:03.935: INFO: stderr: ""
+Oct 26 16:31:03.935: INFO: stdout: "e2e-test-crd-publish-openapi-1409-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n"
+Oct 26 16:31:03.935: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-821 delete e2e-test-crd-publish-openapi-1409-crds test-cr'
+Oct 26 16:31:04.108: INFO: stderr: ""
+Oct 26 16:31:04.109: INFO: stdout: "e2e-test-crd-publish-openapi-1409-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n"
+Oct 26 16:31:04.109: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-821 apply -f -'
+Oct 26 16:31:04.495: INFO: stderr: ""
+Oct 26 16:31:04.495: INFO: stdout: "e2e-test-crd-publish-openapi-1409-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n"
+Oct 26 16:31:04.495: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-821 delete e2e-test-crd-publish-openapi-1409-crds test-cr'
+Oct 26 16:31:04.678: INFO: stderr: ""
+Oct 26 16:31:04.678: INFO: stdout: "e2e-test-crd-publish-openapi-1409-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n"
+STEP: kubectl explain works to explain CR
+Oct 26 16:31:04.678: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 explain e2e-test-crd-publish-openapi-1409-crds'
+Oct 26 16:31:05.205: INFO: stderr: ""
+Oct 26 16:31:05.205: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-1409-crd\nVERSION:  crd-publish-openapi-test-unknown-in-nested.example.com/v1\n\nDESCRIPTION:\n     preserve-unknown-properties in nested field for Testing\n\nFIELDS:\n   apiVersion\t\n     APIVersion defines the versioned schema of this representation of an\n     object. Servers should convert recognized schemas to the latest internal\n     value, and may reject unrecognized values. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n   kind\t\n     Kind is a string value representing the REST resource this object\n     represents. Servers may infer this from the endpoint the client submits\n     requests to. Cannot be updated. In CamelCase. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n   metadata\t\n     Standard object's metadata. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   spec\t\n     Specification of Waldo\n\n   status\t\n     Status of Waldo\n\n"
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:14:47.991: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "deployment-1153" for this suite.
+Oct 26 16:31:13.203: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-821" for this suite.
 
-• [SLOW TEST:10.510 seconds]
-[sig-apps] Deployment
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  deployment should support proportional scaling [Conformance]
+• [SLOW TEST:18.691 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  works for CRD preserving unknown fields in an embedded object [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] Deployment deployment should support proportional scaling [Conformance]","total":280,"completed":88,"skipped":1555,"failed":0}
-SSSSSSS
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance]","total":280,"completed":92,"skipped":1321,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Security Context when creating containers with AllowPrivilegeEscalation 
-  should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]
+[sig-api-machinery] Watchers 
+  should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Security Context
+[BeforeEach] [sig-api-machinery] Watchers
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:14:48.060: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename security-context-test
+Oct 26 16:31:13.240: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename watch
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39
-[It] should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]
+[It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:14:48.528: INFO: Waiting up to 5m0s for pod "alpine-nnp-false-f2a47afa-37ce-424e-af11-d868ed1346ad" in namespace "security-context-test-5858" to be "success or failure"
-Sep 21 16:14:48.552: INFO: Pod "alpine-nnp-false-f2a47afa-37ce-424e-af11-d868ed1346ad": Phase="Pending", Reason="", readiness=false. Elapsed: 24.010584ms
-Sep 21 16:14:50.576: INFO: Pod "alpine-nnp-false-f2a47afa-37ce-424e-af11-d868ed1346ad": Phase="Pending", Reason="", readiness=false. Elapsed: 2.04857257s
-Sep 21 16:14:52.601: INFO: Pod "alpine-nnp-false-f2a47afa-37ce-424e-af11-d868ed1346ad": Phase="Pending", Reason="", readiness=false. Elapsed: 4.073320626s
-Sep 21 16:14:54.626: INFO: Pod "alpine-nnp-false-f2a47afa-37ce-424e-af11-d868ed1346ad": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.098115531s
-Sep 21 16:14:54.626: INFO: Pod "alpine-nnp-false-f2a47afa-37ce-424e-af11-d868ed1346ad" satisfied condition "success or failure"
-[AfterEach] [k8s.io] Security Context
+STEP: creating a watch on configmaps with a certain label
+STEP: creating a new configmap
+STEP: modifying the configmap once
+STEP: changing the label value of the configmap
+STEP: Expecting to observe a delete notification for the watched object
+Oct 26 16:31:13.499: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-5004 /api/v1/namespaces/watch-5004/configmaps/e2e-watch-test-label-changed f47bf8f8-834b-4c73-9291-1a2a9d931579 62244 0 2020-10-26 16:31:13 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Oct 26 16:31:13.499: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-5004 /api/v1/namespaces/watch-5004/configmaps/e2e-watch-test-label-changed f47bf8f8-834b-4c73-9291-1a2a9d931579 62246 0 2020-10-26 16:31:13 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+Oct 26 16:31:13.499: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-5004 /api/v1/namespaces/watch-5004/configmaps/e2e-watch-test-label-changed f47bf8f8-834b-4c73-9291-1a2a9d931579 62247 0 2020-10-26 16:31:13 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+STEP: modifying the configmap a second time
+STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements
+STEP: changing the label value of the configmap back
+STEP: modifying the configmap a third time
+STEP: deleting the configmap
+STEP: Expecting to observe an add notification for the watched object when the label value was restored
+Oct 26 16:31:23.609: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-5004 /api/v1/namespaces/watch-5004/configmaps/e2e-watch-test-label-changed f47bf8f8-834b-4c73-9291-1a2a9d931579 62307 0 2020-10-26 16:31:13 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Oct 26 16:31:23.609: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-5004 /api/v1/namespaces/watch-5004/configmaps/e2e-watch-test-label-changed f47bf8f8-834b-4c73-9291-1a2a9d931579 62308 0 2020-10-26 16:31:13 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
+Oct 26 16:31:23.609: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-5004 /api/v1/namespaces/watch-5004/configmaps/e2e-watch-test-label-changed f47bf8f8-834b-4c73-9291-1a2a9d931579 62309 0 2020-10-26 16:31:13 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:14:54.726: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "security-context-test-5858" for this suite.
+Oct 26 16:31:23.609: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-5004" for this suite.
 
-• [SLOW TEST:6.737 seconds]
-[k8s.io] Security Context
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  when creating containers with AllowPrivilegeEscalation
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:289
-    should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:10.415 seconds]
+[sig-api-machinery] Watchers
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":89,"skipped":1562,"failed":0}
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with configmap pod [LinuxOnly] [Conformance]
+{"msg":"PASSED [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance]","total":280,"completed":93,"skipped":1354,"failed":0}
+SSSSSSSSSSS
+------------------------------
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute poststart http hook properly [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Subpath
+[BeforeEach] [k8s.io] Container Lifecycle Hook
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:14:54.797: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename subpath
+Oct 26 16:31:23.657: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename container-lifecycle-hook
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
-STEP: Setting up data
-[It] should support subpaths with configmap pod [LinuxOnly] [Conformance]
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute poststart http hook properly [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod pod-subpath-test-configmap-j6xs
-STEP: Creating a pod to test atomic-volume-subpath
-Sep 21 16:14:55.264: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-j6xs" in namespace "subpath-9201" to be "success or failure"
-Sep 21 16:14:55.289: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Pending", Reason="", readiness=false. Elapsed: 25.344571ms
-Sep 21 16:14:57.325: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Pending", Reason="", readiness=false. Elapsed: 2.061064264s
-Sep 21 16:14:59.348: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Running", Reason="", readiness=true. Elapsed: 4.084202292s
-Sep 21 16:15:01.379: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Running", Reason="", readiness=true. Elapsed: 6.115073928s
-Sep 21 16:15:03.414: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Running", Reason="", readiness=true. Elapsed: 8.15016044s
-Sep 21 16:15:05.440: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Running", Reason="", readiness=true. Elapsed: 10.175771481s
-Sep 21 16:15:07.465: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Running", Reason="", readiness=true. Elapsed: 12.200712745s
-Sep 21 16:15:09.490: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Running", Reason="", readiness=true. Elapsed: 14.225657589s
-Sep 21 16:15:11.513: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Running", Reason="", readiness=true. Elapsed: 16.248949029s
-Sep 21 16:15:13.536: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Running", Reason="", readiness=true. Elapsed: 18.272301563s
-Sep 21 16:15:15.776: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Running", Reason="", readiness=true. Elapsed: 20.511873186s
-Sep 21 16:15:17.947: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Running", Reason="", readiness=true. Elapsed: 22.682579076s
-Sep 21 16:15:19.971: INFO: Pod "pod-subpath-test-configmap-j6xs": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.707239417s
-STEP: Saw pod success
-Sep 21 16:15:19.971: INFO: Pod "pod-subpath-test-configmap-j6xs" satisfied condition "success or failure"
-Sep 21 16:15:19.995: INFO: Trying to get logs from node 10.241.51.150 pod pod-subpath-test-configmap-j6xs container test-container-subpath-configmap-j6xs: 
-STEP: delete the pod
-Sep 21 16:15:20.167: INFO: Waiting for pod pod-subpath-test-configmap-j6xs to disappear
-Sep 21 16:15:20.191: INFO: Pod pod-subpath-test-configmap-j6xs no longer exists
-STEP: Deleting pod pod-subpath-test-configmap-j6xs
-Sep 21 16:15:20.191: INFO: Deleting pod "pod-subpath-test-configmap-j6xs" in namespace "subpath-9201"
-[AfterEach] [sig-storage] Subpath
+STEP: create the pod with lifecycle hook
+STEP: check poststart hook
+STEP: delete the pod with lifecycle hook
+Oct 26 16:31:32.105: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Oct 26 16:31:32.122: INFO: Pod pod-with-poststart-http-hook still exists
+Oct 26 16:31:34.122: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Oct 26 16:31:34.140: INFO: Pod pod-with-poststart-http-hook still exists
+Oct 26 16:31:36.123: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Oct 26 16:31:36.139: INFO: Pod pod-with-poststart-http-hook still exists
+Oct 26 16:31:38.122: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Oct 26 16:31:38.138: INFO: Pod pod-with-poststart-http-hook no longer exists
+[AfterEach] [k8s.io] Container Lifecycle Hook
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:15:20.214: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "subpath-9201" for this suite.
+Oct 26 16:31:38.138: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-8398" for this suite.
 
-• [SLOW TEST:25.482 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
-  Atomic writer volumes
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
-    should support subpaths with configmap pod [LinuxOnly] [Conformance]
+• [SLOW TEST:14.533 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  when create a pod with lifecycle hook
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
+    should execute poststart http hook properly [NodeConformance] [Conformance]
     /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [LinuxOnly] [Conformance]","total":280,"completed":90,"skipped":1562,"failed":0}
-SSSSSSSSS
+{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance]","total":280,"completed":94,"skipped":1365,"failed":0}
+SSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+[k8s.io] KubeletManagedEtcHosts 
+  should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [k8s.io] KubeletManagedEtcHosts
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:15:20.280: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename emptydir
+Oct 26 16:31:38.190: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+[It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir volume type on node default medium
-Sep 21 16:15:20.705: INFO: Waiting up to 5m0s for pod "pod-9f76f600-1552-442c-8bdb-40597774f68a" in namespace "emptydir-5278" to be "success or failure"
-Sep 21 16:15:20.730: INFO: Pod "pod-9f76f600-1552-442c-8bdb-40597774f68a": Phase="Pending", Reason="", readiness=false. Elapsed: 25.028341ms
-Sep 21 16:15:22.757: INFO: Pod "pod-9f76f600-1552-442c-8bdb-40597774f68a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.051541926s
-Sep 21 16:15:24.794: INFO: Pod "pod-9f76f600-1552-442c-8bdb-40597774f68a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.08920106s
-STEP: Saw pod success
-Sep 21 16:15:24.795: INFO: Pod "pod-9f76f600-1552-442c-8bdb-40597774f68a" satisfied condition "success or failure"
-Sep 21 16:15:24.829: INFO: Trying to get logs from node 10.241.51.147 pod pod-9f76f600-1552-442c-8bdb-40597774f68a container test-container: 
-STEP: delete the pod
-Sep 21 16:15:25.059: INFO: Waiting for pod pod-9f76f600-1552-442c-8bdb-40597774f68a to disappear
-Sep 21 16:15:25.099: INFO: Pod pod-9f76f600-1552-442c-8bdb-40597774f68a no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+STEP: Setting up the test
+STEP: Creating hostNetwork=false pod
+STEP: Creating hostNetwork=true pod
+STEP: Running the test
+STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false
+Oct 26 16:31:44.579: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-5614 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:31:44.579: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:31:44.831: INFO: Exec stderr: ""
+Oct 26 16:31:44.831: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-5614 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:31:44.832: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:31:45.115: INFO: Exec stderr: ""
+Oct 26 16:31:45.115: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-5614 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:31:45.115: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:31:45.365: INFO: Exec stderr: ""
+Oct 26 16:31:45.365: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-5614 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:31:45.365: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:31:45.607: INFO: Exec stderr: ""
+STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount
+Oct 26 16:31:45.607: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-5614 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:31:45.607: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:31:45.869: INFO: Exec stderr: ""
+Oct 26 16:31:45.869: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-5614 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:31:45.869: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:31:46.095: INFO: Exec stderr: ""
+STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true
+Oct 26 16:31:46.095: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-5614 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:31:46.095: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:31:46.338: INFO: Exec stderr: ""
+Oct 26 16:31:46.338: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-5614 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:31:46.338: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:31:46.558: INFO: Exec stderr: ""
+Oct 26 16:31:46.558: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-5614 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:31:46.558: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:31:46.780: INFO: Exec stderr: ""
+Oct 26 16:31:46.781: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-5614 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:31:46.781: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:31:47.029: INFO: Exec stderr: ""
+[AfterEach] [k8s.io] KubeletManagedEtcHosts
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:15:25.099: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-5278" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":91,"skipped":1571,"failed":0}
-SSSSSSSSSSSSS
+Oct 26 16:31:47.029: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-kubelet-etc-hosts-5614" for this suite.
+
+• [SLOW TEST:8.878 seconds]
+[k8s.io] KubeletManagedEtcHosts
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [k8s.io] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":95,"skipped":1381,"failed":0}
+S
 ------------------------------
 [k8s.io] InitContainer [NodeConformance] 
   should not start app containers if init containers fail on a RestartAlways pod [Conformance]
@@ -4616,7 +4997,7 @@ SSSSSSSSSSSSS
 [BeforeEach] [k8s.io] InitContainer [NodeConformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:15:25.168: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 16:31:47.068: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename init-container
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [k8s.io] InitContainer [NodeConformance]
@@ -4624,3286 +5005,2954 @@ STEP: Waiting for a default service account to be provisioned in namespace
 [It] should not start app containers if init containers fail on a RestartAlways pod [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: creating the pod
-Sep 21 16:15:25.560: INFO: PodSpec: initContainers in spec.initContainers
-Sep 21 16:16:09.051: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-3757636b-592c-41f1-b40a-5220a126eb38", GenerateName:"", Namespace:"init-container-9033", SelfLink:"/api/v1/namespaces/init-container-9033/pods/pod-init-3757636b-592c-41f1-b40a-5220a126eb38", UID:"3c920348-81ba-493a-b620-38048eb7a644", ResourceVersion:"62848", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63736301725, loc:(*time.Location)(0x7db5bc0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"560547515"}, Annotations:map[string]string{"cni.projectcalico.org/podIP":"172.30.237.124/32", "cni.projectcalico.org/podIPs":"172.30.237.124/32", "k8s.v1.cni.cncf.io/networks-status":"[{\n    \"name\": \"k8s-pod-network\",\n    \"ips\": [\n        \"172.30.237.124\"\n    ],\n    \"dns\": {}\n}]", "openshift.io/scc":"anyuid"}, OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-wbd4k", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc00099a940), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-wbd4k", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(0xc001daa0f0), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-wbd4k", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(0xc001daa190), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-wbd4k", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(0xc001daa050), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc002eee088), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"10.241.51.147", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc0022f4120), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc002eee140)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc002eee160)}, v1.Toleration{Key:"node.kubernetes.io/memory-pressure", Operator:"Exists", Value:"", Effect:"NoSchedule", TolerationSeconds:(*int64)(nil)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc002eee17c), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc002eee180), PreemptionPolicy:(*v1.PreemptionPolicy)(nil), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301725, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301725, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301725, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736301725, loc:(*time.Location)(0x7db5bc0)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"10.241.51.147", PodIP:"172.30.237.124", PodIPs:[]v1.PodIP{v1.PodIP{IP:"172.30.237.124"}}, StartTime:(*v1.Time)(0xc003556060), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc00339c070)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc00339c0e0)}, Ready:false, RestartCount:3, Image:"docker.io/library/busybox:1.29", ImageID:"docker.io/library/busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"cri-o://3289c969092c246acb0370d97d09ad4859575933b3d993eb06a34e68273d3395", Started:(*bool)(nil)}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc0035560a0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:"", Started:(*bool)(nil)}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc003556080), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.1", ImageID:"", ContainerID:"", Started:(*bool)(0xc002eee1f4)}}, QOSClass:"Burstable", EphemeralContainerStatuses:[]v1.ContainerStatus(nil)}}
+Oct 26 16:31:47.201: INFO: PodSpec: initContainers in spec.initContainers
+Oct 26 16:32:36.741: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-5f7d7a26-cf12-4a21-bd2f-0572be91ffbd", GenerateName:"", Namespace:"init-container-2603", SelfLink:"/api/v1/namespaces/init-container-2603/pods/pod-init-5f7d7a26-cf12-4a21-bd2f-0572be91ffbd", UID:"4ef4ace7-062e-4256-9e1f-24acd18b4356", ResourceVersion:"62946", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63739326707, loc:(*time.Location)(0x7db5bc0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"201857810"}, Annotations:map[string]string{"cni.projectcalico.org/podIP":"172.30.54.183/32", "cni.projectcalico.org/podIPs":"172.30.54.183/32", "k8s.v1.cni.cncf.io/networks-status":"[{\n    \"name\": \"k8s-pod-network\",\n    \"ips\": [\n        \"172.30.54.183\"\n    ],\n    \"dns\": {}\n}]", "openshift.io/scc":"anyuid"}, OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-q59kh", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc00e60b300), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-q59kh", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(0xc003606b40), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-q59kh", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(0xc003606be0), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-q59kh", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(0xc003606aa0), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc00d8b8ce8), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"10.72.119.82", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc0079b0420), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc00d8b8da0)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc00d8b8dc0)}, v1.Toleration{Key:"node.kubernetes.io/memory-pressure", Operator:"Exists", Value:"", Effect:"NoSchedule", TolerationSeconds:(*int64)(nil)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc00d8b8ddc), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc00d8b8de0), PreemptionPolicy:(*v1.PreemptionPolicy)(nil), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326707, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326707, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326707, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326707, loc:(*time.Location)(0x7db5bc0)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"10.72.119.82", PodIP:"172.30.54.183", PodIPs:[]v1.PodIP{v1.PodIP{IP:"172.30.54.183"}}, StartTime:(*v1.Time)(0xc00478c5c0), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc0008c7030)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:3, Image:"docker.io/library/busybox:1.29", ImageID:"docker.io/library/busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"cri-o://13e76caaa8ab0e1ba61e4fa0603bb47501b07af0185ef59f42f5140940392f0f", Started:(*bool)(nil)}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc00478c600), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:"", Started:(*bool)(nil)}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc00478c5e0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.1", ImageID:"", ContainerID:"", Started:(*bool)(0xc00d8b8eaf)}}, QOSClass:"Burstable", EphemeralContainerStatuses:[]v1.ContainerStatus(nil)}}
 [AfterEach] [k8s.io] InitContainer [NodeConformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:16:09.052: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "init-container-9033" for this suite.
+Oct 26 16:32:36.742: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "init-container-2603" for this suite.
 
-• [SLOW TEST:43.955 seconds]
+• [SLOW TEST:49.713 seconds]
 [k8s.io] InitContainer [NodeConformance]
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
   should not start app containers if init containers fail on a RestartAlways pod [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance]","total":280,"completed":92,"skipped":1584,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance]","total":280,"completed":96,"skipped":1382,"failed":0}
+SS
 ------------------------------
-[sig-network] DNS 
-  should support configurable pod DNS nameservers [Conformance]
+[sig-cli] Kubectl client Kubectl run default 
+  should create an rc or deployment from an image  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] DNS
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:16:09.123: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename dns
+Oct 26 16:32:36.783: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubectl
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support configurable pod DNS nameservers [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
+[BeforeEach] Kubectl run default
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1596
+[It] should create an rc or deployment from an image  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod with dnsPolicy=None and customized dnsConfig...
-Sep 21 16:16:09.633: INFO: Created pod &Pod{ObjectMeta:{dns-9547  dns-9547 /api/v1/namespaces/dns-9547/pods/dns-9547 80046894-a15d-46c9-b009-faafe0e98739 62881 0 2020-09-21 16:16:09 +0000 UTC   map[] map[openshift.io/scc:anyuid] [] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-t62r6,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-t62r6,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[pause],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-t62r6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:None,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:&SELinuxOptions{User:,Role:,Type:,Level:s0:c52,c39,},RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:&PodDNSConfig{Nameservers:[1.1.1.1],Searches:[resolv.conf.local],Options:[]PodDNSConfigOption{},},ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-STEP: Verifying customized DNS suffix list is configured on pod...
-Sep 21 16:16:11.682: INFO: ExecWithOptions {Command:[/agnhost dns-suffix] Namespace:dns-9547 PodName:dns-9547 ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:16:11.682: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Verifying customized DNS server is configured on pod...
-Sep 21 16:16:12.018: INFO: ExecWithOptions {Command:[/agnhost dns-server-list] Namespace:dns-9547 PodName:dns-9547 ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:16:12.018: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:16:12.281: INFO: Deleting pod dns-9547...
-[AfterEach] [sig-network] DNS
+STEP: running the image docker.io/library/httpd:2.4.38-alpine
+Oct 26 16:32:36.927: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 run e2e-test-httpd-deployment --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-586'
+Oct 26 16:32:37.111: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Oct 26 16:32:37.111: INFO: stdout: "deployment.apps/e2e-test-httpd-deployment created\n"
+STEP: verifying the pod controlled by e2e-test-httpd-deployment gets created
+[AfterEach] Kubectl run default
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1602
+Oct 26 16:32:39.143: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete deployment e2e-test-httpd-deployment --namespace=kubectl-586'
+Oct 26 16:32:39.308: INFO: stderr: ""
+Oct 26 16:32:39.308: INFO: stdout: "deployment.apps \"e2e-test-httpd-deployment\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:16:12.593: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-9547" for this suite.
-•{"msg":"PASSED [sig-network] DNS should support configurable pod DNS nameservers [Conformance]","total":280,"completed":93,"skipped":1619,"failed":0}
-SSSSSSSSSSSSS
+Oct 26 16:32:39.308: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-586" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run default should create an rc or deployment from an image  [Conformance]","total":280,"completed":97,"skipped":1384,"failed":0}
+SSSSSSS
 ------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+[sig-storage] Projected downwardAPI 
+  should provide container's memory request [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
+[BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:16:12.671: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename secrets
+Oct 26 16:32:39.358: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should provide container's memory request [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-6f42ff40-f8e2-4c2a-aa5c-8def677b41fc
-STEP: Creating a pod to test consume secrets
-Sep 21 16:16:13.331: INFO: Waiting up to 5m0s for pod "pod-secrets-6228b2f4-c2f5-4edb-9171-acd12318ac99" in namespace "secrets-5600" to be "success or failure"
-Sep 21 16:16:13.365: INFO: Pod "pod-secrets-6228b2f4-c2f5-4edb-9171-acd12318ac99": Phase="Pending", Reason="", readiness=false. Elapsed: 33.616042ms
-Sep 21 16:16:15.387: INFO: Pod "pod-secrets-6228b2f4-c2f5-4edb-9171-acd12318ac99": Phase="Pending", Reason="", readiness=false. Elapsed: 2.056309437s
-Sep 21 16:16:17.423: INFO: Pod "pod-secrets-6228b2f4-c2f5-4edb-9171-acd12318ac99": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.092239277s
+STEP: Creating a pod to test downward API volume plugin
+Oct 26 16:32:39.545: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fa18c518-6508-4860-ac20-ef8e0f13987d" in namespace "projected-5698" to be "success or failure"
+Oct 26 16:32:39.570: INFO: Pod "downwardapi-volume-fa18c518-6508-4860-ac20-ef8e0f13987d": Phase="Pending", Reason="", readiness=false. Elapsed: 25.470992ms
+Oct 26 16:32:41.589: INFO: Pod "downwardapi-volume-fa18c518-6508-4860-ac20-ef8e0f13987d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.043793171s
 STEP: Saw pod success
-Sep 21 16:16:17.423: INFO: Pod "pod-secrets-6228b2f4-c2f5-4edb-9171-acd12318ac99" satisfied condition "success or failure"
-Sep 21 16:16:17.456: INFO: Trying to get logs from node 10.241.51.147 pod pod-secrets-6228b2f4-c2f5-4edb-9171-acd12318ac99 container secret-volume-test: 
+Oct 26 16:32:41.589: INFO: Pod "downwardapi-volume-fa18c518-6508-4860-ac20-ef8e0f13987d" satisfied condition "success or failure"
+Oct 26 16:32:41.604: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-fa18c518-6508-4860-ac20-ef8e0f13987d container client-container: 
 STEP: delete the pod
-Sep 21 16:16:17.650: INFO: Waiting for pod pod-secrets-6228b2f4-c2f5-4edb-9171-acd12318ac99 to disappear
-Sep 21 16:16:17.682: INFO: Pod pod-secrets-6228b2f4-c2f5-4edb-9171-acd12318ac99 no longer exists
-[AfterEach] [sig-storage] Secrets
+Oct 26 16:32:41.686: INFO: Waiting for pod downwardapi-volume-fa18c518-6508-4860-ac20-ef8e0f13987d to disappear
+Oct 26 16:32:41.701: INFO: Pod downwardapi-volume-fa18c518-6508-4860-ac20-ef8e0f13987d no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:16:17.682: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-5600" for this suite.
-
-• [SLOW TEST:5.089 seconds]
-[sig-storage] Secrets
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
-  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":94,"skipped":1632,"failed":0}
-SSSSSSSSSSSS
+Oct 26 16:32:41.701: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-5698" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance]","total":280,"completed":98,"skipped":1391,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] ReplicationController 
-  should serve a basic image on each replica with a public image  [Conformance]
+[sig-cli] Kubectl client Kubectl run rc 
+  should create an rc from an image  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] ReplicationController
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:16:17.762: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename replication-controller
+Oct 26 16:32:41.739: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubectl
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should serve a basic image on each replica with a public image  [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
+[BeforeEach] Kubectl run rc
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1632
+[It] should create an rc from an image  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating replication controller my-hostname-basic-2990444e-9f47-4040-b573-4ceefb59dbcf
-Sep 21 16:16:18.082: INFO: Pod name my-hostname-basic-2990444e-9f47-4040-b573-4ceefb59dbcf: Found 0 pods out of 1
-Sep 21 16:16:23.114: INFO: Pod name my-hostname-basic-2990444e-9f47-4040-b573-4ceefb59dbcf: Found 1 pods out of 1
-Sep 21 16:16:23.114: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-2990444e-9f47-4040-b573-4ceefb59dbcf" are running
-Sep 21 16:16:23.148: INFO: Pod "my-hostname-basic-2990444e-9f47-4040-b573-4ceefb59dbcf-sldkv" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-09-21 16:16:18 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-09-21 16:16:20 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-09-21 16:16:20 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-09-21 16:16:18 +0000 UTC Reason: Message:}])
-Sep 21 16:16:23.149: INFO: Trying to dial the pod
-Sep 21 16:16:28.255: INFO: Controller my-hostname-basic-2990444e-9f47-4040-b573-4ceefb59dbcf: Got expected result from replica 1 [my-hostname-basic-2990444e-9f47-4040-b573-4ceefb59dbcf-sldkv]: "my-hostname-basic-2990444e-9f47-4040-b573-4ceefb59dbcf-sldkv", 1 of 1 required successes so far
-[AfterEach] [sig-apps] ReplicationController
+STEP: running the image docker.io/library/httpd:2.4.38-alpine
+Oct 26 16:32:41.878: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 run e2e-test-httpd-rc --image=docker.io/library/httpd:2.4.38-alpine --generator=run/v1 --namespace=kubectl-6488'
+Oct 26 16:32:42.045: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Oct 26 16:32:42.045: INFO: stdout: "replicationcontroller/e2e-test-httpd-rc created\n"
+STEP: verifying the rc e2e-test-httpd-rc was created
+STEP: verifying the pod controlled by rc e2e-test-httpd-rc was created
+STEP: confirm that you can get logs from an rc
+Oct 26 16:32:44.087: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [e2e-test-httpd-rc-c5c9j]
+Oct 26 16:32:44.087: INFO: Waiting up to 5m0s for pod "e2e-test-httpd-rc-c5c9j" in namespace "kubectl-6488" to be "running and ready"
+Oct 26 16:32:44.103: INFO: Pod "e2e-test-httpd-rc-c5c9j": Phase="Pending", Reason="", readiness=false. Elapsed: 16.429601ms
+Oct 26 16:32:46.122: INFO: Pod "e2e-test-httpd-rc-c5c9j": Phase="Running", Reason="", readiness=true. Elapsed: 2.035069107s
+Oct 26 16:32:46.122: INFO: Pod "e2e-test-httpd-rc-c5c9j" satisfied condition "running and ready"
+Oct 26 16:32:46.122: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [e2e-test-httpd-rc-c5c9j]
+Oct 26 16:32:46.122: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 logs rc/e2e-test-httpd-rc --namespace=kubectl-6488'
+Oct 26 16:32:46.318: INFO: stderr: ""
+Oct 26 16:32:46.318: INFO: stdout: "AH00558: httpd: Could not reliably determine the server's fully qualified domain name, using 172.30.194.66. Set the 'ServerName' directive globally to suppress this message\nAH00558: httpd: Could not reliably determine the server's fully qualified domain name, using 172.30.194.66. Set the 'ServerName' directive globally to suppress this message\n[Mon Oct 26 16:32:43.630210 2020] [mpm_event:notice] [pid 1:tid 140313223678824] AH00489: Apache/2.4.38 (Unix) configured -- resuming normal operations\n[Mon Oct 26 16:32:43.630310 2020] [core:notice] [pid 1:tid 140313223678824] AH00094: Command line: 'httpd -D FOREGROUND'\n"
+[AfterEach] Kubectl run rc
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1637
+Oct 26 16:32:46.318: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete rc e2e-test-httpd-rc --namespace=kubectl-6488'
+Oct 26 16:32:46.490: INFO: stderr: ""
+Oct 26 16:32:46.491: INFO: stdout: "replicationcontroller \"e2e-test-httpd-rc\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:16:28.255: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replication-controller-4247" for this suite.
-
-• [SLOW TEST:10.561 seconds]
-[sig-apps] ReplicationController
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] ReplicationController should serve a basic image on each replica with a public image  [Conformance]","total":280,"completed":95,"skipped":1644,"failed":0}
+Oct 26 16:32:46.491: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-6488" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run rc should create an rc from an image  [Conformance]","total":280,"completed":99,"skipped":1421,"failed":0}
 SSS
 ------------------------------
 [sig-network] Services 
-  should serve multiport endpoints from pods  [Conformance]
+  should be able to change the type from NodePort to ExternalName [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:16:28.325: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 16:32:46.536: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename services
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should serve multiport endpoints from pods  [Conformance]
+[It] should be able to change the type from NodePort to ExternalName [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating service multi-endpoint-test in namespace services-4734
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4734 to expose endpoints map[]
-Sep 21 16:16:28.769: INFO: Get endpoints failed (26.867982ms elapsed, ignoring for 5s): endpoints "multi-endpoint-test" not found
-Sep 21 16:16:29.800: INFO: successfully validated that service multi-endpoint-test in namespace services-4734 exposes endpoints map[] (1.057825179s elapsed)
-STEP: Creating pod pod1 in namespace services-4734
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4734 to expose endpoints map[pod1:[100]]
-Sep 21 16:16:32.650: INFO: successfully validated that service multi-endpoint-test in namespace services-4734 exposes endpoints map[pod1:[100]] (2.747036s elapsed)
-STEP: Creating pod pod2 in namespace services-4734
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4734 to expose endpoints map[pod1:[100] pod2:[101]]
-Sep 21 16:16:37.084: INFO: successfully validated that service multi-endpoint-test in namespace services-4734 exposes endpoints map[pod1:[100] pod2:[101]] (4.097553769s elapsed)
-STEP: Deleting pod pod1 in namespace services-4734
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4734 to expose endpoints map[pod2:[101]]
-Sep 21 16:16:37.806: INFO: successfully validated that service multi-endpoint-test in namespace services-4734 exposes endpoints map[pod2:[101]] (270.525207ms elapsed)
-STEP: Deleting pod pod2 in namespace services-4734
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4734 to expose endpoints map[]
-Sep 21 16:16:37.895: INFO: successfully validated that service multi-endpoint-test in namespace services-4734 exposes endpoints map[] (26.524565ms elapsed)
+STEP: creating a service nodeport-service with the type=NodePort in namespace services-3321
+STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service
+STEP: creating service externalsvc in namespace services-3321
+STEP: creating replication controller externalsvc in namespace services-3321
+I1026 16:32:46.816034      26 runners.go:189] Created replication controller with name: externalsvc, namespace: services-3321, replica count: 2
+I1026 16:32:49.866877      26 runners.go:189] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+STEP: changing the NodePort service to type=ExternalName
+Oct 26 16:32:49.942: INFO: Creating new exec pod
+Oct 26 16:32:52.033: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-3321 execpod9687v -- /bin/sh -x -c nslookup nodeport-service'
+Oct 26 16:32:52.452: INFO: stderr: "+ nslookup nodeport-service\n"
+Oct 26 16:32:52.452: INFO: stdout: "Server:\t\t172.21.0.10\nAddress:\t172.21.0.10#53\n\nnodeport-service.services-3321.svc.cluster.local\tcanonical name = externalsvc.services-3321.svc.cluster.local.\nName:\texternalsvc.services-3321.svc.cluster.local\nAddress: 172.21.92.151\n\n"
+STEP: deleting ReplicationController externalsvc in namespace services-3321, will wait for the garbage collector to delete the pods
+Oct 26 16:32:52.549: INFO: Deleting ReplicationController externalsvc took: 32.666446ms
+Oct 26 16:32:52.650: INFO: Terminating ReplicationController externalsvc pods took: 100.506376ms
+Oct 26 16:33:06.822: INFO: Cleaning up the NodePort to ExternalName test service
 [AfterEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:16:38.057: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-4734" for this suite.
+Oct 26 16:33:06.886: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-3321" for this suite.
 [AfterEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
 
-• [SLOW TEST:9.804 seconds]
+• [SLOW TEST:20.396 seconds]
 [sig-network] Services
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should serve multiport endpoints from pods  [Conformance]
+  should be able to change the type from NodePort to ExternalName [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance]","total":280,"completed":100,"skipped":1424,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should deny crd creation [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 16:33:06.932: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename webhook
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Oct 26 16:33:07.711: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 16:33:09.758: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326787, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326787, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326787, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326787, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Oct 26 16:33:12.805: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should deny crd creation [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Registering the crd webhook via the AdmissionRegistration API
+STEP: Creating a custom resource definition that should be denied by the webhook
+Oct 26 16:33:12.879: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 16:33:12.952: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-5143" for this suite.
+STEP: Destroying namespace "webhook-5143-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+
+• [SLOW TEST:6.286 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should deny crd creation [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance]","total":280,"completed":101,"skipped":1456,"failed":0}
+SSSSSSSSSSSSSSSS
+------------------------------
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 16:33:13.218: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename sched-pred
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
+Oct 26 16:33:13.372: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Oct 26 16:33:13.427: INFO: Waiting for terminating namespaces to be deleted...
+Oct 26 16:33:13.448: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.74 before test
+Oct 26 16:33:13.539: INFO: calico-typha-7fd49c779d-46t4v from calico-system started at 2020-10-26 14:37:36 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: sonobuoy from sonobuoy started at 2020-10-26 15:59:53 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: sonobuoy-e2e-job-c6d919d638514ce4 from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container e2e ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: thanos-querier-84d7d58bc-c27ss from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (4 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container oauth-proxy ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container thanos-querier ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: grafana-f999495bf-t4tcv from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container grafana ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container grafana-proxy ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: telemeter-client-d5d9896c4-jpd8z from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (3 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container reload ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container telemeter-client ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: ibmcloud-block-storage-driver-ws85j from kube-system started at 2020-10-26 14:34:32 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: node-exporter-ls6pf from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: sample-webhook-deployment-5f65f8c764-x4nv5 from webhook-5143 started at 2020-10-26 16:33:07 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container sample-webhook ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: dns-default-s9kvh from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container dns ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: prometheus-adapter-674456c5d6-p6bj5 from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container prometheus-adapter ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: console-5dd98d6f6f-zxw5x from openshift-console started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container console ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: calico-node-xlzf4 from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: cluster-samples-operator-848994bd88-rw4rq from openshift-cluster-samples-operator started at 2020-10-26 16:09:41 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container cluster-samples-operator ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container cluster-samples-operator-watch ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: ibm-keepalived-watcher-49jbs from kube-system started at 2020-10-26 14:34:28 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-nbr6w from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: service-ca-7d9db9c5b8-4w4cj from openshift-service-ca started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container service-ca-controller ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: openshift-kube-proxy-7d26g from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: tuned-p4cpc from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: node-ca-4ppdl from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: ibm-master-proxy-static-10.72.119.74 from kube-system started at 2020-10-26 14:34:26 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container pause ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: multus-cc44t from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: alertmanager-main-0 from openshift-monitoring started at 2020-10-26 16:09:57 +0000 UTC (3 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: multus-admission-controller-dnt9r from openshift-multus started at 2020-10-26 16:10:16 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: redhat-operators-5b446b45df-fgb5b from openshift-marketplace started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container redhat-operators ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: packageserver-5f76c64df7-jg7sj from openshift-operator-lifecycle-manager started at 2020-10-26 16:10:08 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container packageserver ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: vpn-5b455f6f88-64q7j from kube-system started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container vpn ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: tigera-operator-5c45c894b4-z9bnx from tigera-operator started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.539: INFO: 	Container tigera-operator ready: true, restart count 0
+Oct 26 16:33:13.539: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.76 before test
+Oct 26 16:33:13.700: INFO: calico-typha-7fd49c779d-766qb from calico-system started at 2020-10-26 14:37:36 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: dns-default-z8pn9 from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container dns ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: openshift-kube-proxy-74ss4 from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: ibm-file-plugin-944b59c94-k7mrm from kube-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: alertmanager-main-2 from openshift-monitoring started at 2020-10-26 14:44:28 +0000 UTC (3 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-2scb4 from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: router-default-84766ff4fd-cftxw from openshift-ingress started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container router ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: ibmcloud-block-storage-driver-b724k from kube-system started at 2020-10-26 14:34:34 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: openshift-service-catalog-apiserver-operator-7cfbcc49dc-hrn66 from openshift-service-catalog-apiserver-operator started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container operator ready: true, restart count 1
+Oct 26 16:33:13.700: INFO: ibm-storage-watcher-77d6cc685c-bvg8q from kube-system started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: olm-operator-7cd6cbf579-zww2l from openshift-operator-lifecycle-manager started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container olm-operator ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: cluster-monitoring-operator-69649564f8-r4h9b from openshift-monitoring started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container cluster-monitoring-operator ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: multus-lbpcn from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.700: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 16:33:13.700: INFO: openshift-service-catalog-controller-manager-operator-d49bz89wp from openshift-service-catalog-controller-manager-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container operator ready: true, restart count 1
+Oct 26 16:33:13.701: INFO: multus-admission-controller-5j529 from openshift-multus started at 2020-10-26 14:35:50 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: node-exporter-zmwjz from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: node-ca-bpxm5 from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: ibm-cloud-provider-ip-158-176-78-170-5cdb49c95f-76fjw from ibm-system started at 2020-10-26 14:42:32 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container ibm-cloud-provider-ip-158-176-78-170 ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: ingress-operator-65878dcc8d-wdnrl from openshift-ingress-operator started at 2020-10-26 14:35:49 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container ingress-operator ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: downloads-74f6b6dcb6-r9ksv from openshift-console started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container download-server ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: catalog-operator-5db6d5849c-z2lgx from openshift-operator-lifecycle-manager started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container catalog-operator ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: calico-kube-controllers-cdc7f7f8f-jqklm from calico-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container calico-kube-controllers ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: downloads-74f6b6dcb6-vqzf9 from openshift-console started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container download-server ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: cluster-node-tuning-operator-575b98978c-zz7s9 from openshift-cluster-node-tuning-operator started at 2020-10-26 14:35:51 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container cluster-node-tuning-operator ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: calico-node-dhd7p from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: kube-storage-version-migrator-operator-86d479ffb5-9cmml from openshift-kube-storage-version-migrator-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container kube-storage-version-migrator-operator ready: true, restart count 1
+Oct 26 16:33:13.701: INFO: ibm-keepalived-watcher-qwg48 from kube-system started at 2020-10-26 14:34:30 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: ibmcloud-block-storage-plugin-5f486d686f-6brjv from kube-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container ibmcloud-block-storage-plugin-container ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: prometheus-k8s-0 from openshift-monitoring started at 2020-10-26 14:45:39 +0000 UTC (7 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: 	Container prometheus ready: true, restart count 1
+Oct 26 16:33:13.701: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: 	Container prometheus-proxy ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: 	Container thanos-sidecar ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: ibm-master-proxy-static-10.72.119.76 from kube-system started at 2020-10-26 14:34:28 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: 	Container pause ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: cluster-image-registry-operator-8444cfb976-gdjk6 from openshift-image-registry started at 2020-10-26 14:35:50 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container cluster-image-registry-operator ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: 	Container cluster-image-registry-operator-watch ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: cluster-storage-operator-9b57569c9-gpqd2 from openshift-cluster-storage-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container cluster-storage-operator ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: service-ca-operator-656f67898-svtt9 from openshift-service-ca-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container operator ready: true, restart count 1
+Oct 26 16:33:13.701: INFO: marketplace-operator-5df598b96b-6vr64 from openshift-marketplace started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container marketplace-operator ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: console-operator-5fc674d8d4-fjctd from openshift-console-operator started at 2020-10-26 14:35:51 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container console-operator ready: true, restart count 1
+Oct 26 16:33:13.701: INFO: tuned-s6flg from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: network-operator-86b67c6f7f-pcq2w from openshift-network-operator started at 2020-10-26 14:34:37 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container network-operator ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: dns-operator-58d86f779c-wbm98 from openshift-dns-operator started at 2020-10-26 14:35:49 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.701: INFO: 	Container dns-operator ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.701: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.82 before test
+Oct 26 16:33:13.870: INFO: prometheus-k8s-1 from openshift-monitoring started at 2020-10-26 14:45:29 +0000 UTC (7 container statuses recorded)
+Oct 26 16:33:13.870: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: 	Container prometheus ready: true, restart count 1
+Oct 26 16:33:13.871: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: 	Container prometheus-proxy ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: 	Container thanos-sidecar ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: console-5dd98d6f6f-xvz7c from openshift-console started at 2020-10-26 14:45:54 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.871: INFO: 	Container console ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: packageserver-5f76c64df7-6nrvc from openshift-operator-lifecycle-manager started at 2020-10-26 16:10:11 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.871: INFO: 	Container packageserver ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: calico-node-c7wjs from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.871: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: dns-default-ggg6q from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.871: INFO: 	Container dns ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: multus-hwkgh from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.871: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: kube-state-metrics-7bfb6f7864-vmgz9 from openshift-monitoring started at 2020-10-26 14:36:32 +0000 UTC (3 container statuses recorded)
+Oct 26 16:33:13.871: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: 	Container kube-state-metrics ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: thanos-querier-84d7d58bc-mpnzq from openshift-monitoring started at 2020-10-26 14:45:18 +0000 UTC (4 container statuses recorded)
+Oct 26 16:33:13.871: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.871: INFO: 	Container oauth-proxy ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: 	Container thanos-querier ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: community-operators-f4df99f64-twf5k from openshift-marketplace started at 2020-10-26 14:38:39 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.872: INFO: 	Container community-operators ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: ibm-master-proxy-static-10.72.119.82 from kube-system started at 2020-10-26 14:34:29 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.872: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: 	Container pause ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: ibmcloud-block-storage-driver-sxlj2 from kube-system started at 2020-10-26 14:34:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.872: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: node-ca-z564l from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.872: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: certified-operators-69bb9f65cc-9mhpt from openshift-marketplace started at 2020-10-26 14:38:38 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.872: INFO: 	Container certified-operators ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: redhat-marketplace-77d9b69f69-qb5s9 from openshift-marketplace started at 2020-10-26 14:38:39 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.872: INFO: 	Container redhat-marketplace ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: alertmanager-main-1 from openshift-monitoring started at 2020-10-26 14:44:39 +0000 UTC (3 container statuses recorded)
+Oct 26 16:33:13.872: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: ibm-keepalived-watcher-98n85 from kube-system started at 2020-10-26 14:34:32 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.872: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 16:33:13.872: INFO: tuned-h578f from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.873: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 16:33:13.873: INFO: registry-pvc-permissions-m9wwv from openshift-image-registry started at 2020-10-26 14:40:09 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.873: INFO: 	Container pvc-permissions ready: false, restart count 0
+Oct 26 16:33:13.873: INFO: calico-typha-7fd49c779d-nmzbk from calico-system started at 2020-10-26 14:35:39 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.873: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 16:33:13.873: INFO: test-k8s-e2e-pvg-master-verification from default started at 2020-10-26 14:37:46 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.873: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
+Oct 26 16:33:13.873: INFO: node-exporter-4lfg7 from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.873: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.873: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 16:33:13.873: INFO: migrator-686fc6cc66-9wpj9 from openshift-kube-storage-version-migrator started at 2020-10-26 14:36:53 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.873: INFO: 	Container migrator ready: true, restart count 0
+Oct 26 16:33:13.873: INFO: prometheus-adapter-674456c5d6-kj77x from openshift-monitoring started at 2020-10-26 14:44:24 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.873: INFO: 	Container prometheus-adapter ready: true, restart count 0
+Oct 26 16:33:13.873: INFO: image-registry-6db5c967cf-d4ljt from openshift-image-registry started at 2020-10-26 14:40:09 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.873: INFO: 	Container registry ready: true, restart count 0
+Oct 26 16:33:13.873: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-x2lkl from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.873: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 16:33:13.873: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 16:33:13.873: INFO: openshift-kube-proxy-z9sjt from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.873: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 16:33:13.874: INFO: openshift-state-metrics-d4d5d6f57-lszmv from openshift-monitoring started at 2020-10-26 14:36:34 +0000 UTC (3 container statuses recorded)
+Oct 26 16:33:13.874: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
+Oct 26 16:33:13.874: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
+Oct 26 16:33:13.874: INFO: 	Container openshift-state-metrics ready: true, restart count 0
+Oct 26 16:33:13.874: INFO: prometheus-operator-55f946f5d-tpnd8 from openshift-monitoring started at 2020-10-26 14:44:11 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.874: INFO: 	Container prometheus-operator ready: true, restart count 0
+Oct 26 16:33:13.874: INFO: router-default-84766ff4fd-7dgrd from openshift-ingress started at 2020-10-26 14:38:40 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.874: INFO: 	Container router ready: true, restart count 0
+Oct 26 16:33:13.874: INFO: multus-admission-controller-9rv7d from openshift-multus started at 2020-10-26 14:36:02 +0000 UTC (2 container statuses recorded)
+Oct 26 16:33:13.874: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 16:33:13.874: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 16:33:13.874: INFO: ibm-cloud-provider-ip-158-176-78-170-5cdb49c95f-89x4j from ibm-system started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 16:33:13.874: INFO: 	Container ibm-cloud-provider-ip-158-176-78-170 ready: true, restart count 0
+[It] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Trying to launch a pod without a label to get a node which can launch it.
+STEP: Explicitly delete pod here to free the resource it takes.
+STEP: Trying to apply a random label on the found node.
+STEP: verifying the node has the label kubernetes.io/e2e-0e09c432-8c4c-4a8c-a010-09b1b3996a3f 90
+STEP: Trying to create a pod(pod1) with hostport 54321 and hostIP 127.0.0.1 and expect scheduled
+STEP: Trying to create another pod(pod2) with hostport 54321 but hostIP 127.0.0.2 on the node which pod1 resides and expect scheduled
+STEP: Trying to create a third pod(pod3) with hostport 54321, hostIP 127.0.0.2 but use UDP protocol on the node which pod2 resides
+STEP: removing the label kubernetes.io/e2e-0e09c432-8c4c-4a8c-a010-09b1b3996a3f off the node 10.72.119.74
+STEP: verifying the node doesn't have the label kubernetes.io/e2e-0e09c432-8c4c-4a8c-a010-09b1b3996a3f
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 16:33:26.318: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-2638" for this suite.
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
+
+• [SLOW TEST:13.134 seconds]
+[sig-scheduling] SchedulerPredicates [Serial]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
+  validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Services should serve multiport endpoints from pods  [Conformance]","total":280,"completed":96,"skipped":1647,"failed":0}
-SSSSSSSSS
+{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]","total":280,"completed":102,"skipped":1472,"failed":0}
+SSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Variable Expansion 
-  should allow substituting values in a container's args [NodeConformance] [Conformance]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should be able to deny pod and configmap creation [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Variable Expansion
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:16:38.131: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename var-expansion
+Oct 26 16:33:26.353: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should allow substituting values in a container's args [NodeConformance] [Conformance]
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Oct 26 16:33:27.004: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 16:33:29.052: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326807, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326807, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326807, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326806, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Oct 26 16:33:32.109: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should be able to deny pod and configmap creation [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test substitution in container's args
-Sep 21 16:16:38.569: INFO: Waiting up to 5m0s for pod "var-expansion-a0c246be-30a8-4975-89f7-07cec799f630" in namespace "var-expansion-1270" to be "success or failure"
-Sep 21 16:16:38.597: INFO: Pod "var-expansion-a0c246be-30a8-4975-89f7-07cec799f630": Phase="Pending", Reason="", readiness=false. Elapsed: 27.483736ms
-Sep 21 16:16:40.620: INFO: Pod "var-expansion-a0c246be-30a8-4975-89f7-07cec799f630": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.050992288s
-STEP: Saw pod success
-Sep 21 16:16:40.621: INFO: Pod "var-expansion-a0c246be-30a8-4975-89f7-07cec799f630" satisfied condition "success or failure"
-Sep 21 16:16:40.644: INFO: Trying to get logs from node 10.241.51.147 pod var-expansion-a0c246be-30a8-4975-89f7-07cec799f630 container dapi-container: 
-STEP: delete the pod
-Sep 21 16:16:40.872: INFO: Waiting for pod var-expansion-a0c246be-30a8-4975-89f7-07cec799f630 to disappear
-Sep 21 16:16:40.899: INFO: Pod var-expansion-a0c246be-30a8-4975-89f7-07cec799f630 no longer exists
-[AfterEach] [k8s.io] Variable Expansion
+STEP: Registering the webhook via the AdmissionRegistration API
+STEP: create a pod that should be denied by the webhook
+STEP: create a pod that causes the webhook to hang
+STEP: create a configmap that should be denied by the webhook
+STEP: create a configmap that should be admitted by the webhook
+STEP: update (PUT) the admitted configmap to a non-compliant one should be rejected by the webhook
+STEP: update (PATCH) the admitted configmap to a non-compliant one should be rejected by the webhook
+STEP: create a namespace that bypass the webhook
+STEP: create a configmap that violates the webhook policy but is in a whitelisted namespace
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:16:40.899: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "var-expansion-1270" for this suite.
-•{"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance]","total":280,"completed":97,"skipped":1656,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:33:42.674: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-1846" for this suite.
+STEP: Destroying namespace "webhook-1846-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+
+• [SLOW TEST:16.578 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should be able to deny pod and configmap creation [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance]","total":280,"completed":103,"skipped":1489,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Lease 
+  lease API should be available [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
+[BeforeEach] [k8s.io] Lease
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:16:40.981: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 16:33:42.933: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename lease-test
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+[It] lease API should be available [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating projection with secret that has name projected-secret-test-map-9c81edc5-8708-4b39-9f70-3eb9c0761774
-STEP: Creating a pod to test consume secrets
-Sep 21 16:16:41.437: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-a571df2d-d436-4e0f-a240-f13c36ed9720" in namespace "projected-5201" to be "success or failure"
-Sep 21 16:16:41.462: INFO: Pod "pod-projected-secrets-a571df2d-d436-4e0f-a240-f13c36ed9720": Phase="Pending", Reason="", readiness=false. Elapsed: 25.113683ms
-Sep 21 16:16:43.489: INFO: Pod "pod-projected-secrets-a571df2d-d436-4e0f-a240-f13c36ed9720": Phase="Pending", Reason="", readiness=false. Elapsed: 2.051247655s
-Sep 21 16:16:45.514: INFO: Pod "pod-projected-secrets-a571df2d-d436-4e0f-a240-f13c36ed9720": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.076685359s
-STEP: Saw pod success
-Sep 21 16:16:45.514: INFO: Pod "pod-projected-secrets-a571df2d-d436-4e0f-a240-f13c36ed9720" satisfied condition "success or failure"
-Sep 21 16:16:45.539: INFO: Trying to get logs from node 10.241.51.147 pod pod-projected-secrets-a571df2d-d436-4e0f-a240-f13c36ed9720 container projected-secret-volume-test: 
-STEP: delete the pod
-Sep 21 16:16:45.724: INFO: Waiting for pod pod-projected-secrets-a571df2d-d436-4e0f-a240-f13c36ed9720 to disappear
-Sep 21 16:16:45.750: INFO: Pod pod-projected-secrets-a571df2d-d436-4e0f-a240-f13c36ed9720 no longer exists
-[AfterEach] [sig-storage] Projected secret
+[AfterEach] [k8s.io] Lease
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:16:45.750: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-5201" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":98,"skipped":1679,"failed":0}
-SSSSSSSSSSS
+Oct 26 16:33:43.306: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "lease-test-1925" for this suite.
+•{"msg":"PASSED [k8s.io] Lease lease API should be available [Conformance]","total":280,"completed":104,"skipped":1545,"failed":0}
+SS
 ------------------------------
-[sig-cli] Kubectl client Kubectl run default 
-  should create an rc or deployment from an image  [Conformance]
+[k8s.io] Container Runtime blackbox test on terminated container 
+  should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [k8s.io] Container Runtime
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:16:45.831: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubectl
+Oct 26 16:33:43.347: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename container-runtime
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[BeforeEach] Kubectl run default
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1596
-[It] should create an rc or deployment from an image  [Conformance]
+[It] should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: running the image docker.io/library/httpd:2.4.38-alpine
-Sep 21 16:16:46.078: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 run e2e-test-httpd-deployment --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-5820'
-Sep 21 16:16:46.272: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-Sep 21 16:16:46.273: INFO: stdout: "deployment.apps/e2e-test-httpd-deployment created\n"
-STEP: verifying the pod controlled by e2e-test-httpd-deployment gets created
-[AfterEach] Kubectl run default
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1602
-Sep 21 16:16:48.338: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete deployment e2e-test-httpd-deployment --namespace=kubectl-5820'
-Sep 21 16:16:48.526: INFO: stderr: ""
-Sep 21 16:16:48.526: INFO: stdout: "deployment.apps \"e2e-test-httpd-deployment\" deleted\n"
-[AfterEach] [sig-cli] Kubectl client
+STEP: create the container
+STEP: wait for the container to reach Succeeded
+STEP: get the container status
+STEP: the container should be terminated
+STEP: the termination message should be set
+Oct 26 16:33:46.644: INFO: Expected: &{DONE} to match Container's Termination Message: DONE --
+STEP: delete the container
+[AfterEach] [k8s.io] Container Runtime
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:16:48.526: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-5820" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run default should create an rc or deployment from an image  [Conformance]","total":280,"completed":99,"skipped":1690,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:33:46.711: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-runtime-6111" for this suite.
+•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]","total":280,"completed":105,"skipped":1547,"failed":0}
+SSSSSSSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] 
+  should include custom resource definition resources in discovery documents [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:16:48.619: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename configmap
+Oct 26 16:33:46.752: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename custom-resource-definition
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+[It] should include custom resource definition resources in discovery documents [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name configmap-test-volume-map-1d0d5bfe-6826-4b2d-bda5-a7602c69b79e
-STEP: Creating a pod to test consume configMaps
-Sep 21 16:16:49.118: INFO: Waiting up to 5m0s for pod "pod-configmaps-1ca09323-11df-44e7-9f5f-59742a6d2257" in namespace "configmap-5514" to be "success or failure"
-Sep 21 16:16:49.142: INFO: Pod "pod-configmaps-1ca09323-11df-44e7-9f5f-59742a6d2257": Phase="Pending", Reason="", readiness=false. Elapsed: 23.653164ms
-Sep 21 16:16:51.177: INFO: Pod "pod-configmaps-1ca09323-11df-44e7-9f5f-59742a6d2257": Phase="Pending", Reason="", readiness=false. Elapsed: 2.059275912s
-Sep 21 16:16:53.203: INFO: Pod "pod-configmaps-1ca09323-11df-44e7-9f5f-59742a6d2257": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.084911421s
-STEP: Saw pod success
-Sep 21 16:16:53.203: INFO: Pod "pod-configmaps-1ca09323-11df-44e7-9f5f-59742a6d2257" satisfied condition "success or failure"
-Sep 21 16:16:53.226: INFO: Trying to get logs from node 10.241.51.147 pod pod-configmaps-1ca09323-11df-44e7-9f5f-59742a6d2257 container configmap-volume-test: 
-STEP: delete the pod
-Sep 21 16:16:53.572: INFO: Waiting for pod pod-configmaps-1ca09323-11df-44e7-9f5f-59742a6d2257 to disappear
-Sep 21 16:16:53.598: INFO: Pod pod-configmaps-1ca09323-11df-44e7-9f5f-59742a6d2257 no longer exists
-[AfterEach] [sig-storage] ConfigMap
+STEP: fetching the /apis discovery document
+STEP: finding the apiextensions.k8s.io API group in the /apis discovery document
+STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis discovery document
+STEP: fetching the /apis/apiextensions.k8s.io discovery document
+STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis/apiextensions.k8s.io discovery document
+STEP: fetching the /apis/apiextensions.k8s.io/v1 discovery document
+STEP: finding customresourcedefinitions resources in the /apis/apiextensions.k8s.io/v1 discovery document
+[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:16:53.598: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-5514" for this suite.
-
-• [SLOW TEST:5.105 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
-  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":100,"skipped":1719,"failed":0}
-SSSSSSSSSSSSSSS
+Oct 26 16:33:46.928: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "custom-resource-definition-9294" for this suite.
+•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance]","total":280,"completed":106,"skipped":1557,"failed":0}
+SSSSSS
 ------------------------------
-[sig-storage] HostPath 
-  should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+[k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class 
+  should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] HostPath
+[BeforeEach] [k8s.io] [sig-node] Pods Extended
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:16:53.724: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename hostpath
+Oct 26 16:33:46.970: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename pods
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] HostPath
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37
-[It] should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] Pods Set QOS Class
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:178
+[It] should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test hostPath mode
-Sep 21 16:16:54.109: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "hostpath-6042" to be "success or failure"
-Sep 21 16:16:54.134: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 24.587965ms
-Sep 21 16:16:56.163: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 2.053604135s
-Sep 21 16:16:58.187: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.0774365s
-STEP: Saw pod success
-Sep 21 16:16:58.187: INFO: Pod "pod-host-path-test" satisfied condition "success or failure"
-Sep 21 16:16:58.216: INFO: Trying to get logs from node 10.241.51.147 pod pod-host-path-test container test-container-1: 
-STEP: delete the pod
-Sep 21 16:16:58.383: INFO: Waiting for pod pod-host-path-test to disappear
-Sep 21 16:16:58.409: INFO: Pod pod-host-path-test no longer exists
-[AfterEach] [sig-storage] HostPath
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying QOS class is set on the pod
+[AfterEach] [k8s.io] [sig-node] Pods Extended
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:16:58.409: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "hostpath-6042" for this suite.
-•{"msg":"PASSED [sig-storage] HostPath should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":101,"skipped":1734,"failed":0}
-SSSSSSSSSSSSSSSSS
+Oct 26 16:33:47.208: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-5115" for this suite.
+•{"msg":"PASSED [k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]","total":280,"completed":107,"skipped":1563,"failed":0}
+SSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl cluster-info 
-  should check if Kubernetes master services is included in cluster-info  [Conformance]
+[k8s.io] Security Context When creating a container with runAsUser 
+  should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [k8s.io] Security Context
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:16:58.470: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubectl
+Oct 26 16:33:47.249: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename security-context-test
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[It] should check if Kubernetes master services is included in cluster-info  [Conformance]
+[BeforeEach] [k8s.io] Security Context
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39
+[It] should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: validating cluster-info
-Sep 21 16:16:58.731: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 cluster-info'
-Sep 21 16:16:58.884: INFO: stderr: ""
-Sep 21 16:16:58.884: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://172.21.0.1:443\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n"
-[AfterEach] [sig-cli] Kubectl client
+Oct 26 16:33:47.456: INFO: Waiting up to 5m0s for pod "busybox-user-65534-aef09855-20fc-4c75-a29d-0279ea0766fc" in namespace "security-context-test-1905" to be "success or failure"
+Oct 26 16:33:47.470: INFO: Pod "busybox-user-65534-aef09855-20fc-4c75-a29d-0279ea0766fc": Phase="Pending", Reason="", readiness=false. Elapsed: 14.79202ms
+Oct 26 16:33:49.486: INFO: Pod "busybox-user-65534-aef09855-20fc-4c75-a29d-0279ea0766fc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030772693s
+Oct 26 16:33:51.503: INFO: Pod "busybox-user-65534-aef09855-20fc-4c75-a29d-0279ea0766fc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.04720339s
+Oct 26 16:33:51.503: INFO: Pod "busybox-user-65534-aef09855-20fc-4c75-a29d-0279ea0766fc" satisfied condition "success or failure"
+[AfterEach] [k8s.io] Security Context
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:16:58.885: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-8125" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes master services is included in cluster-info  [Conformance]","total":280,"completed":102,"skipped":1751,"failed":0}
-SSSSSS
+Oct 26 16:33:51.503: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "security-context-test-1905" for this suite.
+•{"msg":"PASSED [k8s.io] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":108,"skipped":1573,"failed":0}
+SSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+[sig-storage] Secrets 
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
+[BeforeEach] [sig-storage] Secrets
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:16:58.987: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 16:33:51.538: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename secrets
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating projection with secret that has name projected-secret-test-85acc2b3-0af1-468c-afcd-efc9f19df612
-STEP: Creating a pod to test consume secrets
-Sep 21 16:16:59.427: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-8b0c9da3-327b-4ab9-acd6-74cd54f1fed4" in namespace "projected-8865" to be "success or failure"
-Sep 21 16:16:59.455: INFO: Pod "pod-projected-secrets-8b0c9da3-327b-4ab9-acd6-74cd54f1fed4": Phase="Pending", Reason="", readiness=false. Elapsed: 27.755573ms
-Sep 21 16:17:01.480: INFO: Pod "pod-projected-secrets-8b0c9da3-327b-4ab9-acd6-74cd54f1fed4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.052164663s
-STEP: Saw pod success
-Sep 21 16:17:01.480: INFO: Pod "pod-projected-secrets-8b0c9da3-327b-4ab9-acd6-74cd54f1fed4" satisfied condition "success or failure"
-Sep 21 16:17:01.508: INFO: Trying to get logs from node 10.241.51.147 pod pod-projected-secrets-8b0c9da3-327b-4ab9-acd6-74cd54f1fed4 container projected-secret-volume-test: 
-STEP: delete the pod
-Sep 21 16:17:01.904: INFO: Waiting for pod pod-projected-secrets-8b0c9da3-327b-4ab9-acd6-74cd54f1fed4 to disappear
-Sep 21 16:17:01.932: INFO: Pod pod-projected-secrets-8b0c9da3-327b-4ab9-acd6-74cd54f1fed4 no longer exists
-[AfterEach] [sig-storage] Projected secret
+Oct 26 16:33:51.678: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
+STEP: Creating secret with name s-test-opt-del-1671b94a-3a32-4ca8-8126-30b8dd9361f7
+STEP: Creating secret with name s-test-opt-upd-9e3460f9-3b2c-470c-97b0-9e6415000642
+STEP: Creating the pod
+STEP: Deleting secret s-test-opt-del-1671b94a-3a32-4ca8-8126-30b8dd9361f7
+STEP: Updating secret s-test-opt-upd-9e3460f9-3b2c-470c-97b0-9e6415000642
+STEP: Creating secret with name s-test-opt-create-32530452-4778-4add-a406-e71d9ff1510e
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] Secrets
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:17:01.932: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-8865" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":103,"skipped":1757,"failed":0}
-SSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:35:23.892: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-7919" for this suite.
+
+• [SLOW TEST:92.392 seconds]
+[sig-storage] Secrets
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] Variable Expansion 
-  should allow composing env vars into new env vars [NodeConformance] [Conformance]
+{"msg":"PASSED [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":109,"skipped":1588,"failed":0}
+SSSSSSSSSSS
+------------------------------
+[sig-network] DNS 
+  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Variable Expansion
+[BeforeEach] [sig-network] DNS
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:17:02.030: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename var-expansion
+Oct 26 16:35:23.930: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename dns
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should allow composing env vars into new env vars [NodeConformance] [Conformance]
+[It] should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test env composition
-Sep 21 16:17:02.430: INFO: Waiting up to 5m0s for pod "var-expansion-ff899bde-cda3-4313-8407-1fe7f4f1da25" in namespace "var-expansion-3492" to be "success or failure"
-Sep 21 16:17:02.670: INFO: Pod "var-expansion-ff899bde-cda3-4313-8407-1fe7f4f1da25": Phase="Pending", Reason="", readiness=false. Elapsed: 240.153071ms
-Sep 21 16:17:04.697: INFO: Pod "var-expansion-ff899bde-cda3-4313-8407-1fe7f4f1da25": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.267006893s
-STEP: Saw pod success
-Sep 21 16:17:04.697: INFO: Pod "var-expansion-ff899bde-cda3-4313-8407-1fe7f4f1da25" satisfied condition "success or failure"
-Sep 21 16:17:04.724: INFO: Trying to get logs from node 10.241.51.147 pod var-expansion-ff899bde-cda3-4313-8407-1fe7f4f1da25 container dapi-container: 
-STEP: delete the pod
-Sep 21 16:17:05.159: INFO: Waiting for pod var-expansion-ff899bde-cda3-4313-8407-1fe7f4f1da25 to disappear
-Sep 21 16:17:05.460: INFO: Pod var-expansion-ff899bde-cda3-4313-8407-1fe7f4f1da25 no longer exists
-[AfterEach] [k8s.io] Variable Expansion
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-7926.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-7926.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-7926.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
+
+STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-7926.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-7926.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-7926.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+
+STEP: creating a pod to probe /etc/hosts
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Oct 26 16:35:28.370: INFO: DNS probes using dns-7926/dns-test-df6cec2b-fcc6-4bda-9ace-4de1f005bc6c succeeded
+
+STEP: deleting the pod
+[AfterEach] [sig-network] DNS
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:17:05.461: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "var-expansion-3492" for this suite.
-•{"msg":"PASSED [k8s.io] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance]","total":280,"completed":104,"skipped":1778,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:35:28.421: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-7926" for this suite.
+•{"msg":"PASSED [sig-network] DNS should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]","total":280,"completed":110,"skipped":1599,"failed":0}
+SSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir wrapper volumes 
-  should not cause race condition when used for configmaps [Serial] [Conformance]
+[sig-storage] Projected configMap 
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir wrapper volumes
+[BeforeEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:17:05.525: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename emptydir-wrapper
+Oct 26 16:35:28.475: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not cause race condition when used for configmaps [Serial] [Conformance]
+[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating 50 configmaps
-STEP: Creating RC which spawns configmap-volume pods
-Sep 21 16:17:08.862: INFO: Pod name wrapped-volume-race-5dd947f6-f76f-43bf-9ab4-9bab6b5e6903: Found 0 pods out of 5
-Sep 21 16:17:13.922: INFO: Pod name wrapped-volume-race-5dd947f6-f76f-43bf-9ab4-9bab6b5e6903: Found 5 pods out of 5
-STEP: Ensuring each pod is running
-STEP: deleting ReplicationController wrapped-volume-race-5dd947f6-f76f-43bf-9ab4-9bab6b5e6903 in namespace emptydir-wrapper-1989, will wait for the garbage collector to delete the pods
-Sep 21 16:17:14.172: INFO: Deleting ReplicationController wrapped-volume-race-5dd947f6-f76f-43bf-9ab4-9bab6b5e6903 took: 55.820939ms
-Sep 21 16:17:14.272: INFO: Terminating ReplicationController wrapped-volume-race-5dd947f6-f76f-43bf-9ab4-9bab6b5e6903 pods took: 100.231081ms
-STEP: Creating RC which spawns configmap-volume pods
-Sep 21 16:17:22.670: INFO: Pod name wrapped-volume-race-fb6b0e65-4989-4566-b479-d43d3a377495: Found 0 pods out of 5
-Sep 21 16:17:27.720: INFO: Pod name wrapped-volume-race-fb6b0e65-4989-4566-b479-d43d3a377495: Found 5 pods out of 5
-STEP: Ensuring each pod is running
-STEP: deleting ReplicationController wrapped-volume-race-fb6b0e65-4989-4566-b479-d43d3a377495 in namespace emptydir-wrapper-1989, will wait for the garbage collector to delete the pods
-Sep 21 16:17:28.011: INFO: Deleting ReplicationController wrapped-volume-race-fb6b0e65-4989-4566-b479-d43d3a377495 took: 78.699946ms
-Sep 21 16:17:29.011: INFO: Terminating ReplicationController wrapped-volume-race-fb6b0e65-4989-4566-b479-d43d3a377495 pods took: 1.000248677s
-STEP: Creating RC which spawns configmap-volume pods
-Sep 21 16:17:37.609: INFO: Pod name wrapped-volume-race-def74b7f-5180-467a-8f62-bb9a2b646321: Found 0 pods out of 5
-Sep 21 16:17:42.666: INFO: Pod name wrapped-volume-race-def74b7f-5180-467a-8f62-bb9a2b646321: Found 5 pods out of 5
-STEP: Ensuring each pod is running
-STEP: deleting ReplicationController wrapped-volume-race-def74b7f-5180-467a-8f62-bb9a2b646321 in namespace emptydir-wrapper-1989, will wait for the garbage collector to delete the pods
-Sep 21 16:17:43.124: INFO: Deleting ReplicationController wrapped-volume-race-def74b7f-5180-467a-8f62-bb9a2b646321 took: 251.922978ms
-Sep 21 16:17:43.224: INFO: Terminating ReplicationController wrapped-volume-race-def74b7f-5180-467a-8f62-bb9a2b646321 pods took: 100.155902ms
-STEP: Cleaning up the configMaps
-[AfterEach] [sig-storage] EmptyDir wrapper volumes
+Oct 26 16:35:28.631: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
+STEP: Creating configMap with name cm-test-opt-del-3e7fd6a5-8dc3-40e7-a90e-b83122751f5d
+STEP: Creating configMap with name cm-test-opt-upd-43e248bb-a220-4cd0-9496-43d0193af5b7
+STEP: Creating the pod
+STEP: Deleting configmap cm-test-opt-del-3e7fd6a5-8dc3-40e7-a90e-b83122751f5d
+STEP: Updating configmap cm-test-opt-upd-43e248bb-a220-4cd0-9496-43d0193af5b7
+STEP: Creating configMap with name cm-test-opt-create-02e94066-6a99-4bd5-811b-62aa5df1b707
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:18:02.163: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-wrapper-1989" for this suite.
+Oct 26 16:35:37.101: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-2351" for this suite.
 
-• [SLOW TEST:56.724 seconds]
-[sig-storage] EmptyDir wrapper volumes
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
-  should not cause race condition when used for configmaps [Serial] [Conformance]
+• [SLOW TEST:8.666 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance]","total":280,"completed":105,"skipped":1808,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":111,"skipped":1609,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Kubelet when scheduling a busybox Pod with hostAliases 
-  should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
+[sig-cli] Kubectl client Kubectl version 
+  should check is all data is printed  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Kubelet
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:18:02.251: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubelet-test
+Oct 26 16:35:37.144: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubectl
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
-[It] should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
+[It] should check is all data is printed  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [k8s.io] Kubelet
+Oct 26 16:35:37.295: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 version'
+Oct 26 16:35:37.402: INFO: stderr: ""
+Oct 26 16:35:37.402: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"17\", GitVersion:\"v1.17.1\", GitCommit:\"d224476cd0730baca2b6e357d144171ed74192d6\", GitTreeState:\"clean\", BuildDate:\"2020-01-14T21:04:32Z\", GoVersion:\"go1.13.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"17+\", GitVersion:\"v1.17.1+45f8ddb\", GitCommit:\"45f8ddb\", GitTreeState:\"clean\", BuildDate:\"2020-09-26T11:14:22Z\", GoVersion:\"go1.13.4\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n"
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:18:05.006: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubelet-test-6071" for this suite.
-•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox Pod with hostAliases should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":106,"skipped":1883,"failed":0}
+Oct 26 16:35:37.402: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-6209" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl version should check is all data is printed  [Conformance]","total":280,"completed":112,"skipped":1639,"failed":0}
 SSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  should perform rolling updates and roll backs of template modifications [Conformance]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  works for multiple CRDs of same group and version but different kinds [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] StatefulSet
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:18:05.358: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename statefulset
+Oct 26 16:35:37.446: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename crd-publish-openapi
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
-STEP: Creating service test in namespace statefulset-3782
-[It] should perform rolling updates and roll backs of template modifications [Conformance]
+[It] works for multiple CRDs of same group and version but different kinds [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a new StatefulSet
-Sep 21 16:18:05.764: INFO: Found 0 stateful pods, waiting for 3
-Sep 21 16:18:15.792: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
-Sep 21 16:18:15.793: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
-Sep 21 16:18:15.793: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
-Sep 21 16:18:15.881: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-3782 ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Sep 21 16:18:16.253: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Sep 21 16:18:16.253: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Sep 21 16:18:16.253: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-STEP: Updating StatefulSet template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine
-Sep 21 16:18:26.408: INFO: Updating stateful set ss2
-STEP: Creating a new revision
-STEP: Updating Pods in reverse ordinal order
-Sep 21 16:18:36.561: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-3782 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Sep 21 16:18:36.971: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Sep 21 16:18:36.971: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Sep 21 16:18:36.971: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Sep 21 16:18:57.116: INFO: Waiting for StatefulSet statefulset-3782/ss2 to complete update
-Sep 21 16:18:57.116: INFO: Waiting for Pod statefulset-3782/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
-STEP: Rolling back to a previous revision
-Sep 21 16:19:07.166: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-3782 ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Sep 21 16:19:07.586: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Sep 21 16:19:07.586: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Sep 21 16:19:07.586: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Sep 21 16:19:17.788: INFO: Updating stateful set ss2
-STEP: Rolling back update in reverse ordinal order
-Sep 21 16:19:27.931: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-3782 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Sep 21 16:19:28.382: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Sep 21 16:19:28.383: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Sep 21 16:19:28.383: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Sep 21 16:19:38.558: INFO: Waiting for StatefulSet statefulset-3782/ss2 to complete update
-Sep 21 16:19:38.558: INFO: Waiting for Pod statefulset-3782/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
-Sep 21 16:19:38.558: INFO: Waiting for Pod statefulset-3782/ss2-1 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
-Sep 21 16:19:38.558: INFO: Waiting for Pod statefulset-3782/ss2-2 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
-Sep 21 16:19:48.618: INFO: Waiting for StatefulSet statefulset-3782/ss2 to complete update
-Sep 21 16:19:48.618: INFO: Waiting for Pod statefulset-3782/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
-Sep 21 16:19:48.618: INFO: Waiting for Pod statefulset-3782/ss2-1 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
-Sep 21 16:19:58.608: INFO: Waiting for StatefulSet statefulset-3782/ss2 to complete update
-Sep 21 16:19:58.608: INFO: Waiting for Pod statefulset-3782/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
-Sep 21 16:19:58.608: INFO: Waiting for Pod statefulset-3782/ss2-1 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
-Sep 21 16:20:08.623: INFO: Waiting for StatefulSet statefulset-3782/ss2 to complete update
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
-Sep 21 16:20:18.611: INFO: Deleting all statefulset in ns statefulset-3782
-Sep 21 16:20:18.641: INFO: Scaling statefulset ss2 to 0
-Sep 21 16:20:38.769: INFO: Waiting for statefulset status.replicas updated to 0
-Sep 21 16:20:38.802: INFO: Deleting statefulset ss2
-[AfterEach] [sig-apps] StatefulSet
+STEP: CRs in the same group and version but different kinds (two CRDs) show up in OpenAPI documentation
+Oct 26 16:35:37.584: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:35:45.890: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:20:38.914: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-3782" for this suite.
+Oct 26 16:36:17.975: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-8237" for this suite.
 
-• [SLOW TEST:153.621 seconds]
-[sig-apps] StatefulSet
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-    should perform rolling updates and roll backs of template modifications [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:40.563 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  works for multiple CRDs of same group and version but different kinds [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance]","total":280,"completed":107,"skipped":1903,"failed":0}
-SS
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance]","total":280,"completed":113,"skipped":1659,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+[k8s.io] Variable Expansion 
+  should allow substituting values in a container's command [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [k8s.io] Variable Expansion
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:20:38.979: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename emptydir
+Oct 26 16:36:18.010: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename var-expansion
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+[It] should allow substituting values in a container's command [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0666 on tmpfs
-Sep 21 16:20:39.394: INFO: Waiting up to 5m0s for pod "pod-3b10c8b6-a5c7-42e7-b961-153abb04f62c" in namespace "emptydir-3550" to be "success or failure"
-Sep 21 16:20:39.425: INFO: Pod "pod-3b10c8b6-a5c7-42e7-b961-153abb04f62c": Phase="Pending", Reason="", readiness=false. Elapsed: 30.251595ms
-Sep 21 16:20:41.447: INFO: Pod "pod-3b10c8b6-a5c7-42e7-b961-153abb04f62c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.052404169s
+STEP: Creating a pod to test substitution in container's command
+Oct 26 16:36:18.253: INFO: Waiting up to 5m0s for pod "var-expansion-ab3face0-b573-4743-a8f7-d1b0b539fdcd" in namespace "var-expansion-6986" to be "success or failure"
+Oct 26 16:36:18.268: INFO: Pod "var-expansion-ab3face0-b573-4743-a8f7-d1b0b539fdcd": Phase="Pending", Reason="", readiness=false. Elapsed: 15.083037ms
+Oct 26 16:36:20.285: INFO: Pod "var-expansion-ab3face0-b573-4743-a8f7-d1b0b539fdcd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031596749s
+Oct 26 16:36:22.301: INFO: Pod "var-expansion-ab3face0-b573-4743-a8f7-d1b0b539fdcd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.048091147s
 STEP: Saw pod success
-Sep 21 16:20:41.447: INFO: Pod "pod-3b10c8b6-a5c7-42e7-b961-153abb04f62c" satisfied condition "success or failure"
-Sep 21 16:20:41.476: INFO: Trying to get logs from node 10.241.51.147 pod pod-3b10c8b6-a5c7-42e7-b961-153abb04f62c container test-container: 
+Oct 26 16:36:22.301: INFO: Pod "var-expansion-ab3face0-b573-4743-a8f7-d1b0b539fdcd" satisfied condition "success or failure"
+Oct 26 16:36:22.319: INFO: Trying to get logs from node 10.72.119.74 pod var-expansion-ab3face0-b573-4743-a8f7-d1b0b539fdcd container dapi-container: 
 STEP: delete the pod
-Sep 21 16:20:41.695: INFO: Waiting for pod pod-3b10c8b6-a5c7-42e7-b961-153abb04f62c to disappear
-Sep 21 16:20:41.717: INFO: Pod pod-3b10c8b6-a5c7-42e7-b961-153abb04f62c no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+Oct 26 16:36:22.404: INFO: Waiting for pod var-expansion-ab3face0-b573-4743-a8f7-d1b0b539fdcd to disappear
+Oct 26 16:36:22.421: INFO: Pod var-expansion-ab3face0-b573-4743-a8f7-d1b0b539fdcd no longer exists
+[AfterEach] [k8s.io] Variable Expansion
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:20:41.717: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-3550" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":108,"skipped":1905,"failed":0}
-SSSSSSS
+Oct 26 16:36:22.421: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "var-expansion-6986" for this suite.
+•{"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance]","total":280,"completed":114,"skipped":1684,"failed":0}
+SSSSSS
 ------------------------------
-[k8s.io] Pods 
-  should be updated [NodeConformance] [Conformance]
+[sig-storage] Projected downwardAPI 
+  should provide container's cpu limit [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Pods
+[BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:20:41.781: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename pods
+Oct 26 16:36:22.457: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
-[It] should be updated [NodeConformance] [Conformance]
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should provide container's cpu limit [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-STEP: submitting the pod to kubernetes
-STEP: verifying the pod is in kubernetes
-STEP: updating the pod
-Sep 21 16:20:44.848: INFO: Successfully updated pod "pod-update-f9b43ee3-a7cd-4b58-996f-654e5c040031"
-STEP: verifying the updated pod is in kubernetes
-Sep 21 16:20:45.103: INFO: Pod update OK
-[AfterEach] [k8s.io] Pods
+STEP: Creating a pod to test downward API volume plugin
+Oct 26 16:36:22.670: INFO: Waiting up to 5m0s for pod "downwardapi-volume-6fd28854-b28d-4fcf-88f2-857095f1bc5f" in namespace "projected-4711" to be "success or failure"
+Oct 26 16:36:22.686: INFO: Pod "downwardapi-volume-6fd28854-b28d-4fcf-88f2-857095f1bc5f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.936343ms
+Oct 26 16:36:24.704: INFO: Pod "downwardapi-volume-6fd28854-b28d-4fcf-88f2-857095f1bc5f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.03379413s
+STEP: Saw pod success
+Oct 26 16:36:24.704: INFO: Pod "downwardapi-volume-6fd28854-b28d-4fcf-88f2-857095f1bc5f" satisfied condition "success or failure"
+Oct 26 16:36:24.719: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-6fd28854-b28d-4fcf-88f2-857095f1bc5f container client-container: 
+STEP: delete the pod
+Oct 26 16:36:24.804: INFO: Waiting for pod downwardapi-volume-6fd28854-b28d-4fcf-88f2-857095f1bc5f to disappear
+Oct 26 16:36:24.820: INFO: Pod downwardapi-volume-6fd28854-b28d-4fcf-88f2-857095f1bc5f no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:20:45.103: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-2734" for this suite.
-•{"msg":"PASSED [k8s.io] Pods should be updated [NodeConformance] [Conformance]","total":280,"completed":109,"skipped":1912,"failed":0}
-S
+Oct 26 16:36:24.820: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-4711" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance]","total":280,"completed":115,"skipped":1690,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-network] Proxy version v1 
-  should proxy through a service and a pod  [Conformance]
+[sig-api-machinery] ResourceQuota 
+  should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] version v1
+[BeforeEach] [sig-api-machinery] ResourceQuota
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:20:45.150: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename proxy
+Oct 26 16:36:24.857: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename resourcequota
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should proxy through a service and a pod  [Conformance]
+[It] should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: starting an echo server on multiple ports
-STEP: creating replication controller proxy-service-dvsmz in namespace proxy-138
-I0921 16:20:45.531294      24 runners.go:189] Created replication controller with name: proxy-service-dvsmz, namespace: proxy-138, replica count: 1
-I0921 16:20:46.582131      24 runners.go:189] proxy-service-dvsmz Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-I0921 16:20:47.582362      24 runners.go:189] proxy-service-dvsmz Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
-I0921 16:20:48.582707      24 runners.go:189] proxy-service-dvsmz Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
-I0921 16:20:49.582964      24 runners.go:189] proxy-service-dvsmz Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
-I0921 16:20:50.583217      24 runners.go:189] proxy-service-dvsmz Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
-I0921 16:20:51.583508      24 runners.go:189] proxy-service-dvsmz Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
-I0921 16:20:52.583708      24 runners.go:189] proxy-service-dvsmz Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
-I0921 16:20:53.584015      24 runners.go:189] proxy-service-dvsmz Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
-I0921 16:20:54.584263      24 runners.go:189] proxy-service-dvsmz Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
-I0921 16:20:55.584731      24 runners.go:189] proxy-service-dvsmz Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-Sep 21 16:20:55.617: INFO: setup took 10.253768745s, starting test cases
-STEP: running 16 cases, 20 attempts per case, 320 total attempts
-Sep 21 16:20:55.677: INFO: (0) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf/proxy/: test (200; 59.218905ms)
-Sep 21 16:20:55.684: INFO: (0) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:1080/proxy/: testt... (200; 79.990992ms)
-Sep 21 16:20:55.703: INFO: (0) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 85.135743ms)
-Sep 21 16:20:55.703: INFO: (0) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 84.930848ms)
-Sep 21 16:20:55.703: INFO: (0) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname2/proxy/: bar (200; 84.979743ms)
-Sep 21 16:20:55.703: INFO: (0) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:460/proxy/: tls baz (200; 85.57928ms)
-Sep 21 16:20:55.715: INFO: (0) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname1/proxy/: tls baz (200; 97.228565ms)
-Sep 21 16:20:55.715: INFO: (0) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 97.67285ms)
-Sep 21 16:20:55.715: INFO: (0) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname1/proxy/: foo (200; 97.733575ms)
-Sep 21 16:20:55.718: INFO: (0) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname1/proxy/: foo (200; 100.668016ms)
-Sep 21 16:20:55.720: INFO: (0) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:443/proxy/: t... (200; 65.605464ms)
-Sep 21 16:20:55.801: INFO: (1) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 70.764006ms)
-Sep 21 16:20:55.802: INFO: (1) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:462/proxy/: tls qux (200; 72.005817ms)
-Sep 21 16:20:55.803: INFO: (1) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf/proxy/: test (200; 72.497406ms)
-Sep 21 16:20:55.803: INFO: (1) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:443/proxy/: testtestt... (200; 68.128786ms)
-Sep 21 16:20:56.068: INFO: (2) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:460/proxy/: tls baz (200; 69.69408ms)
-Sep 21 16:20:56.068: INFO: (2) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 70.388526ms)
-Sep 21 16:20:56.080: INFO: (2) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:443/proxy/: test (200; 83.066751ms)
-Sep 21 16:20:56.088: INFO: (2) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname2/proxy/: bar (200; 90.056996ms)
-Sep 21 16:20:56.088: INFO: (2) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname1/proxy/: tls baz (200; 90.355283ms)
-Sep 21 16:20:56.092: INFO: (2) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname1/proxy/: foo (200; 93.611293ms)
-Sep 21 16:20:56.115: INFO: (2) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname2/proxy/: tls qux (200; 116.915765ms)
-Sep 21 16:20:56.171: INFO: (3) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 55.656274ms)
-Sep 21 16:20:56.173: INFO: (3) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 57.133165ms)
-Sep 21 16:20:56.178: INFO: (3) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:443/proxy/: test (200; 61.842725ms)
-Sep 21 16:20:56.180: INFO: (3) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:1080/proxy/: t... (200; 64.219373ms)
-Sep 21 16:20:56.180: INFO: (3) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:462/proxy/: tls qux (200; 64.549428ms)
-Sep 21 16:20:56.180: INFO: (3) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 64.610688ms)
-Sep 21 16:20:56.180: INFO: (3) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:1080/proxy/: testtesttest (200; 69.359866ms)
-Sep 21 16:20:56.282: INFO: (4) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 68.615941ms)
-Sep 21 16:20:56.282: INFO: (4) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:443/proxy/: t... (200; 68.793165ms)
-Sep 21 16:20:56.282: INFO: (4) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 69.167259ms)
-Sep 21 16:20:56.283: INFO: (4) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 71.217843ms)
-Sep 21 16:20:56.297: INFO: (4) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname1/proxy/: foo (200; 84.248521ms)
-Sep 21 16:20:56.307: INFO: (4) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname2/proxy/: tls qux (200; 93.803424ms)
-Sep 21 16:20:56.307: INFO: (4) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname1/proxy/: tls baz (200; 94.723996ms)
-Sep 21 16:20:56.307: INFO: (4) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname2/proxy/: bar (200; 94.432403ms)
-Sep 21 16:20:56.307: INFO: (4) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname1/proxy/: foo (200; 94.04787ms)
-Sep 21 16:20:56.346: INFO: (5) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf/proxy/: test (200; 38.905983ms)
-Sep 21 16:20:56.376: INFO: (5) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 68.637932ms)
-Sep 21 16:20:56.381: INFO: (5) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname1/proxy/: tls baz (200; 72.784946ms)
-Sep 21 16:20:56.388: INFO: (5) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 79.820565ms)
-Sep 21 16:20:56.388: INFO: (5) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:1080/proxy/: t... (200; 79.976151ms)
-Sep 21 16:20:56.388: INFO: (5) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 79.944077ms)
-Sep 21 16:20:56.389: INFO: (5) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 80.486877ms)
-Sep 21 16:20:56.393: INFO: (5) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:1080/proxy/: testtestt... (200; 64.548709ms)
-Sep 21 16:20:56.473: INFO: (6) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:462/proxy/: tls qux (200; 64.363277ms)
-Sep 21 16:20:56.474: INFO: (6) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 66.124692ms)
-Sep 21 16:20:56.474: INFO: (6) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname1/proxy/: foo (200; 66.136442ms)
-Sep 21 16:20:56.474: INFO: (6) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf/proxy/: test (200; 66.106825ms)
-Sep 21 16:20:56.492: INFO: (6) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 83.990404ms)
-Sep 21 16:20:56.497: INFO: (6) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname1/proxy/: tls baz (200; 88.844573ms)
-Sep 21 16:20:56.498: INFO: (6) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname2/proxy/: bar (200; 89.495457ms)
-Sep 21 16:20:56.498: INFO: (6) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname1/proxy/: foo (200; 89.86006ms)
-Sep 21 16:20:56.551: INFO: (7) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 52.222828ms)
-Sep 21 16:20:56.558: INFO: (7) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:1080/proxy/: testt... (200; 59.193543ms)
-Sep 21 16:20:56.559: INFO: (7) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 59.911486ms)
-Sep 21 16:20:56.560: INFO: (7) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf/proxy/: test (200; 62.08728ms)
-Sep 21 16:20:56.567: INFO: (7) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:462/proxy/: tls qux (200; 67.826054ms)
-Sep 21 16:20:56.567: INFO: (7) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:443/proxy/: testt... (200; 65.138344ms)
-Sep 21 16:20:56.671: INFO: (8) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf/proxy/: test (200; 69.451189ms)
-Sep 21 16:20:56.671: INFO: (8) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname2/proxy/: tls qux (200; 69.048883ms)
-Sep 21 16:20:56.680: INFO: (8) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname1/proxy/: tls baz (200; 78.282831ms)
-Sep 21 16:20:56.681: INFO: (8) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname1/proxy/: foo (200; 79.357219ms)
-Sep 21 16:20:56.685: INFO: (8) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 82.990463ms)
-Sep 21 16:20:56.688: INFO: (8) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname1/proxy/: foo (200; 85.985932ms)
-Sep 21 16:20:56.688: INFO: (8) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname2/proxy/: bar (200; 86.533395ms)
-Sep 21 16:20:56.727: INFO: (9) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:1080/proxy/: testt... (200; 57.854819ms)
-Sep 21 16:20:56.750: INFO: (9) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:460/proxy/: tls baz (200; 61.550952ms)
-Sep 21 16:20:56.750: INFO: (9) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 60.638511ms)
-Sep 21 16:20:56.756: INFO: (9) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 66.898755ms)
-Sep 21 16:20:56.756: INFO: (9) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:462/proxy/: tls qux (200; 66.772828ms)
-Sep 21 16:20:56.757: INFO: (9) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf/proxy/: test (200; 68.357377ms)
-Sep 21 16:20:56.758: INFO: (9) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 68.965379ms)
-Sep 21 16:20:56.763: INFO: (9) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname1/proxy/: tls baz (200; 73.983095ms)
-Sep 21 16:20:56.767: INFO: (9) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname1/proxy/: foo (200; 77.10675ms)
-Sep 21 16:20:56.775: INFO: (9) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname2/proxy/: bar (200; 85.354332ms)
-Sep 21 16:20:56.775: INFO: (9) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname1/proxy/: foo (200; 85.88687ms)
-Sep 21 16:20:56.775: INFO: (9) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname2/proxy/: tls qux (200; 85.46847ms)
-Sep 21 16:20:56.821: INFO: (10) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:460/proxy/: tls baz (200; 45.677542ms)
-Sep 21 16:20:56.838: INFO: (10) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 62.287598ms)
-Sep 21 16:20:56.838: INFO: (10) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:462/proxy/: tls qux (200; 62.805623ms)
-Sep 21 16:20:56.858: INFO: (10) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:443/proxy/: t... (200; 81.420796ms)
-Sep 21 16:20:56.858: INFO: (10) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf/proxy/: test (200; 81.415995ms)
-Sep 21 16:20:56.858: INFO: (10) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 81.981614ms)
-Sep 21 16:20:56.858: INFO: (10) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:1080/proxy/: testtest (200; 96.323676ms)
-Sep 21 16:20:56.978: INFO: (11) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:1080/proxy/: testt... (200; 110.882843ms)
-Sep 21 16:20:56.993: INFO: (11) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 111.450071ms)
-Sep 21 16:20:56.993: INFO: (11) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:443/proxy/: testtest (200; 67.980588ms)
-Sep 21 16:20:57.265: INFO: (12) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:1080/proxy/: t... (200; 68.712708ms)
-Sep 21 16:20:57.267: INFO: (12) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname1/proxy/: tls baz (200; 71.20314ms)
-Sep 21 16:20:57.267: INFO: (12) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname1/proxy/: foo (200; 70.798427ms)
-Sep 21 16:20:57.282: INFO: (12) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname2/proxy/: bar (200; 86.559751ms)
-Sep 21 16:20:57.287: INFO: (12) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname1/proxy/: foo (200; 90.411577ms)
-Sep 21 16:20:57.287: INFO: (12) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname2/proxy/: tls qux (200; 91.006663ms)
-Sep 21 16:20:57.287: INFO: (12) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 90.350526ms)
-Sep 21 16:20:57.347: INFO: (13) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:443/proxy/: testtest (200; 69.479159ms)
-Sep 21 16:20:57.357: INFO: (13) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 69.378806ms)
-Sep 21 16:20:57.357: INFO: (13) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 69.411433ms)
-Sep 21 16:20:57.364: INFO: (13) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:1080/proxy/: t... (200; 75.600716ms)
-Sep 21 16:20:57.365: INFO: (13) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:462/proxy/: tls qux (200; 77.631783ms)
-Sep 21 16:20:57.365: INFO: (13) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 77.80983ms)
-Sep 21 16:20:57.365: INFO: (13) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname2/proxy/: tls qux (200; 78.478367ms)
-Sep 21 16:20:57.365: INFO: (13) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 77.336713ms)
-Sep 21 16:20:57.396: INFO: (13) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname1/proxy/: tls baz (200; 108.764386ms)
-Sep 21 16:20:57.396: INFO: (13) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname1/proxy/: foo (200; 108.470848ms)
-Sep 21 16:20:57.396: INFO: (13) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname1/proxy/: foo (200; 108.105682ms)
-Sep 21 16:20:57.407: INFO: (13) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 119.433379ms)
-Sep 21 16:20:57.407: INFO: (13) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname2/proxy/: bar (200; 118.689414ms)
-Sep 21 16:20:57.456: INFO: (14) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:443/proxy/: testt... (200; 66.972148ms)
-Sep 21 16:20:57.475: INFO: (14) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf/proxy/: test (200; 67.130926ms)
-Sep 21 16:20:57.475: INFO: (14) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 67.212624ms)
-Sep 21 16:20:57.478: INFO: (14) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname2/proxy/: tls qux (200; 71.311438ms)
-Sep 21 16:20:57.480: INFO: (14) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname1/proxy/: tls baz (200; 72.103652ms)
-Sep 21 16:20:57.480: INFO: (14) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 72.978503ms)
-Sep 21 16:20:57.493: INFO: (14) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname2/proxy/: bar (200; 85.600613ms)
-Sep 21 16:20:57.493: INFO: (14) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname1/proxy/: foo (200; 85.693029ms)
-Sep 21 16:20:57.493: INFO: (14) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname1/proxy/: foo (200; 85.576079ms)
-Sep 21 16:20:57.549: INFO: (15) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 55.48413ms)
-Sep 21 16:20:57.549: INFO: (15) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:460/proxy/: tls baz (200; 55.709702ms)
-Sep 21 16:20:57.550: INFO: (15) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf/proxy/: test (200; 56.775226ms)
-Sep 21 16:20:57.566: INFO: (15) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:1080/proxy/: t... (200; 72.59623ms)
-Sep 21 16:20:57.567: INFO: (15) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 72.408562ms)
-Sep 21 16:20:57.568: INFO: (15) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 73.481489ms)
-Sep 21 16:20:57.570: INFO: (15) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:443/proxy/: testtest (200; 106.595301ms)
-Sep 21 16:20:57.703: INFO: (16) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:462/proxy/: tls qux (200; 108.506053ms)
-Sep 21 16:20:57.703: INFO: (16) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 108.501886ms)
-Sep 21 16:20:57.703: INFO: (16) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:1080/proxy/: testt... (200; 108.659938ms)
-Sep 21 16:20:57.703: INFO: (16) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname2/proxy/: tls qux (200; 108.738272ms)
-Sep 21 16:20:57.708: INFO: (16) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname1/proxy/: tls baz (200; 113.181985ms)
-Sep 21 16:20:57.718: INFO: (16) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 122.88222ms)
-Sep 21 16:20:57.720: INFO: (16) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname1/proxy/: foo (200; 126.442588ms)
-Sep 21 16:20:57.720: INFO: (16) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname1/proxy/: foo (200; 125.920923ms)
-Sep 21 16:20:57.774: INFO: (17) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 53.403676ms)
-Sep 21 16:20:57.782: INFO: (17) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:462/proxy/: tls qux (200; 60.910038ms)
-Sep 21 16:20:57.782: INFO: (17) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 61.376607ms)
-Sep 21 16:20:57.784: INFO: (17) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:1080/proxy/: t... (200; 62.747185ms)
-Sep 21 16:20:57.784: INFO: (17) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:443/proxy/: test (200; 65.244137ms)
-Sep 21 16:20:57.786: INFO: (17) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 64.908078ms)
-Sep 21 16:20:57.788: INFO: (17) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:1080/proxy/: testtest (200; 62.734824ms)
-Sep 21 16:20:57.882: INFO: (18) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:460/proxy/: tls baz (200; 62.709296ms)
-Sep 21 16:20:57.882: INFO: (18) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 62.413146ms)
-Sep 21 16:20:57.886: INFO: (18) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 66.851433ms)
-Sep 21 16:20:57.887: INFO: (18) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 67.847973ms)
-Sep 21 16:20:57.892: INFO: (18) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:462/proxy/: tls qux (200; 73.138818ms)
-Sep 21 16:20:57.895: INFO: (18) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:1080/proxy/: t... (200; 75.50366ms)
-Sep 21 16:20:57.895: INFO: (18) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 75.572322ms)
-Sep 21 16:20:57.895: INFO: (18) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:1080/proxy/: testt... (200; 67.812165ms)
-Sep 21 16:20:58.014: INFO: (19) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:462/proxy/: tls qux (200; 68.891307ms)
-Sep 21 16:20:58.015: INFO: (19) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:160/proxy/: foo (200; 68.884573ms)
-Sep 21 16:20:58.015: INFO: (19) /api/v1/namespaces/proxy-138/pods/http:proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 69.694023ms)
-Sep 21 16:20:58.027: INFO: (19) /api/v1/namespaces/proxy-138/pods/https:proxy-service-dvsmz-wzkjf:460/proxy/: tls baz (200; 82.079466ms)
-Sep 21 16:20:58.027: INFO: (19) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname1/proxy/: tls baz (200; 82.109995ms)
-Sep 21 16:20:58.027: INFO: (19) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:162/proxy/: bar (200; 82.034218ms)
-Sep 21 16:20:58.037: INFO: (19) /api/v1/namespaces/proxy-138/services/https:proxy-service-dvsmz:tlsportname2/proxy/: tls qux (200; 90.78106ms)
-Sep 21 16:20:58.055: INFO: (19) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname2/proxy/: bar (200; 108.677202ms)
-Sep 21 16:20:58.058: INFO: (19) /api/v1/namespaces/proxy-138/services/proxy-service-dvsmz:portname1/proxy/: foo (200; 112.872225ms)
-Sep 21 16:20:58.059: INFO: (19) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname2/proxy/: bar (200; 113.755946ms)
-Sep 21 16:20:58.059: INFO: (19) /api/v1/namespaces/proxy-138/services/http:proxy-service-dvsmz:portname1/proxy/: foo (200; 113.660451ms)
-Sep 21 16:20:58.240: INFO: (19) /api/v1/namespaces/proxy-138/pods/proxy-service-dvsmz-wzkjf:1080/proxy/: testtest (200; 295.288401ms)
-STEP: deleting ReplicationController proxy-service-dvsmz in namespace proxy-138, will wait for the garbage collector to delete the pods
-Sep 21 16:20:58.368: INFO: Deleting ReplicationController proxy-service-dvsmz took: 56.393388ms
-Sep 21 16:20:58.468: INFO: Terminating ReplicationController proxy-service-dvsmz pods took: 100.292469ms
-[AfterEach] version v1
+STEP: Counting existing ResourceQuota
+STEP: Creating a ResourceQuota
+STEP: Ensuring resource quota status is calculated
+[AfterEach] [sig-api-machinery] ResourceQuota
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:21:07.369: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "proxy-138" for this suite.
+Oct 26 16:36:32.070: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-2294" for this suite.
 
-• [SLOW TEST:22.291 seconds]
-[sig-network] Proxy
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  version v1
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:57
-    should proxy through a service and a pod  [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:7.253 seconds]
+[sig-api-machinery] ResourceQuota
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Proxy version v1 should proxy through a service and a pod  [Conformance]","total":280,"completed":110,"skipped":1913,"failed":0}
-SSSS
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]","total":280,"completed":116,"skipped":1721,"failed":0}
+SSSSSSSSSSSSSSSS
 ------------------------------
-[sig-network] Networking Granular Checks: Pods 
-  should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  patching/updating a mutating webhook should work [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Networking
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:21:07.441: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename pod-network-test
+Oct 26 16:36:32.111: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Oct 26 16:36:33.030: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 16:36:35.089: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326993, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326993, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326993, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739326993, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Oct 26 16:36:38.148: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] patching/updating a mutating webhook should work [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Performing setup for networking test in namespace pod-network-test-6919
-STEP: creating a selector
-STEP: Creating the service pods in kubernetes
-Sep 21 16:21:07.810: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
-STEP: Creating test pods
-Sep 21 16:21:31.196: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.237.101:8080/dial?request=hostname&protocol=udp&host=172.30.237.100&port=8081&tries=1'] Namespace:pod-network-test-6919 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:21:31.196: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:21:31.412: INFO: Waiting for responses: map[]
-Sep 21 16:21:31.445: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.237.101:8080/dial?request=hostname&protocol=udp&host=172.30.139.63&port=8081&tries=1'] Namespace:pod-network-test-6919 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:21:31.445: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:21:31.733: INFO: Waiting for responses: map[]
-Sep 21 16:21:31.757: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.237.101:8080/dial?request=hostname&protocol=udp&host=172.30.81.26&port=8081&tries=1'] Namespace:pod-network-test-6919 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:21:31.757: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:21:32.108: INFO: Waiting for responses: map[]
-[AfterEach] [sig-network] Networking
+STEP: Creating a mutating webhook configuration
+STEP: Updating a mutating webhook configuration's rules to not include the create operation
+STEP: Creating a configMap that should not be mutated
+STEP: Patching a mutating webhook configuration's rules to include the create operation
+STEP: Creating a configMap that should be mutated
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:21:32.108: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pod-network-test-6919" for this suite.
+Oct 26 16:36:38.377: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-277" for this suite.
+STEP: Destroying namespace "webhook-277-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:24.741 seconds]
-[sig-network] Networking
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
-  Granular Checks: Pods
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
-    should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:6.525 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  patching/updating a mutating webhook should work [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":111,"skipped":1917,"failed":0}
-[sig-storage] Secrets 
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance]","total":280,"completed":117,"skipped":1737,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
+[BeforeEach] [sig-storage] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:21:32.182: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename secrets
+Oct 26 16:36:38.640: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename configmap
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:21:32.539: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
-STEP: Creating secret with name s-test-opt-del-c6cd4903-856a-4b4e-bc0d-80571fc90faf
-STEP: Creating secret with name s-test-opt-upd-cb05ec7b-ee36-4627-ba9b-b0db723a083f
-STEP: Creating the pod
-STEP: Deleting secret s-test-opt-del-c6cd4903-856a-4b4e-bc0d-80571fc90faf
-STEP: Updating secret s-test-opt-upd-cb05ec7b-ee36-4627-ba9b-b0db723a083f
-STEP: Creating secret with name s-test-opt-create-8087064c-d75b-4c57-a6e5-01ea27ae1d35
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] Secrets
+STEP: Creating configMap with name configmap-test-volume-map-86409a35-b8bc-4577-8004-f5671a481870
+STEP: Creating a pod to test consume configMaps
+Oct 26 16:36:38.896: INFO: Waiting up to 5m0s for pod "pod-configmaps-9aa44636-82a3-43d2-aa16-ad53797e0eba" in namespace "configmap-834" to be "success or failure"
+Oct 26 16:36:38.916: INFO: Pod "pod-configmaps-9aa44636-82a3-43d2-aa16-ad53797e0eba": Phase="Pending", Reason="", readiness=false. Elapsed: 19.773276ms
+Oct 26 16:36:40.933: INFO: Pod "pod-configmaps-9aa44636-82a3-43d2-aa16-ad53797e0eba": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.036890008s
+STEP: Saw pod success
+Oct 26 16:36:40.933: INFO: Pod "pod-configmaps-9aa44636-82a3-43d2-aa16-ad53797e0eba" satisfied condition "success or failure"
+Oct 26 16:36:40.947: INFO: Trying to get logs from node 10.72.119.74 pod pod-configmaps-9aa44636-82a3-43d2-aa16-ad53797e0eba container configmap-volume-test: 
+STEP: delete the pod
+Oct 26 16:36:41.042: INFO: Waiting for pod pod-configmaps-9aa44636-82a3-43d2-aa16-ad53797e0eba to disappear
+Oct 26 16:36:41.057: INFO: Pod pod-configmaps-9aa44636-82a3-43d2-aa16-ad53797e0eba no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 16:36:41.057: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-834" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":118,"skipped":1772,"failed":0}
+S
+------------------------------
+[sig-api-machinery] ResourceQuota 
+  should create a ResourceQuota and capture the life of a secret. [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] ResourceQuota
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 16:36:41.098: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename resourcequota
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should create a ResourceQuota and capture the life of a secret. [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Discovering how many secrets are in namespace by default
+STEP: Counting existing ResourceQuota
+STEP: Creating a ResourceQuota
+STEP: Ensuring resource quota status is calculated
+STEP: Creating a Secret
+STEP: Ensuring resource quota status captures secret creation
+STEP: Deleting a secret
+STEP: Ensuring resource quota status released usage
+[AfterEach] [sig-api-machinery] ResourceQuota
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:22:48.368: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-1429" for this suite.
+Oct 26 16:36:58.511: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-6420" for this suite.
 
-• [SLOW TEST:76.257 seconds]
-[sig-storage] Secrets
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
+• [SLOW TEST:17.449 seconds]
+[sig-api-machinery] ResourceQuota
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should create a ResourceQuota and capture the life of a secret. [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":112,"skipped":1917,"failed":0}
-SSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance]","total":280,"completed":119,"skipped":1773,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl expose 
-  should create services for rc  [Conformance]
+[k8s.io] Pods 
+  should support remote command execution over websockets [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [k8s.io] Pods
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:22:48.440: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubectl
+Oct 26 16:36:58.547: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename pods
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[It] should create services for rc  [Conformance]
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
+[It] should support remote command execution over websockets [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating Agnhost RC
-Sep 21 16:22:48.738: INFO: namespace kubectl-9093
-Sep 21 16:22:48.738: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-9093'
-Sep 21 16:22:49.444: INFO: stderr: ""
-Sep 21 16:22:49.444: INFO: stdout: "replicationcontroller/agnhost-master created\n"
-STEP: Waiting for Agnhost master to start.
-Sep 21 16:22:50.467: INFO: Selector matched 1 pods for map[app:agnhost]
-Sep 21 16:22:50.467: INFO: Found 0 / 1
-Sep 21 16:22:51.496: INFO: Selector matched 1 pods for map[app:agnhost]
-Sep 21 16:22:51.496: INFO: Found 0 / 1
-Sep 21 16:22:52.470: INFO: Selector matched 1 pods for map[app:agnhost]
-Sep 21 16:22:52.470: INFO: Found 1 / 1
-Sep 21 16:22:52.470: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
-Sep 21 16:22:52.500: INFO: Selector matched 1 pods for map[app:agnhost]
-Sep 21 16:22:52.500: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
-Sep 21 16:22:52.500: INFO: wait on agnhost-master startup in kubectl-9093 
-Sep 21 16:22:52.500: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 logs agnhost-master-bnx4m agnhost-master --namespace=kubectl-9093'
-Sep 21 16:22:52.709: INFO: stderr: ""
-Sep 21 16:22:52.709: INFO: stdout: "Paused\n"
-STEP: exposing RC
-Sep 21 16:22:52.709: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 expose rc agnhost-master --name=rm2 --port=1234 --target-port=6379 --namespace=kubectl-9093'
-Sep 21 16:22:52.942: INFO: stderr: ""
-Sep 21 16:22:52.942: INFO: stdout: "service/rm2 exposed\n"
-Sep 21 16:22:52.963: INFO: Service rm2 in namespace kubectl-9093 found.
-STEP: exposing service
-Sep 21 16:22:55.036: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=kubectl-9093'
-Sep 21 16:22:55.301: INFO: stderr: ""
-Sep 21 16:22:55.301: INFO: stdout: "service/rm3 exposed\n"
-Sep 21 16:22:55.358: INFO: Service rm3 in namespace kubectl-9093 found.
-[AfterEach] [sig-cli] Kubectl client
+Oct 26 16:36:58.716: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+[AfterEach] [k8s.io] Pods
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:22:57.482: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-9093" for this suite.
-
-• [SLOW TEST:9.164 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Kubectl expose
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1295
-    should create services for rc  [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Oct 26 16:37:01.041: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-561" for this suite.
+•{"msg":"PASSED [k8s.io] Pods should support remote command execution over websockets [NodeConformance] [Conformance]","total":280,"completed":120,"skipped":1801,"failed":0}
+SS
 ------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Kubectl expose should create services for rc  [Conformance]","total":280,"completed":113,"skipped":1929,"failed":0}
-[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition 
-  getting/updating/patching custom resource definition status sub-resource works  [Conformance]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  works for multiple CRDs of different groups [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:22:57.604: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename custom-resource-definition
+Oct 26 16:37:01.077: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename crd-publish-openapi
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] getting/updating/patching custom resource definition status sub-resource works  [Conformance]
+[It] works for multiple CRDs of different groups [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:22:57.888: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+STEP: CRs in different groups (two CRDs) show up in OpenAPI documentation
+Oct 26 16:37:01.207: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:37:09.288: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:22:58.209: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "custom-resource-definition-8624" for this suite.
-•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works  [Conformance]","total":280,"completed":114,"skipped":1929,"failed":0}
-SSSSSSSSS
+Oct 26 16:37:40.745: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-7812" for this suite.
+
+• [SLOW TEST:39.701 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  works for multiple CRDs of different groups [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-api-machinery] ResourceQuota 
-  should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance]","total":280,"completed":121,"skipped":1803,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] ReplicaSet 
+  should adopt matching pods on creation and release no longer matching pods [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
+[BeforeEach] [sig-apps] ReplicaSet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:22:58.371: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename resourcequota
+Oct 26 16:37:40.779: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename replicaset
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]
+[It] should adopt matching pods on creation and release no longer matching pods [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Counting existing ResourceQuota
-STEP: Creating a ResourceQuota
-STEP: Ensuring resource quota status is calculated
-[AfterEach] [sig-api-machinery] ResourceQuota
+STEP: Given a Pod with a 'name' label pod-adoption-release is created
+STEP: When a replicaset with a matching selector is created
+STEP: Then the orphan pod is adopted
+STEP: When the matched label of one of its pods change
+Oct 26 16:37:44.069: INFO: Pod name pod-adoption-release: Found 1 pods out of 1
+STEP: Then the pod is released
+[AfterEach] [sig-apps] ReplicaSet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:23:05.788: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-8062" for this suite.
+Oct 26 16:37:45.153: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replicaset-4846" for this suite.
+•{"msg":"PASSED [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance]","total":280,"completed":122,"skipped":1863,"failed":0}
 
-• [SLOW TEST:7.535 seconds]
-[sig-api-machinery] ResourceQuota
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]","total":280,"completed":115,"skipped":1938,"failed":0}
-SSSSSSSS
-------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+[sig-network] DNS 
+  should provide DNS for pods for Subdomain [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
+[BeforeEach] [sig-network] DNS
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:23:05.907: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 16:37:45.200: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename dns
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+[It] should provide DNS for pods for Subdomain [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:23:06.298: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2c9a6795-b55a-4319-9342-610353a1fe57" in namespace "projected-3431" to be "success or failure"
-Sep 21 16:23:06.329: INFO: Pod "downwardapi-volume-2c9a6795-b55a-4319-9342-610353a1fe57": Phase="Pending", Reason="", readiness=false. Elapsed: 30.764003ms
-Sep 21 16:23:08.357: INFO: Pod "downwardapi-volume-2c9a6795-b55a-4319-9342-610353a1fe57": Phase="Pending", Reason="", readiness=false. Elapsed: 2.058092706s
-Sep 21 16:23:12.636: INFO: Pod "downwardapi-volume-2c9a6795-b55a-4319-9342-610353a1fe57": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.337447211s
-STEP: Saw pod success
-Sep 21 16:23:12.636: INFO: Pod "downwardapi-volume-2c9a6795-b55a-4319-9342-610353a1fe57" satisfied condition "success or failure"
-Sep 21 16:23:14.631: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-2c9a6795-b55a-4319-9342-610353a1fe57 container client-container: 
-STEP: delete the pod
-Sep 21 16:23:18.740: INFO: Waiting for pod downwardapi-volume-2c9a6795-b55a-4319-9342-610353a1fe57 to disappear
-Sep 21 16:23:18.765: INFO: Pod downwardapi-volume-2c9a6795-b55a-4319-9342-610353a1fe57 no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
+STEP: Creating a test headless service
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-3577.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-querier-2.dns-test-service-2.dns-3577.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-3577.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-querier-2.dns-test-service-2.dns-3577.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-3577.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service-2.dns-3577.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-3577.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service-2.dns-3577.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3577.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
+
+STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-3577.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-querier-2.dns-test-service-2.dns-3577.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-3577.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-querier-2.dns-test-service-2.dns-3577.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-3577.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service-2.dns-3577.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-3577.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service-2.dns-3577.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3577.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+
+STEP: creating a pod to probe DNS
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Oct 26 16:37:49.564: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-3577.svc.cluster.local from pod dns-3577/dns-test-5158e651-e54f-47b6-a375-f6a447ec7049: the server could not find the requested resource (get pods dns-test-5158e651-e54f-47b6-a375-f6a447ec7049)
+Oct 26 16:37:49.595: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-3577.svc.cluster.local from pod dns-3577/dns-test-5158e651-e54f-47b6-a375-f6a447ec7049: the server could not find the requested resource (get pods dns-test-5158e651-e54f-47b6-a375-f6a447ec7049)
+Oct 26 16:37:49.719: INFO: Unable to read jessie_udp@dns-test-service-2.dns-3577.svc.cluster.local from pod dns-3577/dns-test-5158e651-e54f-47b6-a375-f6a447ec7049: the server could not find the requested resource (get pods dns-test-5158e651-e54f-47b6-a375-f6a447ec7049)
+Oct 26 16:37:49.742: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-3577.svc.cluster.local from pod dns-3577/dns-test-5158e651-e54f-47b6-a375-f6a447ec7049: the server could not find the requested resource (get pods dns-test-5158e651-e54f-47b6-a375-f6a447ec7049)
+Oct 26 16:37:49.790: INFO: Lookups using dns-3577/dns-test-5158e651-e54f-47b6-a375-f6a447ec7049 failed for: [wheezy_udp@dns-test-service-2.dns-3577.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-3577.svc.cluster.local jessie_udp@dns-test-service-2.dns-3577.svc.cluster.local jessie_tcp@dns-test-service-2.dns-3577.svc.cluster.local]
+
+Oct 26 16:37:55.099: INFO: DNS probes using dns-3577/dns-test-5158e651-e54f-47b6-a375-f6a447ec7049 succeeded
+
+STEP: deleting the pod
+STEP: deleting the test headless service
+[AfterEach] [sig-network] DNS
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:23:18.765: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-3431" for this suite.
+Oct 26 16:37:55.223: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-3577" for this suite.
 
-• [SLOW TEST:12.957 seconds]
-[sig-storage] Projected downwardAPI
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:34
-  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+• [SLOW TEST:10.057 seconds]
+[sig-network] DNS
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should provide DNS for pods for Subdomain [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":116,"skipped":1946,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Subdomain [Conformance]","total":280,"completed":123,"skipped":1863,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-node] Downward API 
-  should provide pod UID as env vars [NodeConformance] [Conformance]
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] Downward API
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:23:18.867: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
+Oct 26 16:37:55.258: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide pod UID as env vars [NodeConformance] [Conformance]
+[It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward api env vars
-Sep 21 16:23:19.307: INFO: Waiting up to 5m0s for pod "downward-api-5fad414a-1f3b-402b-97db-555cdb6a691d" in namespace "downward-api-1337" to be "success or failure"
-Sep 21 16:23:19.339: INFO: Pod "downward-api-5fad414a-1f3b-402b-97db-555cdb6a691d": Phase="Pending", Reason="", readiness=false. Elapsed: 32.259649ms
-Sep 21 16:23:21.398: INFO: Pod "downward-api-5fad414a-1f3b-402b-97db-555cdb6a691d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.09092026s
-Sep 21 16:23:23.422: INFO: Pod "downward-api-5fad414a-1f3b-402b-97db-555cdb6a691d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.115092302s
+STEP: Creating a pod to test emptydir 0777 on node default medium
+Oct 26 16:37:55.457: INFO: Waiting up to 5m0s for pod "pod-66bbd4f0-c56d-4e5c-be74-29cf82b8cd82" in namespace "emptydir-9763" to be "success or failure"
+Oct 26 16:37:55.473: INFO: Pod "pod-66bbd4f0-c56d-4e5c-be74-29cf82b8cd82": Phase="Pending", Reason="", readiness=false. Elapsed: 15.236948ms
+Oct 26 16:37:57.490: INFO: Pod "pod-66bbd4f0-c56d-4e5c-be74-29cf82b8cd82": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.032461344s
 STEP: Saw pod success
-Sep 21 16:23:23.422: INFO: Pod "downward-api-5fad414a-1f3b-402b-97db-555cdb6a691d" satisfied condition "success or failure"
-Sep 21 16:23:23.445: INFO: Trying to get logs from node 10.241.51.147 pod downward-api-5fad414a-1f3b-402b-97db-555cdb6a691d container dapi-container: 
+Oct 26 16:37:57.490: INFO: Pod "pod-66bbd4f0-c56d-4e5c-be74-29cf82b8cd82" satisfied condition "success or failure"
+Oct 26 16:37:57.509: INFO: Trying to get logs from node 10.72.119.74 pod pod-66bbd4f0-c56d-4e5c-be74-29cf82b8cd82 container test-container: 
 STEP: delete the pod
-Sep 21 16:23:23.598: INFO: Waiting for pod downward-api-5fad414a-1f3b-402b-97db-555cdb6a691d to disappear
-Sep 21 16:23:23.624: INFO: Pod downward-api-5fad414a-1f3b-402b-97db-555cdb6a691d no longer exists
-[AfterEach] [sig-node] Downward API
+Oct 26 16:37:57.594: INFO: Waiting for pod pod-66bbd4f0-c56d-4e5c-be74-29cf82b8cd82 to disappear
+Oct 26 16:37:57.610: INFO: Pod pod-66bbd4f0-c56d-4e5c-be74-29cf82b8cd82 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:23:23.624: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-1337" for this suite.
-•{"msg":"PASSED [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance]","total":280,"completed":117,"skipped":1987,"failed":0}
+Oct 26 16:37:57.610: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-9763" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":124,"skipped":1890,"failed":0}
 SSSS
 ------------------------------
-[k8s.io] [sig-node] Pods Extended [k8s.io] Delete Grace Period 
-  should be submitted and removed [Conformance]
+[sig-cli] Kubectl client Kubectl run pod 
+  should create a pod from an image when restart is Never  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] [sig-node] Pods Extended
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:23:23.706: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename pods
+Oct 26 16:37:57.655: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubectl
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Delete Grace Period
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:46
-[It] should be submitted and removed [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
+[BeforeEach] Kubectl run pod
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1861
+[It] should create a pod from an image when restart is Never  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-STEP: setting up selector
-STEP: submitting the pod to kubernetes
-STEP: verifying the pod is in kubernetes
-Sep 21 16:23:28.470: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-468631458 proxy -p 0'
-STEP: deleting the pod gracefully
-STEP: verifying the kubelet observed the termination notice
-Sep 21 16:23:38.915: INFO: no pod exists with the name we were looking for, assuming the termination request was observed and completed
-[AfterEach] [k8s.io] [sig-node] Pods Extended
+STEP: running the image docker.io/library/httpd:2.4.38-alpine
+Oct 26 16:37:57.808: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 run e2e-test-httpd-pod --restart=Never --generator=run-pod/v1 --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-3828'
+Oct 26 16:37:58.009: INFO: stderr: ""
+Oct 26 16:37:58.009: INFO: stdout: "pod/e2e-test-httpd-pod created\n"
+STEP: verifying the pod e2e-test-httpd-pod was created
+[AfterEach] Kubectl run pod
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1866
+Oct 26 16:37:58.024: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete pods e2e-test-httpd-pod --namespace=kubectl-3828'
+Oct 26 16:38:06.780: INFO: stderr: ""
+Oct 26 16:38:06.780: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:23:38.940: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-2239" for this suite.
+Oct 26 16:38:06.780: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-3828" for this suite.
 
-• [SLOW TEST:15.496 seconds]
-[k8s.io] [sig-node] Pods Extended
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  [k8s.io] Delete Grace Period
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-    should be submitted and removed [Conformance]
+• [SLOW TEST:9.166 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Kubectl run pod
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1857
+    should create a pod from an image when restart is Never  [Conformance]
     /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] [sig-node] Pods Extended [k8s.io] Delete Grace Period should be submitted and removed [Conformance]","total":280,"completed":118,"skipped":1991,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never  [Conformance]","total":280,"completed":125,"skipped":1894,"failed":0}
+SSSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  works for multiple CRDs of different groups [Conformance]
+[sig-storage] EmptyDir wrapper volumes 
+  should not cause race condition when used for configmaps [Serial] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+[BeforeEach] [sig-storage] EmptyDir wrapper volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:23:39.204: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename crd-publish-openapi
+Oct 26 16:38:06.821: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir-wrapper
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] works for multiple CRDs of different groups [Conformance]
+[It] should not cause race condition when used for configmaps [Serial] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: CRs in different groups (two CRDs) show up in OpenAPI documentation
-Sep 21 16:23:39.438: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:23:48.453: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+STEP: Creating 50 configmaps
+STEP: Creating RC which spawns configmap-volume pods
+Oct 26 16:38:07.947: INFO: Pod name wrapped-volume-race-8a741f49-f3cb-44f2-9785-78e31475587a: Found 0 pods out of 5
+Oct 26 16:38:12.978: INFO: Pod name wrapped-volume-race-8a741f49-f3cb-44f2-9785-78e31475587a: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-8a741f49-f3cb-44f2-9785-78e31475587a in namespace emptydir-wrapper-1715, will wait for the garbage collector to delete the pods
+Oct 26 16:38:13.159: INFO: Deleting ReplicationController wrapped-volume-race-8a741f49-f3cb-44f2-9785-78e31475587a took: 26.264346ms
+Oct 26 16:38:13.559: INFO: Terminating ReplicationController wrapped-volume-race-8a741f49-f3cb-44f2-9785-78e31475587a pods took: 400.285323ms
+STEP: Creating RC which spawns configmap-volume pods
+Oct 26 16:38:19.938: INFO: Pod name wrapped-volume-race-0ab21807-b217-4c58-9545-a55e7be64563: Found 0 pods out of 5
+Oct 26 16:38:24.964: INFO: Pod name wrapped-volume-race-0ab21807-b217-4c58-9545-a55e7be64563: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-0ab21807-b217-4c58-9545-a55e7be64563 in namespace emptydir-wrapper-1715, will wait for the garbage collector to delete the pods
+Oct 26 16:38:27.154: INFO: Deleting ReplicationController wrapped-volume-race-0ab21807-b217-4c58-9545-a55e7be64563 took: 27.68912ms
+Oct 26 16:38:27.254: INFO: Terminating ReplicationController wrapped-volume-race-0ab21807-b217-4c58-9545-a55e7be64563 pods took: 100.285994ms
+STEP: Creating RC which spawns configmap-volume pods
+Oct 26 16:38:38.516: INFO: Pod name wrapped-volume-race-aca1843e-a833-469e-b896-1de8e51812e6: Found 0 pods out of 5
+Oct 26 16:38:43.538: INFO: Pod name wrapped-volume-race-aca1843e-a833-469e-b896-1de8e51812e6: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-aca1843e-a833-469e-b896-1de8e51812e6 in namespace emptydir-wrapper-1715, will wait for the garbage collector to delete the pods
+Oct 26 16:38:43.720: INFO: Deleting ReplicationController wrapped-volume-race-aca1843e-a833-469e-b896-1de8e51812e6 took: 35.123041ms
+Oct 26 16:38:44.021: INFO: Terminating ReplicationController wrapped-volume-race-aca1843e-a833-469e-b896-1de8e51812e6 pods took: 300.488332ms
+STEP: Cleaning up the configMaps
+[AfterEach] [sig-storage] EmptyDir wrapper volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:24:23.496: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-1999" for this suite.
+Oct 26 16:39:01.181: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-wrapper-1715" for this suite.
 
-• [SLOW TEST:44.361 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for multiple CRDs of different groups [Conformance]
+• [SLOW TEST:54.405 seconds]
+[sig-storage] EmptyDir wrapper volumes
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
+  should not cause race condition when used for configmaps [Serial] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance]","total":280,"completed":119,"skipped":2013,"failed":0}
-S
-------------------------------
-[sig-storage] Projected configMap 
-  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:24:23.565: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name projected-configmap-test-volume-84d2672b-eed9-4da3-898e-83031ce545d2
-STEP: Creating a pod to test consume configMaps
-Sep 21 16:24:24.039: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-d961c0ab-d6fb-488e-b946-97378a5306e0" in namespace "projected-2086" to be "success or failure"
-Sep 21 16:24:24.062: INFO: Pod "pod-projected-configmaps-d961c0ab-d6fb-488e-b946-97378a5306e0": Phase="Pending", Reason="", readiness=false. Elapsed: 22.522225ms
-Sep 21 16:24:26.094: INFO: Pod "pod-projected-configmaps-d961c0ab-d6fb-488e-b946-97378a5306e0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.0550668s
-STEP: Saw pod success
-Sep 21 16:24:26.094: INFO: Pod "pod-projected-configmaps-d961c0ab-d6fb-488e-b946-97378a5306e0" satisfied condition "success or failure"
-Sep 21 16:24:26.120: INFO: Trying to get logs from node 10.241.51.147 pod pod-projected-configmaps-d961c0ab-d6fb-488e-b946-97378a5306e0 container projected-configmap-volume-test: 
-STEP: delete the pod
-Sep 21 16:24:26.357: INFO: Waiting for pod pod-projected-configmaps-d961c0ab-d6fb-488e-b946-97378a5306e0 to disappear
-Sep 21 16:24:26.608: INFO: Pod pod-projected-configmaps-d961c0ab-d6fb-488e-b946-97378a5306e0 no longer exists
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:24:26.609: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-2086" for this suite.
-•{"msg":"PASSED [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":280,"completed":120,"skipped":2014,"failed":0}
-SSS
+{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance]","total":280,"completed":126,"skipped":1904,"failed":0}
+SSSSS
 ------------------------------
-[sig-cli] Kubectl client Proxy server 
-  should support --unix-socket=/path  [Conformance]
+[sig-cli] Kubectl client Kubectl replace 
+  should update a single-container pod's image  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:24:26.680: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 16:39:01.226: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename kubectl
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[It] should support --unix-socket=/path  [Conformance]
+[BeforeEach] Kubectl replace
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1897
+[It] should update a single-container pod's image  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Starting the proxy
-Sep 21 16:24:27.019: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-468631458 proxy --unix-socket=/tmp/kubectl-proxy-unix343611501/test'
-STEP: retrieving proxy /api/ output
+STEP: running the image docker.io/library/httpd:2.4.38-alpine
+Oct 26 16:39:01.390: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 run e2e-test-httpd-pod --generator=run-pod/v1 --image=docker.io/library/httpd:2.4.38-alpine --labels=run=e2e-test-httpd-pod --namespace=kubectl-2080'
+Oct 26 16:39:01.595: INFO: stderr: ""
+Oct 26 16:39:01.595: INFO: stdout: "pod/e2e-test-httpd-pod created\n"
+STEP: verifying the pod e2e-test-httpd-pod is running
+STEP: verifying the pod e2e-test-httpd-pod was created
+Oct 26 16:39:06.646: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pod e2e-test-httpd-pod --namespace=kubectl-2080 -o json'
+Oct 26 16:39:06.807: INFO: stderr: ""
+Oct 26 16:39:06.807: INFO: stdout: "{\n    \"apiVersion\": \"v1\",\n    \"kind\": \"Pod\",\n    \"metadata\": {\n        \"annotations\": {\n            \"cni.projectcalico.org/podIP\": \"172.30.194.124/32\",\n            \"cni.projectcalico.org/podIPs\": \"172.30.194.124/32\",\n            \"k8s.v1.cni.cncf.io/networks-status\": \"[{\\n    \\\"name\\\": \\\"k8s-pod-network\\\",\\n    \\\"ips\\\": [\\n        \\\"172.30.194.124\\\"\\n    ],\\n    \\\"dns\\\": {}\\n}]\",\n            \"openshift.io/scc\": \"anyuid\"\n        },\n        \"creationTimestamp\": \"2020-10-26T16:39:01Z\",\n        \"labels\": {\n            \"run\": \"e2e-test-httpd-pod\"\n        },\n        \"name\": \"e2e-test-httpd-pod\",\n        \"namespace\": \"kubectl-2080\",\n        \"resourceVersion\": \"67866\",\n        \"selfLink\": \"/api/v1/namespaces/kubectl-2080/pods/e2e-test-httpd-pod\",\n        \"uid\": \"01b8b898-782b-4f67-9afe-3da5ecd7e78f\"\n    },\n    \"spec\": {\n        \"containers\": [\n            {\n                \"image\": \"docker.io/library/httpd:2.4.38-alpine\",\n                \"imagePullPolicy\": \"IfNotPresent\",\n                \"name\": \"e2e-test-httpd-pod\",\n                \"resources\": {},\n                \"securityContext\": {\n                    \"capabilities\": {\n                        \"drop\": [\n                            \"MKNOD\"\n                        ]\n                    }\n                },\n                \"terminationMessagePath\": \"/dev/termination-log\",\n                \"terminationMessagePolicy\": \"File\",\n                \"volumeMounts\": [\n                    {\n                        \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n                        \"name\": \"default-token-fmqm6\",\n                        \"readOnly\": true\n                    }\n                ]\n            }\n        ],\n        \"dnsPolicy\": \"ClusterFirst\",\n        \"enableServiceLinks\": true,\n        \"imagePullSecrets\": [\n            {\n                \"name\": \"default-dockercfg-hmdzp\"\n            }\n        ],\n        \"nodeName\": \"10.72.119.74\",\n        \"priority\": 0,\n        \"restartPolicy\": \"Always\",\n        \"schedulerName\": \"default-scheduler\",\n        \"securityContext\": {\n            \"seLinuxOptions\": {\n                \"level\": \"s0:c54,c39\"\n            }\n        },\n        \"serviceAccount\": \"default\",\n        \"serviceAccountName\": \"default\",\n        \"terminationGracePeriodSeconds\": 30,\n        \"tolerations\": [\n            {\n                \"effect\": \"NoExecute\",\n                \"key\": \"node.kubernetes.io/not-ready\",\n                \"operator\": \"Exists\",\n                \"tolerationSeconds\": 300\n            },\n            {\n                \"effect\": \"NoExecute\",\n                \"key\": \"node.kubernetes.io/unreachable\",\n                \"operator\": \"Exists\",\n                \"tolerationSeconds\": 300\n            }\n        ],\n        \"volumes\": [\n            {\n                \"name\": \"default-token-fmqm6\",\n                \"secret\": {\n                    \"defaultMode\": 420,\n                    \"secretName\": \"default-token-fmqm6\"\n                }\n            }\n        ]\n    },\n    \"status\": {\n        \"conditions\": [\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2020-10-26T16:39:01Z\",\n                \"status\": \"True\",\n                \"type\": \"Initialized\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2020-10-26T16:39:03Z\",\n                \"status\": \"True\",\n                \"type\": \"Ready\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2020-10-26T16:39:03Z\",\n                \"status\": \"True\",\n                \"type\": \"ContainersReady\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2020-10-26T16:39:01Z\",\n                \"status\": \"True\",\n                \"type\": \"PodScheduled\"\n            }\n        ],\n        \"containerStatuses\": [\n            {\n                \"containerID\": \"cri-o://efceba6c483e37d9abcad394d1d3343f30c7482eb8707c76736ffc7abf033f4d\",\n                \"image\": \"docker.io/library/httpd:2.4.38-alpine\",\n                \"imageID\": \"docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4\",\n                \"lastState\": {},\n                \"name\": \"e2e-test-httpd-pod\",\n                \"ready\": true,\n                \"restartCount\": 0,\n                \"started\": true,\n                \"state\": {\n                    \"running\": {\n                        \"startedAt\": \"2020-10-26T16:39:03Z\"\n                    }\n                }\n            }\n        ],\n        \"hostIP\": \"10.72.119.74\",\n        \"phase\": \"Running\",\n        \"podIP\": \"172.30.194.124\",\n        \"podIPs\": [\n            {\n                \"ip\": \"172.30.194.124\"\n            }\n        ],\n        \"qosClass\": \"BestEffort\",\n        \"startTime\": \"2020-10-26T16:39:01Z\"\n    }\n}\n"
+STEP: replace the image in the pod
+Oct 26 16:39:06.807: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 replace -f - --namespace=kubectl-2080'
+Oct 26 16:39:07.402: INFO: stderr: ""
+Oct 26 16:39:07.402: INFO: stdout: "pod/e2e-test-httpd-pod replaced\n"
+STEP: verifying the pod e2e-test-httpd-pod has the right image docker.io/library/busybox:1.29
+[AfterEach] Kubectl replace
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1902
+Oct 26 16:39:07.424: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete pods e2e-test-httpd-pod --namespace=kubectl-2080'
+Oct 26 16:39:09.243: INFO: stderr: ""
+Oct 26 16:39:09.243: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n"
 [AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:24:27.091: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-7649" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support --unix-socket=/path  [Conformance]","total":280,"completed":121,"skipped":2017,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:39:09.243: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-2080" for this suite.
+
+• [SLOW TEST:8.056 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Kubectl replace
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1893
+    should update a single-container pod's image  [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] Docker Containers 
-  should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
+{"msg":"PASSED [sig-cli] Kubectl client Kubectl replace should update a single-container pod's image  [Conformance]","total":280,"completed":127,"skipped":1909,"failed":0}
+SSSSSSSSSSSSSSSS
+------------------------------
+[sig-network] Proxy version v1 
+  should proxy through a service and a pod  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Docker Containers
+[BeforeEach] version v1
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:24:27.221: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename containers
+Oct 26 16:39:09.283: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename proxy
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
+[It] should proxy through a service and a pod  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test override arguments
-Sep 21 16:24:27.719: INFO: Waiting up to 5m0s for pod "client-containers-1ac32aa2-152d-48cf-9b23-0652701f25d8" in namespace "containers-8758" to be "success or failure"
-Sep 21 16:24:27.761: INFO: Pod "client-containers-1ac32aa2-152d-48cf-9b23-0652701f25d8": Phase="Pending", Reason="", readiness=false. Elapsed: 42.484045ms
-Sep 21 16:24:29.822: INFO: Pod "client-containers-1ac32aa2-152d-48cf-9b23-0652701f25d8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.103005558s
-STEP: Saw pod success
-Sep 21 16:24:29.822: INFO: Pod "client-containers-1ac32aa2-152d-48cf-9b23-0652701f25d8" satisfied condition "success or failure"
-Sep 21 16:24:29.853: INFO: Trying to get logs from node 10.241.51.147 pod client-containers-1ac32aa2-152d-48cf-9b23-0652701f25d8 container test-container: 
-STEP: delete the pod
-Sep 21 16:24:30.021: INFO: Waiting for pod client-containers-1ac32aa2-152d-48cf-9b23-0652701f25d8 to disappear
-Sep 21 16:24:30.047: INFO: Pod client-containers-1ac32aa2-152d-48cf-9b23-0652701f25d8 no longer exists
-[AfterEach] [k8s.io] Docker Containers
+STEP: starting an echo server on multiple ports
+STEP: creating replication controller proxy-service-dmn6m in namespace proxy-8351
+I1026 16:39:09.483828      26 runners.go:189] Created replication controller with name: proxy-service-dmn6m, namespace: proxy-8351, replica count: 1
+I1026 16:39:10.534326      26 runners.go:189] proxy-service-dmn6m Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+I1026 16:39:11.534609      26 runners.go:189] proxy-service-dmn6m Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I1026 16:39:12.534781      26 runners.go:189] proxy-service-dmn6m Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I1026 16:39:13.535048      26 runners.go:189] proxy-service-dmn6m Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+Oct 26 16:39:13.550: INFO: setup took 4.136634631s, starting test cases
+STEP: running 16 cases, 20 attempts per case, 320 total attempts
+Oct 26 16:39:13.586: INFO: (0) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:1080/proxy/: ... (200; 33.782788ms)
+Oct 26 16:39:13.586: INFO: (0) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 35.596015ms)
+Oct 26 16:39:13.588: INFO: (0) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 37.802183ms)
+Oct 26 16:39:13.589: INFO: (0) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 36.82065ms)
+Oct 26 16:39:13.594: INFO: (0) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 43.207848ms)
+Oct 26 16:39:13.594: INFO: (0) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 41.698771ms)
+Oct 26 16:39:13.595: INFO: (0) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 41.438925ms)
+Oct 26 16:39:13.595: INFO: (0) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 41.284041ms)
+Oct 26 16:39:13.595: INFO: (0) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 44.562689ms)
+Oct 26 16:39:13.595: INFO: (0) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 43.207186ms)
+Oct 26 16:39:13.597: INFO: (0) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 43.969763ms)
+Oct 26 16:39:13.601: INFO: (0) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 49.247572ms)
+Oct 26 16:39:13.605: INFO: (0) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 53.973074ms)
+Oct 26 16:39:13.608: INFO: (0) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 55.43793ms)
+Oct 26 16:39:13.609: INFO: (0) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: ... (200; 33.953321ms)
+Oct 26 16:39:13.645: INFO: (1) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 34.885965ms)
+Oct 26 16:39:13.647: INFO: (1) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 38.00947ms)
+Oct 26 16:39:13.648: INFO: (1) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 38.511103ms)
+Oct 26 16:39:13.648: INFO: (1) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 38.489251ms)
+Oct 26 16:39:13.649: INFO: (1) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 39.359195ms)
+Oct 26 16:39:13.649: INFO: (1) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 39.472748ms)
+Oct 26 16:39:13.649: INFO: (1) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 39.602554ms)
+Oct 26 16:39:13.649: INFO: (1) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 39.486601ms)
+Oct 26 16:39:13.649: INFO: (1) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: test (200; 31.714752ms)
+Oct 26 16:39:13.691: INFO: (2) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:1080/proxy/: ... (200; 31.960398ms)
+Oct 26 16:39:13.692: INFO: (2) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 32.321795ms)
+Oct 26 16:39:13.692: INFO: (2) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 33.286173ms)
+Oct 26 16:39:13.692: INFO: (2) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 33.149931ms)
+Oct 26 16:39:13.692: INFO: (2) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 33.314133ms)
+Oct 26 16:39:13.693: INFO: (2) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 34.018968ms)
+Oct 26 16:39:13.695: INFO: (2) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 37.307518ms)
+Oct 26 16:39:13.702: INFO: (2) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 42.754189ms)
+Oct 26 16:39:13.706: INFO: (2) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 46.90189ms)
+Oct 26 16:39:13.707: INFO: (2) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 47.436909ms)
+Oct 26 16:39:13.707: INFO: (2) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 47.350557ms)
+Oct 26 16:39:13.707: INFO: (2) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 47.552992ms)
+Oct 26 16:39:13.729: INFO: (3) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: ... (200; 35.868949ms)
+Oct 26 16:39:13.743: INFO: (3) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 35.946318ms)
+Oct 26 16:39:13.743: INFO: (3) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 36.660564ms)
+Oct 26 16:39:13.743: INFO: (3) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 36.125479ms)
+Oct 26 16:39:13.744: INFO: (3) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 36.233438ms)
+Oct 26 16:39:13.744: INFO: (3) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 36.312498ms)
+Oct 26 16:39:13.744: INFO: (3) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 36.729905ms)
+Oct 26 16:39:13.744: INFO: (3) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 36.548656ms)
+Oct 26 16:39:13.744: INFO: (3) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 36.815425ms)
+Oct 26 16:39:13.749: INFO: (3) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 41.928395ms)
+Oct 26 16:39:13.751: INFO: (3) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 43.865868ms)
+Oct 26 16:39:13.753: INFO: (3) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 45.819067ms)
+Oct 26 16:39:13.758: INFO: (3) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 50.851996ms)
+Oct 26 16:39:13.760: INFO: (3) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 52.172496ms)
+Oct 26 16:39:13.760: INFO: (3) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 52.439097ms)
+Oct 26 16:39:13.780: INFO: (4) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: test<... (200; 25.865927ms)
+Oct 26 16:39:13.788: INFO: (4) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 24.430328ms)
+Oct 26 16:39:13.788: INFO: (4) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 24.760726ms)
+Oct 26 16:39:13.788: INFO: (4) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 26.252394ms)
+Oct 26 16:39:13.788: INFO: (4) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 26.53199ms)
+Oct 26 16:39:13.791: INFO: (4) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:1080/proxy/: ... (200; 27.695928ms)
+Oct 26 16:39:13.791: INFO: (4) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 29.334525ms)
+Oct 26 16:39:13.791: INFO: (4) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 30.238332ms)
+Oct 26 16:39:13.791: INFO: (4) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 30.491293ms)
+Oct 26 16:39:13.808: INFO: (4) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 46.105146ms)
+Oct 26 16:39:13.812: INFO: (4) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 50.049528ms)
+Oct 26 16:39:13.813: INFO: (4) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 52.152327ms)
+Oct 26 16:39:13.814: INFO: (4) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 51.48175ms)
+Oct 26 16:39:13.814: INFO: (4) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 53.583963ms)
+Oct 26 16:39:13.814: INFO: (4) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 54.114463ms)
+Oct 26 16:39:13.836: INFO: (5) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 21.473226ms)
+Oct 26 16:39:13.845: INFO: (5) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 29.242667ms)
+Oct 26 16:39:13.845: INFO: (5) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 27.960061ms)
+Oct 26 16:39:13.845: INFO: (5) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 28.950258ms)
+Oct 26 16:39:13.845: INFO: (5) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 28.20175ms)
+Oct 26 16:39:13.845: INFO: (5) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 29.161124ms)
+Oct 26 16:39:13.845: INFO: (5) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 27.844246ms)
+Oct 26 16:39:13.845: INFO: (5) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 28.061885ms)
+Oct 26 16:39:13.848: INFO: (5) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: ... (200; 32.500606ms)
+Oct 26 16:39:13.852: INFO: (5) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 37.689791ms)
+Oct 26 16:39:13.862: INFO: (5) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 44.540752ms)
+Oct 26 16:39:13.862: INFO: (5) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 45.887755ms)
+Oct 26 16:39:13.862: INFO: (5) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 45.542374ms)
+Oct 26 16:39:13.864: INFO: (5) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 49.412813ms)
+Oct 26 16:39:13.865: INFO: (5) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 48.4163ms)
+Oct 26 16:39:13.887: INFO: (6) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 22.31279ms)
+Oct 26 16:39:13.893: INFO: (6) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 28.139619ms)
+Oct 26 16:39:13.894: INFO: (6) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 28.401161ms)
+Oct 26 16:39:13.894: INFO: (6) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 28.414956ms)
+Oct 26 16:39:13.895: INFO: (6) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: test<... (200; 30.359953ms)
+Oct 26 16:39:13.896: INFO: (6) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 30.756888ms)
+Oct 26 16:39:13.896: INFO: (6) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:1080/proxy/: ... (200; 30.455495ms)
+Oct 26 16:39:13.897: INFO: (6) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 31.988016ms)
+Oct 26 16:39:13.900: INFO: (6) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 35.693964ms)
+Oct 26 16:39:13.906: INFO: (6) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 41.660479ms)
+Oct 26 16:39:13.907: INFO: (6) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 41.405986ms)
+Oct 26 16:39:13.907: INFO: (6) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 41.470969ms)
+Oct 26 16:39:13.907: INFO: (6) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 41.843217ms)
+Oct 26 16:39:13.907: INFO: (6) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 41.675654ms)
+Oct 26 16:39:13.929: INFO: (7) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 21.820629ms)
+Oct 26 16:39:13.936: INFO: (7) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 27.621331ms)
+Oct 26 16:39:13.938: INFO: (7) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 30.462436ms)
+Oct 26 16:39:13.939: INFO: (7) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: test (200; 31.796607ms)
+Oct 26 16:39:13.939: INFO: (7) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:1080/proxy/: ... (200; 31.234693ms)
+Oct 26 16:39:13.939: INFO: (7) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 31.868791ms)
+Oct 26 16:39:13.939: INFO: (7) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 31.604844ms)
+Oct 26 16:39:13.948: INFO: (7) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 39.974147ms)
+Oct 26 16:39:13.955: INFO: (7) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 47.169852ms)
+Oct 26 16:39:13.955: INFO: (7) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 47.301812ms)
+Oct 26 16:39:13.955: INFO: (7) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 47.425168ms)
+Oct 26 16:39:13.956: INFO: (7) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 47.512982ms)
+Oct 26 16:39:13.956: INFO: (7) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 47.748168ms)
+Oct 26 16:39:13.977: INFO: (8) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 21.078021ms)
+Oct 26 16:39:13.983: INFO: (8) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 26.753002ms)
+Oct 26 16:39:13.983: INFO: (8) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 26.849879ms)
+Oct 26 16:39:13.990: INFO: (8) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 33.68754ms)
+Oct 26 16:39:13.991: INFO: (8) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: ... (200; 35.81372ms)
+Oct 26 16:39:13.993: INFO: (8) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 35.768003ms)
+Oct 26 16:39:14.002: INFO: (8) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 44.937607ms)
+Oct 26 16:39:14.004: INFO: (8) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 47.969426ms)
+Oct 26 16:39:14.011: INFO: (8) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 54.23787ms)
+Oct 26 16:39:14.011: INFO: (8) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 54.844494ms)
+Oct 26 16:39:14.011: INFO: (8) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 54.231032ms)
+Oct 26 16:39:14.013: INFO: (8) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 56.613939ms)
+Oct 26 16:39:14.036: INFO: (9) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 23.370297ms)
+Oct 26 16:39:14.045: INFO: (9) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: test (200; 40.255897ms)
+Oct 26 16:39:14.054: INFO: (9) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 40.180295ms)
+Oct 26 16:39:14.054: INFO: (9) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:1080/proxy/: ... (200; 40.143222ms)
+Oct 26 16:39:14.058: INFO: (9) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 44.9425ms)
+Oct 26 16:39:14.065: INFO: (9) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 51.090152ms)
+Oct 26 16:39:14.067: INFO: (9) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 53.280647ms)
+Oct 26 16:39:14.067: INFO: (9) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 53.604761ms)
+Oct 26 16:39:14.067: INFO: (9) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 53.653301ms)
+Oct 26 16:39:14.067: INFO: (9) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 53.652054ms)
+Oct 26 16:39:14.093: INFO: (10) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 25.205074ms)
+Oct 26 16:39:14.098: INFO: (10) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:1080/proxy/: ... (200; 29.832854ms)
+Oct 26 16:39:14.098: INFO: (10) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 30.06459ms)
+Oct 26 16:39:14.099: INFO: (10) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 31.200313ms)
+Oct 26 16:39:14.099: INFO: (10) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 30.791515ms)
+Oct 26 16:39:14.100: INFO: (10) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 32.255755ms)
+Oct 26 16:39:14.100: INFO: (10) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: test (200; 32.906454ms)
+Oct 26 16:39:14.110: INFO: (10) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 42.64693ms)
+Oct 26 16:39:14.119: INFO: (10) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 51.348254ms)
+Oct 26 16:39:14.120: INFO: (10) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 52.070072ms)
+Oct 26 16:39:14.120: INFO: (10) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 52.44954ms)
+Oct 26 16:39:14.120: INFO: (10) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 52.544216ms)
+Oct 26 16:39:14.121: INFO: (10) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 52.602662ms)
+Oct 26 16:39:14.144: INFO: (11) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 22.705431ms)
+Oct 26 16:39:14.154: INFO: (11) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 33.151506ms)
+Oct 26 16:39:14.154: INFO: (11) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 33.478422ms)
+Oct 26 16:39:14.155: INFO: (11) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 33.412123ms)
+Oct 26 16:39:14.155: INFO: (11) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:1080/proxy/: ... (200; 33.479194ms)
+Oct 26 16:39:14.155: INFO: (11) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 33.587489ms)
+Oct 26 16:39:14.155: INFO: (11) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 34.194072ms)
+Oct 26 16:39:14.155: INFO: (11) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 34.396527ms)
+Oct 26 16:39:14.156: INFO: (11) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 34.586299ms)
+Oct 26 16:39:14.156: INFO: (11) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: test (200; 36.520704ms)
+Oct 26 16:39:14.224: INFO: (12) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:1080/proxy/: ... (200; 38.22215ms)
+Oct 26 16:39:14.224: INFO: (12) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 38.457164ms)
+Oct 26 16:39:14.225: INFO: (12) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 38.842069ms)
+Oct 26 16:39:14.225: INFO: (12) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 38.802468ms)
+Oct 26 16:39:14.225: INFO: (12) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 38.708682ms)
+Oct 26 16:39:14.225: INFO: (12) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 38.863627ms)
+Oct 26 16:39:14.242: INFO: (12) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 56.014113ms)
+Oct 26 16:39:14.254: INFO: (12) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 67.904698ms)
+Oct 26 16:39:14.258: INFO: (12) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 72.085295ms)
+Oct 26 16:39:14.259: INFO: (12) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 72.831466ms)
+Oct 26 16:39:14.259: INFO: (12) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 72.977932ms)
+Oct 26 16:39:14.259: INFO: (12) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 73.063106ms)
+Oct 26 16:39:14.281: INFO: (13) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 21.517533ms)
+Oct 26 16:39:14.290: INFO: (13) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: ... (200; 30.431338ms)
+Oct 26 16:39:14.291: INFO: (13) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 31.553306ms)
+Oct 26 16:39:14.291: INFO: (13) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 31.657397ms)
+Oct 26 16:39:14.292: INFO: (13) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 31.961057ms)
+Oct 26 16:39:14.292: INFO: (13) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 32.563165ms)
+Oct 26 16:39:14.300: INFO: (13) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 39.370123ms)
+Oct 26 16:39:14.300: INFO: (13) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 39.90675ms)
+Oct 26 16:39:14.300: INFO: (13) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 40.188201ms)
+Oct 26 16:39:14.300: INFO: (13) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 40.298494ms)
+Oct 26 16:39:14.300: INFO: (13) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 39.557135ms)
+Oct 26 16:39:14.321: INFO: (14) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:1080/proxy/: ... (200; 20.711566ms)
+Oct 26 16:39:14.328: INFO: (14) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 27.219881ms)
+Oct 26 16:39:14.329: INFO: (14) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: test<... (200; 28.190324ms)
+Oct 26 16:39:14.329: INFO: (14) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 28.101601ms)
+Oct 26 16:39:14.329: INFO: (14) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 28.298635ms)
+Oct 26 16:39:14.329: INFO: (14) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 28.644631ms)
+Oct 26 16:39:14.330: INFO: (14) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 29.634059ms)
+Oct 26 16:39:14.330: INFO: (14) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 29.667741ms)
+Oct 26 16:39:14.331: INFO: (14) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 29.752572ms)
+Oct 26 16:39:14.338: INFO: (14) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 37.545114ms)
+Oct 26 16:39:14.342: INFO: (14) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 40.835149ms)
+Oct 26 16:39:14.344: INFO: (14) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 42.684904ms)
+Oct 26 16:39:14.344: INFO: (14) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 42.879087ms)
+Oct 26 16:39:14.344: INFO: (14) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 42.927182ms)
+Oct 26 16:39:14.344: INFO: (14) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 43.275787ms)
+Oct 26 16:39:14.367: INFO: (15) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 22.63741ms)
+Oct 26 16:39:14.375: INFO: (15) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 29.697975ms)
+Oct 26 16:39:14.375: INFO: (15) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: ... (200; 30.567669ms)
+Oct 26 16:39:14.376: INFO: (15) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 30.637971ms)
+Oct 26 16:39:14.376: INFO: (15) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 30.75458ms)
+Oct 26 16:39:14.377: INFO: (15) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 31.516837ms)
+Oct 26 16:39:14.380: INFO: (15) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 34.731561ms)
+Oct 26 16:39:14.383: INFO: (15) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 38.011882ms)
+Oct 26 16:39:14.385: INFO: (15) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 40.124511ms)
+Oct 26 16:39:14.386: INFO: (15) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 40.818376ms)
+Oct 26 16:39:14.387: INFO: (15) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 42.048827ms)
+Oct 26 16:39:14.388: INFO: (15) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 42.691714ms)
+Oct 26 16:39:14.388: INFO: (15) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 42.83127ms)
+Oct 26 16:39:14.413: INFO: (16) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 24.86109ms)
+Oct 26 16:39:14.417: INFO: (16) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 27.22062ms)
+Oct 26 16:39:14.417: INFO: (16) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 29.010908ms)
+Oct 26 16:39:14.418: INFO: (16) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: ... (200; 27.302756ms)
+Oct 26 16:39:14.418: INFO: (16) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 28.789084ms)
+Oct 26 16:39:14.418: INFO: (16) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 28.426299ms)
+Oct 26 16:39:14.419: INFO: (16) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 29.442437ms)
+Oct 26 16:39:14.419: INFO: (16) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 30.276279ms)
+Oct 26 16:39:14.426: INFO: (16) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 37.055788ms)
+Oct 26 16:39:14.435: INFO: (16) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 44.432885ms)
+Oct 26 16:39:14.443: INFO: (16) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 53.552166ms)
+Oct 26 16:39:14.443: INFO: (16) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 52.775481ms)
+Oct 26 16:39:14.443: INFO: (16) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 52.97131ms)
+Oct 26 16:39:14.443: INFO: (16) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 53.927063ms)
+Oct 26 16:39:14.465: INFO: (17) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: test<... (200; 27.345866ms)
+Oct 26 16:39:14.472: INFO: (17) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 27.818868ms)
+Oct 26 16:39:14.472: INFO: (17) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:1080/proxy/: ... (200; 27.745466ms)
+Oct 26 16:39:14.472: INFO: (17) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 27.751433ms)
+Oct 26 16:39:14.474: INFO: (17) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 29.681765ms)
+Oct 26 16:39:14.474: INFO: (17) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 29.92311ms)
+Oct 26 16:39:14.474: INFO: (17) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:460/proxy/: tls baz (200; 30.236707ms)
+Oct 26 16:39:14.475: INFO: (17) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 30.256818ms)
+Oct 26 16:39:14.482: INFO: (17) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 38.504034ms)
+Oct 26 16:39:14.489: INFO: (17) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 44.81869ms)
+Oct 26 16:39:14.489: INFO: (17) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 44.613974ms)
+Oct 26 16:39:14.489: INFO: (17) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 44.810826ms)
+Oct 26 16:39:14.489: INFO: (17) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 44.82121ms)
+Oct 26 16:39:14.489: INFO: (17) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 44.90755ms)
+Oct 26 16:39:14.512: INFO: (18) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 22.100153ms)
+Oct 26 16:39:14.517: INFO: (18) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 25.912068ms)
+Oct 26 16:39:14.518: INFO: (18) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: ... (200; 33.978808ms)
+Oct 26 16:39:14.525: INFO: (18) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs/proxy/: test (200; 35.314491ms)
+Oct 26 16:39:14.525: INFO: (18) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 34.168384ms)
+Oct 26 16:39:14.525: INFO: (18) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 35.815211ms)
+Oct 26 16:39:14.525: INFO: (18) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 34.570034ms)
+Oct 26 16:39:14.532: INFO: (18) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 41.424476ms)
+Oct 26 16:39:14.532: INFO: (18) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 41.232967ms)
+Oct 26 16:39:14.535: INFO: (18) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 44.546614ms)
+Oct 26 16:39:14.537: INFO: (18) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 46.177292ms)
+Oct 26 16:39:14.538: INFO: (18) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 47.114655ms)
+Oct 26 16:39:14.538: INFO: (18) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 47.271354ms)
+Oct 26 16:39:14.565: INFO: (19) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:443/proxy/: test (200; 31.553063ms)
+Oct 26 16:39:14.573: INFO: (19) /api/v1/namespaces/proxy-8351/pods/https:proxy-service-dmn6m-jnlfs:462/proxy/: tls qux (200; 34.217145ms)
+Oct 26 16:39:14.573: INFO: (19) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:1080/proxy/: test<... (200; 34.201808ms)
+Oct 26 16:39:14.573: INFO: (19) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 34.364812ms)
+Oct 26 16:39:14.574: INFO: (19) /api/v1/namespaces/proxy-8351/pods/proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 34.52943ms)
+Oct 26 16:39:14.574: INFO: (19) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:1080/proxy/: ... (200; 34.95001ms)
+Oct 26 16:39:14.574: INFO: (19) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:160/proxy/: foo (200; 35.079987ms)
+Oct 26 16:39:14.574: INFO: (19) /api/v1/namespaces/proxy-8351/pods/http:proxy-service-dmn6m-jnlfs:162/proxy/: bar (200; 35.236275ms)
+Oct 26 16:39:14.579: INFO: (19) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname2/proxy/: bar (200; 40.636941ms)
+Oct 26 16:39:14.580: INFO: (19) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname1/proxy/: foo (200; 41.54125ms)
+Oct 26 16:39:14.582: INFO: (19) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname1/proxy/: tls baz (200; 43.038686ms)
+Oct 26 16:39:14.582: INFO: (19) /api/v1/namespaces/proxy-8351/services/https:proxy-service-dmn6m:tlsportname2/proxy/: tls qux (200; 43.168012ms)
+Oct 26 16:39:14.583: INFO: (19) /api/v1/namespaces/proxy-8351/services/proxy-service-dmn6m:portname2/proxy/: bar (200; 44.497584ms)
+Oct 26 16:39:14.584: INFO: (19) /api/v1/namespaces/proxy-8351/services/http:proxy-service-dmn6m:portname1/proxy/: foo (200; 44.461797ms)
+STEP: deleting ReplicationController proxy-service-dmn6m in namespace proxy-8351, will wait for the garbage collector to delete the pods
+Oct 26 16:39:14.672: INFO: Deleting ReplicationController proxy-service-dmn6m took: 24.365773ms
+Oct 26 16:39:14.972: INFO: Terminating ReplicationController proxy-service-dmn6m pods took: 300.198024ms
+[AfterEach] version v1
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:24:30.047: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "containers-8758" for this suite.
-•{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]","total":280,"completed":122,"skipped":2064,"failed":0}
-S
+Oct 26 16:39:17.372: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "proxy-8351" for this suite.
+
+• [SLOW TEST:8.131 seconds]
+[sig-network] Proxy
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  version v1
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:57
+    should proxy through a service and a pod  [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] 
-  custom resource defaulting for requests and from storage works  [Conformance]
+{"msg":"PASSED [sig-network] Proxy version v1 should proxy through a service and a pod  [Conformance]","total":280,"completed":128,"skipped":1925,"failed":0}
+SSSS
+------------------------------
+[k8s.io] Kubelet when scheduling a read only busybox container 
+  should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+[BeforeEach] [k8s.io] Kubelet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:24:30.115: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename custom-resource-definition
+Oct 26 16:39:17.415: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubelet-test
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] custom resource defaulting for requests and from storage works  [Conformance]
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:24:30.370: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+[AfterEach] [k8s.io] Kubelet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:24:32.456: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "custom-resource-definition-4217" for this suite.
-•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works  [Conformance]","total":280,"completed":123,"skipped":2065,"failed":0}
-SSS
+Oct 26 16:39:19.703: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-413" for this suite.
+•{"msg":"PASSED [k8s.io] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":129,"skipped":1929,"failed":0}
+SSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected configMap 
-  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
+[k8s.io] Container Runtime blackbox test on terminated container 
+  should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
+[BeforeEach] [k8s.io] Container Runtime
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:24:32.549: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 16:39:19.756: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename container-runtime
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
+[It] should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name projected-configmap-test-volume-53f4c8f6-c5df-4b25-91bc-b98632a27d04
-STEP: Creating a pod to test consume configMaps
-Sep 21 16:24:33.052: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-c406c77e-0c1d-4dd0-bea8-c72880aab023" in namespace "projected-2151" to be "success or failure"
-Sep 21 16:24:33.084: INFO: Pod "pod-projected-configmaps-c406c77e-0c1d-4dd0-bea8-c72880aab023": Phase="Pending", Reason="", readiness=false. Elapsed: 31.713902ms
-Sep 21 16:24:35.112: INFO: Pod "pod-projected-configmaps-c406c77e-0c1d-4dd0-bea8-c72880aab023": Phase="Pending", Reason="", readiness=false. Elapsed: 2.059967362s
-Sep 21 16:24:37.149: INFO: Pod "pod-projected-configmaps-c406c77e-0c1d-4dd0-bea8-c72880aab023": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.097236614s
-STEP: Saw pod success
-Sep 21 16:24:37.149: INFO: Pod "pod-projected-configmaps-c406c77e-0c1d-4dd0-bea8-c72880aab023" satisfied condition "success or failure"
-Sep 21 16:24:37.418: INFO: Trying to get logs from node 10.241.51.147 pod pod-projected-configmaps-c406c77e-0c1d-4dd0-bea8-c72880aab023 container projected-configmap-volume-test: 
-STEP: delete the pod
-Sep 21 16:24:37.625: INFO: Waiting for pod pod-projected-configmaps-c406c77e-0c1d-4dd0-bea8-c72880aab023 to disappear
-Sep 21 16:24:37.658: INFO: Pod pod-projected-configmaps-c406c77e-0c1d-4dd0-bea8-c72880aab023 no longer exists
-[AfterEach] [sig-storage] Projected configMap
+STEP: create the container
+STEP: wait for the container to reach Succeeded
+STEP: get the container status
+STEP: the container should be terminated
+STEP: the termination message should be set
+Oct 26 16:39:23.136: INFO: Expected: &{} to match Container's Termination Message:  --
+STEP: delete the container
+[AfterEach] [k8s.io] Container Runtime
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:24:37.658: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-2151" for this suite.
-
-• [SLOW TEST:5.191 seconds]
-[sig-storage] Projected configMap
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
-  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","total":280,"completed":124,"skipped":2068,"failed":0}
-S
+Oct 26 16:39:23.208: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-runtime-6703" for this suite.
+•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":280,"completed":130,"skipped":1944,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Kubelet when scheduling a busybox command in a pod 
-  should print the output to logs [NodeConformance] [Conformance]
+[sig-network] Services 
+  should be able to change the type from ExternalName to ClusterIP [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Kubelet
+[BeforeEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:24:37.740: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubelet-test
+Oct 26 16:39:23.246: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename services
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
-[It] should print the output to logs [NodeConformance] [Conformance]
+[BeforeEach] [sig-network] Services
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
+[It] should be able to change the type from ExternalName to ClusterIP [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [k8s.io] Kubelet
+STEP: creating a service externalname-service with the type=ExternalName in namespace services-7195
+STEP: changing the ExternalName service to type=ClusterIP
+STEP: creating replication controller externalname-service in namespace services-7195
+I1026 16:39:23.495703      26 runners.go:189] Created replication controller with name: externalname-service, namespace: services-7195, replica count: 2
+Oct 26 16:39:26.546: INFO: Creating new exec pod
+I1026 16:39:26.546198      26 runners.go:189] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+Oct 26 16:39:29.656: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-7195 execpods2958 -- /bin/sh -x -c nc -zv -t -w 2 externalname-service 80'
+Oct 26 16:39:30.117: INFO: stderr: "+ nc -zv -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n"
+Oct 26 16:39:30.117: INFO: stdout: ""
+Oct 26 16:39:30.117: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-7195 execpods2958 -- /bin/sh -x -c nc -zv -t -w 2 172.21.62.214 80'
+Oct 26 16:39:30.558: INFO: stderr: "+ nc -zv -t -w 2 172.21.62.214 80\nConnection to 172.21.62.214 80 port [tcp/http] succeeded!\n"
+Oct 26 16:39:30.558: INFO: stdout: ""
+Oct 26 16:39:30.558: INFO: Cleaning up the ExternalName to ClusterIP test service
+[AfterEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:24:40.239: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubelet-test-346" for this suite.
-•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance]","total":280,"completed":125,"skipped":2069,"failed":0}
-SSSSSSSS
+Oct 26 16:39:30.637: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-7195" for this suite.
+[AfterEach] [sig-network] Services
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
+
+• [SLOW TEST:7.433 seconds]
+[sig-network] Services
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should be able to change the type from ExternalName to ClusterIP [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates that NodeSelector is respected if not matching  [Conformance]
+{"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance]","total":280,"completed":131,"skipped":1967,"failed":0}
+SSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] Deployment 
+  deployment should support rollover [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+[BeforeEach] [sig-apps] Deployment
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:24:40.314: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename sched-pred
+Oct 26 16:39:30.680: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename deployment
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
-Sep 21 16:24:40.560: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-Sep 21 16:24:40.681: INFO: Waiting for terminating namespaces to be deleted...
-Sep 21 16:24:40.724: INFO: 
-Logging pods the kubelet thinks is on node 10.241.51.147 before test
-Sep 21 16:24:40.874: INFO: multus-7xt8r from openshift-multus started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container kube-multus ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: ibm-master-proxy-static-10.241.51.147 from kube-system started at 2020-09-21 14:11:47 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: 	Container pause ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: busybox-scheduling-b6e7c711-fc10-4ccc-8c2d-bc79e8c4d4fa from kubelet-test-346 started at 2020-09-21 16:24:38 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container busybox-scheduling-b6e7c711-fc10-4ccc-8c2d-bc79e8c4d4fa ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: sonobuoy-e2e-job-36e1517c33ed4f81 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container e2e ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: node-exporter-46mwx from openshift-monitoring started at 2020-09-21 14:11:49 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: 	Container node-exporter ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: node-ca-6r9nd from openshift-image-registry started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container node-ca ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: dns-default-f6vtl from openshift-dns started at 2020-09-21 14:11:49 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container dns ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: 	Container dns-node-resolver ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: sonobuoy from sonobuoy started at 2020-09-21 15:39:44 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: tuned-6qvw2 from openshift-cluster-node-tuning-operator started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container tuned ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: calico-typha-6fc9ff8b5f-9pmb4 from calico-system started at 2020-09-21 14:13:14 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container calico-typha ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-shgq6 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: 	Container systemd-logs ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: ibm-keepalived-watcher-h4dbn from kube-system started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: calico-node-t2n87 from calico-system started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container calico-node ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: ibmcloud-block-storage-driver-8r6lc from kube-system started at 2020-09-21 14:12:04 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: openshift-kube-proxy-tvbvw from openshift-kube-proxy started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container kube-proxy ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: multus-admission-controller-7rkpv from openshift-multus started at 2020-09-21 16:14:37 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:40.874: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: 	Container multus-admission-controller ready: true, restart count 0
-Sep 21 16:24:40.874: INFO: 
-Logging pods the kubelet thinks is on node 10.241.51.149 before test
-Sep 21 16:24:41.069: INFO: router-default-84dfcdc9b5-tk98j from openshift-ingress started at 2020-09-21 14:09:06 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.069: INFO: 	Container router ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: downloads-7989988989-9rhh4 from openshift-console started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.069: INFO: 	Container download-server ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: tuned-gpfhj from openshift-cluster-node-tuning-operator started at 2020-09-21 14:08:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.069: INFO: 	Container tuned ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: node-exporter-jnrzm from openshift-monitoring started at 2020-09-21 14:08:24 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.069: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: 	Container node-exporter ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: dns-default-958xc from openshift-dns started at 2020-09-21 14:09:05 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.069: INFO: 	Container dns ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: 	Container dns-node-resolver ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: alertmanager-main-0 from openshift-monitoring started at 2020-09-21 14:14:12 +0000 UTC (3 container statuses recorded)
-Sep 21 16:24:41.069: INFO: 	Container alertmanager ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: 	Container alertmanager-proxy ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: 	Container config-reloader ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: ibm-keepalived-watcher-r4tn7 from kube-system started at 2020-09-21 14:06:04 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.069: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: marketplace-operator-56c6694c8f-x9rcq from openshift-marketplace started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.069: INFO: 	Container marketplace-operator ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: downloads-7989988989-cjvjw from openshift-console started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.069: INFO: 	Container download-server ready: true, restart count 0
-Sep 21 16:24:41.069: INFO: ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-44pnn from ibm-system started at 2020-09-21 14:12:56 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container ibm-cloud-provider-ip-169-48-211-85 ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: tigera-operator-679798d94d-hdlvj from tigera-operator started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container tigera-operator ready: true, restart count 2
-Sep 21 16:24:41.070: INFO: cluster-image-registry-operator-695bf78ffc-95l5m from openshift-image-registry started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container cluster-image-registry-operator ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container cluster-image-registry-operator-watch ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: console-operator-58fd84b95c-dc6vq from openshift-console-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container console-operator ready: true, restart count 1
-Sep 21 16:24:41.070: INFO: migrator-6c94c8c775-vxsnp from openshift-kube-storage-version-migrator started at 2020-09-21 14:07:55 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container migrator ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: cluster-storage-operator-5bcf8d9f8f-dls4n from openshift-cluster-storage-operator started at 2020-09-21 14:07:36 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container cluster-storage-operator ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: image-registry-775dd777d7-5qtdd from openshift-image-registry started at 2020-09-21 14:11:01 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container registry ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: ibm-master-proxy-static-10.241.51.149 from kube-system started at 2020-09-21 14:06:02 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container pause ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: cluster-node-tuning-operator-bdbf98554-22kv8 from openshift-cluster-node-tuning-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container cluster-node-tuning-operator ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: redhat-operators-54dcdd9697-pcwgz from openshift-marketplace started at 2020-09-21 14:09:34 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container redhat-operators ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: registry-pvc-permissions-rlfxm from openshift-image-registry started at 2020-09-21 14:11:01 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container pvc-permissions ready: false, restart count 0
-Sep 21 16:24:41.070: INFO: prometheus-operator-8d96bcc68-ct2jw from openshift-monitoring started at 2020-09-21 16:14:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container prometheus-operator ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: calico-node-9gqht from calico-system started at 2020-09-21 14:07:18 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container calico-node ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: openshift-service-catalog-controller-manager-operator-b9ccczlcs from openshift-service-catalog-controller-manager-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container operator ready: true, restart count 1
-Sep 21 16:24:41.070: INFO: olm-operator-b5f47bf98-8d8cw from openshift-operator-lifecycle-manager started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container olm-operator ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: console-8549dd6c4f-8vk99 from openshift-console started at 2020-09-21 16:14:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container console ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: node-ca-bxmsf from openshift-image-registry started at 2020-09-21 14:09:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container node-ca ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: ibm-file-plugin-66fdcfc4c9-tm6qk from kube-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: packageserver-65fd8fc8fc-2mnzj from openshift-operator-lifecycle-manager started at 2020-09-21 16:14:19 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container packageserver ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: cluster-samples-operator-5cfb985bf-n7nxd from openshift-cluster-samples-operator started at 2020-09-21 14:09:54 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container cluster-samples-operator ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container cluster-samples-operator-watch ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: openshift-service-catalog-apiserver-operator-6988647bc8-t9z7w from openshift-service-catalog-apiserver-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container operator ready: true, restart count 1
-Sep 21 16:24:41.070: INFO: ibmcloud-block-storage-plugin-68d5c65db9-7q8bv from kube-system started at 2020-09-21 14:07:36 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container ibmcloud-block-storage-plugin-container ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: openshift-state-metrics-678b6c786f-698gn from openshift-monitoring started at 2020-09-21 14:08:22 +0000 UTC (3 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container openshift-state-metrics ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: redhat-marketplace-7b9bcdf96b-p4khw from openshift-marketplace started at 2020-09-21 14:09:34 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container redhat-marketplace ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: multus-hprgb from openshift-multus started at 2020-09-21 14:06:32 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container kube-multus ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: calico-typha-6fc9ff8b5f-gzhgw from calico-system started at 2020-09-21 14:07:18 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container calico-typha ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: kube-state-metrics-7455b6c8-455w7 from openshift-monitoring started at 2020-09-21 14:08:22 +0000 UTC (3 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container kube-state-metrics ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: prometheus-adapter-5f78ddc679-vb486 from openshift-monitoring started at 2020-09-21 14:08:28 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container prometheus-adapter ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: network-operator-f456cfdbb-4m6pj from openshift-network-operator started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container network-operator ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: cluster-monitoring-operator-75f4ddd85f-579q6 from openshift-monitoring started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container cluster-monitoring-operator ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: kube-storage-version-migrator-operator-86d8fc6986-vl4fl from openshift-kube-storage-version-migrator-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container kube-storage-version-migrator-operator ready: true, restart count 1
-Sep 21 16:24:41.070: INFO: calico-kube-controllers-79d75767dd-xkvb6 from calico-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container calico-kube-controllers ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: multus-admission-controller-gkchr from openshift-multus started at 2020-09-21 14:07:34 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container multus-admission-controller ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: prometheus-adapter-5f78ddc679-9hj7w from openshift-monitoring started at 2020-09-21 14:08:28 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container prometheus-adapter ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: community-operators-6566687cd-smnd2 from openshift-marketplace started at 2020-09-21 14:09:35 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container community-operators ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: ibm-storage-watcher-7956f954f-26xx6 from kube-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: service-ca-bd8866b9b-qzggv from openshift-service-ca started at 2020-09-21 14:07:50 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container service-ca-controller ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: certified-operators-5cfd68758d-q6c6n from openshift-marketplace started at 2020-09-21 16:14:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container certified-operators ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: ibmcloud-block-storage-driver-pldbm from kube-system started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: dns-operator-7886f4f4f4-rk7dv from openshift-dns-operator started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container dns-operator ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: prometheus-k8s-1 from openshift-monitoring started at 2020-09-21 16:14:17 +0000 UTC (7 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container prometheus ready: true, restart count 1
-Sep 21 16:24:41.070: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container prometheus-proxy ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container thanos-sidecar ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: openshift-kube-proxy-9svk2 from openshift-kube-proxy started at 2020-09-21 14:06:38 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container kube-proxy ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: service-ca-operator-9f84dfdb5-wdhvg from openshift-service-ca-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container operator ready: true, restart count 1
-Sep 21 16:24:41.070: INFO: ingress-operator-57546dfcd9-2sg5v from openshift-ingress-operator started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container ingress-operator ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: catalog-operator-69d4f5c59f-lqjgz from openshift-operator-lifecycle-manager started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container catalog-operator ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-lgbbw from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.070: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 	Container systemd-logs ready: true, restart count 0
-Sep 21 16:24:41.070: INFO: 
-Logging pods the kubelet thinks is on node 10.241.51.150 before test
-Sep 21 16:24:41.242: INFO: ibm-keepalived-watcher-bntgc from kube-system started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: alertmanager-main-2 from openshift-monitoring started at 2020-09-21 16:14:17 +0000 UTC (3 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container alertmanager ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container alertmanager-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container config-reloader ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-xzz27 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container systemd-logs ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: node-exporter-flb5l from openshift-monitoring started at 2020-09-21 14:12:21 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container node-exporter ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: calico-node-dn4jv from calico-system started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container calico-node ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: console-8549dd6c4f-b5jnn from openshift-console started at 2020-09-21 14:18:13 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container console ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: grafana-c745ff78c-9pgfx from openshift-monitoring started at 2020-09-21 14:13:42 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container grafana ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container grafana-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: dns-default-z8v9j from openshift-dns started at 2020-09-21 14:12:21 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container dns ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container dns-node-resolver ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: calico-typha-6fc9ff8b5f-cbhcz from calico-system started at 2020-09-21 14:13:14 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container calico-typha ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-ghw5q from ibm-system started at 2020-09-21 16:07:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container ibm-cloud-provider-ip-169-48-211-85 ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: multus-n5qc8 from openshift-multus started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container kube-multus ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: openshift-kube-proxy-9bv2v from openshift-kube-proxy started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container kube-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: multus-admission-controller-snglx from openshift-multus started at 2020-09-21 14:13:31 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container multus-admission-controller ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: alertmanager-main-1 from openshift-monitoring started at 2020-09-21 14:14:02 +0000 UTC (3 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container alertmanager ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container alertmanager-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container config-reloader ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: router-default-84dfcdc9b5-mt4b7 from openshift-ingress started at 2020-09-21 16:07:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container router ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: ibm-master-proxy-static-10.241.51.150 from kube-system started at 2020-09-21 14:12:19 +0000 UTC (2 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container pause ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: telemeter-client-68bdb7f795-nnhfl from openshift-monitoring started at 2020-09-21 14:13:34 +0000 UTC (3 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container reload ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container telemeter-client ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: prometheus-k8s-0 from openshift-monitoring started at 2020-09-21 14:15:10 +0000 UTC (7 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container prometheus ready: true, restart count 1
-Sep 21 16:24:41.242: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container prometheus-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container thanos-sidecar ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: ibmcloud-block-storage-driver-jf9s4 from kube-system started at 2020-09-21 14:12:26 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: thanos-querier-8f4c5c746-v982g from openshift-monitoring started at 2020-09-21 14:14:42 +0000 UTC (4 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container oauth-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container thanos-querier ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: vpn-64d8d4987c-lxt95 from kube-system started at 2020-09-21 16:07:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container vpn ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: thanos-querier-8f4c5c746-xv2gj from openshift-monitoring started at 2020-09-21 16:14:05 +0000 UTC (4 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container oauth-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: 	Container thanos-querier ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: node-ca-ql97q from openshift-image-registry started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container node-ca ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: tuned-5n2xm from openshift-cluster-node-tuning-operator started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container tuned ready: true, restart count 0
-Sep 21 16:24:41.242: INFO: packageserver-65fd8fc8fc-qlpmv from openshift-operator-lifecycle-manager started at 2020-09-21 16:14:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:24:41.242: INFO: 	Container packageserver ready: true, restart count 0
-[It] validates that NodeSelector is respected if not matching  [Conformance]
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
+[It] deployment should support rollover [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Trying to schedule Pod with nonempty NodeSelector.
-STEP: Considering event: 
-Type = [Warning], Name = [restricted-pod.1636d92798931012], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match node selector.]
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+Oct 26 16:39:30.872: INFO: Pod name rollover-pod: Found 0 pods out of 1
+Oct 26 16:39:35.892: INFO: Pod name rollover-pod: Found 1 pods out of 1
+STEP: ensuring each pod is running
+Oct 26 16:39:35.893: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready
+Oct 26 16:39:37.910: INFO: Creating deployment "test-rollover-deployment"
+Oct 26 16:39:37.946: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations
+Oct 26 16:39:39.981: INFO: Check revision of new replica set for deployment "test-rollover-deployment"
+Oct 26 16:39:40.013: INFO: Ensure that both replica sets have 1 created replica
+Oct 26 16:39:40.051: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update
+Oct 26 16:39:40.092: INFO: Updating deployment test-rollover-deployment
+Oct 26 16:39:40.092: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller
+Oct 26 16:39:42.126: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2
+Oct 26 16:39:42.157: INFO: Make sure deployment "test-rollover-deployment" is complete
+Oct 26 16:39:42.199: INFO: all replica sets need to contain the pod-template-hash label
+Oct 26 16:39:42.199: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327178, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327178, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327180, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327177, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 16:39:44.246: INFO: all replica sets need to contain the pod-template-hash label
+Oct 26 16:39:44.246: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327178, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327178, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327182, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327177, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 16:39:46.233: INFO: all replica sets need to contain the pod-template-hash label
+Oct 26 16:39:46.233: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327178, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327178, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327182, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327177, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 16:39:48.233: INFO: all replica sets need to contain the pod-template-hash label
+Oct 26 16:39:48.233: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327178, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327178, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327182, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327177, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 16:39:50.234: INFO: all replica sets need to contain the pod-template-hash label
+Oct 26 16:39:50.234: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327178, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327178, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327182, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327177, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 16:39:52.239: INFO: all replica sets need to contain the pod-template-hash label
+Oct 26 16:39:52.239: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327178, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327178, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327182, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327177, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 16:39:54.245: INFO: 
+Oct 26 16:39:54.245: INFO: Ensure that both old replica sets have no replicas
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
+Oct 26 16:39:54.289: INFO: Deployment "test-rollover-deployment":
+&Deployment{ObjectMeta:{test-rollover-deployment  deployment-1217 /apis/apps/v1/namespaces/deployment-1217/deployments/test-rollover-deployment 61d8814e-11ba-4cef-9072-7f43bf01af26 68821 2 2020-10-26 16:39:37 +0000 UTC   map[name:rollover-pod] map[deployment.kubernetes.io/revision:2] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00d550df8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-10-26 16:39:38 +0000 UTC,LastTransitionTime:2020-10-26 16:39:38 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rollover-deployment-574d6dfbff" has successfully progressed.,LastUpdateTime:2020-10-26 16:39:52 +0000 UTC,LastTransitionTime:2020-10-26 16:39:37 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},}
+
+Oct 26 16:39:54.305: INFO: New ReplicaSet "test-rollover-deployment-574d6dfbff" of Deployment "test-rollover-deployment":
+&ReplicaSet{ObjectMeta:{test-rollover-deployment-574d6dfbff  deployment-1217 /apis/apps/v1/namespaces/deployment-1217/replicasets/test-rollover-deployment-574d6dfbff 11d32c8a-0699-4905-a938-35b5215ceca7 68810 2 2020-10-26 16:39:40 +0000 UTC   map[name:rollover-pod pod-template-hash:574d6dfbff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-rollover-deployment 61d8814e-11ba-4cef-9072-7f43bf01af26 0xc00d551267 0xc00d551268}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 574d6dfbff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod pod-template-hash:574d6dfbff] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00d5512d8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},}
+Oct 26 16:39:54.305: INFO: All old ReplicaSets of Deployment "test-rollover-deployment":
+Oct 26 16:39:54.305: INFO: &ReplicaSet{ObjectMeta:{test-rollover-controller  deployment-1217 /apis/apps/v1/namespaces/deployment-1217/replicasets/test-rollover-controller 2438d2b4-9b96-45d4-a6ed-f042448331bb 68819 2 2020-10-26 16:39:30 +0000 UTC   map[name:rollover-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2] [{apps/v1 Deployment test-rollover-deployment 61d8814e-11ba-4cef-9072-7f43bf01af26 0xc00d551197 0xc00d551198}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod pod:httpd] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc00d5511f8  ClusterFirst map[]     false false false  PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
+Oct 26 16:39:54.306: INFO: &ReplicaSet{ObjectMeta:{test-rollover-deployment-f6c94f66c  deployment-1217 /apis/apps/v1/namespaces/deployment-1217/replicasets/test-rollover-deployment-f6c94f66c a27ca52c-3d85-4663-8c8a-8286b84a2528 68726 2 2020-10-26 16:39:37 +0000 UTC   map[name:rollover-pod pod-template-hash:f6c94f66c] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-rollover-deployment 61d8814e-11ba-4cef-9072-7f43bf01af26 0xc00d551340 0xc00d551341}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: f6c94f66c,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod pod-template-hash:f6c94f66c] map[] [] []  []} {[] [] [{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00d5513b8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
+Oct 26 16:39:54.321: INFO: Pod "test-rollover-deployment-574d6dfbff-c8rhf" is available:
+&Pod{ObjectMeta:{test-rollover-deployment-574d6dfbff-c8rhf test-rollover-deployment-574d6dfbff- deployment-1217 /api/v1/namespaces/deployment-1217/pods/test-rollover-deployment-574d6dfbff-c8rhf 87f297c2-d282-4376-a5a3-88732dd4c027 68752 0 2020-10-26 16:39:40 +0000 UTC   map[name:rollover-pod pod-template-hash:574d6dfbff] map[cni.projectcalico.org/podIP:172.30.194.67/32 cni.projectcalico.org/podIPs:172.30.194.67/32 k8s.v1.cni.cncf.io/networks-status:[{
+    "name": "k8s-pod-network",
+    "ips": [
+        "172.30.194.67"
+    ],
+    "dns": {}
+}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet test-rollover-deployment-574d6dfbff 11d32c8a-0699-4905-a938-35b5215ceca7 0xc00d53d3f7 0xc00d53d3f8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-6rwsl,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-6rwsl,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-6rwsl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.74,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-bl2wm,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:39:40 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:39:42 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:39:42 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 16:39:40 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.74,PodIP:172.30.194.67,StartTime:2020-10-26 16:39:40 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 16:39:41 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:gcr.io/kubernetes-e2e-test-images/agnhost@sha256:02b9e10ea1f7439a03d9bc3ee16d984172ca7eac1818792ad4d721c5c8f72ff4,ContainerID:cri-o://b61d176ac2bdfd47eac748ef9f028cb901d2d46ae5d8013a7b0c77d549c36670,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.194.67,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+[AfterEach] [sig-apps] Deployment
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:24:42.692: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "sched-pred-5405" for this suite.
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
-•{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching  [Conformance]","total":280,"completed":126,"skipped":2077,"failed":0}
-SSSSSSSSSSSSSSSS
+Oct 26 16:39:54.321: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-1217" for this suite.
+
+• [SLOW TEST:23.687 seconds]
+[sig-apps] Deployment
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  deployment should support rollover [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with secret pod [LinuxOnly] [Conformance]
+{"msg":"PASSED [sig-apps] Deployment deployment should support rollover [Conformance]","total":280,"completed":132,"skipped":1987,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client Kubectl api-versions 
+  should check if v1 is in available api versions  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Subpath
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:24:42.774: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename subpath
+Oct 26 16:39:54.367: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubectl
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
-STEP: Setting up data
-[It] should support subpaths with secret pod [LinuxOnly] [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
+[It] should check if v1 is in available api versions  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod pod-subpath-test-secret-jrrt
-STEP: Creating a pod to test atomic-volume-subpath
-Sep 21 16:24:43.172: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-jrrt" in namespace "subpath-9844" to be "success or failure"
-Sep 21 16:24:43.197: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Pending", Reason="", readiness=false. Elapsed: 24.663612ms
-Sep 21 16:24:45.229: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Pending", Reason="", readiness=false. Elapsed: 2.05711733s
-Sep 21 16:24:47.254: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Running", Reason="", readiness=true. Elapsed: 4.081710018s
-Sep 21 16:24:49.280: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Running", Reason="", readiness=true. Elapsed: 6.107375096s
-Sep 21 16:24:51.308: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Running", Reason="", readiness=true. Elapsed: 8.135426568s
-Sep 21 16:24:53.333: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Running", Reason="", readiness=true. Elapsed: 10.160959104s
-Sep 21 16:24:55.366: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Running", Reason="", readiness=true. Elapsed: 12.194175339s
-Sep 21 16:24:57.390: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Running", Reason="", readiness=true. Elapsed: 14.217753635s
-Sep 21 16:24:59.422: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Running", Reason="", readiness=true. Elapsed: 16.249305417s
-Sep 21 16:25:01.457: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Running", Reason="", readiness=true. Elapsed: 18.285032433s
-Sep 21 16:25:03.483: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Running", Reason="", readiness=true. Elapsed: 20.31099771s
-Sep 21 16:25:05.720: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Running", Reason="", readiness=true. Elapsed: 22.547569988s
-Sep 21 16:25:07.746: INFO: Pod "pod-subpath-test-secret-jrrt": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.57341631s
-STEP: Saw pod success
-Sep 21 16:25:07.746: INFO: Pod "pod-subpath-test-secret-jrrt" satisfied condition "success or failure"
-Sep 21 16:25:07.988: INFO: Trying to get logs from node 10.241.51.147 pod pod-subpath-test-secret-jrrt container test-container-subpath-secret-jrrt: 
-STEP: delete the pod
-Sep 21 16:25:08.153: INFO: Waiting for pod pod-subpath-test-secret-jrrt to disappear
-Sep 21 16:25:08.177: INFO: Pod pod-subpath-test-secret-jrrt no longer exists
-STEP: Deleting pod pod-subpath-test-secret-jrrt
-Sep 21 16:25:08.177: INFO: Deleting pod "pod-subpath-test-secret-jrrt" in namespace "subpath-9844"
-[AfterEach] [sig-storage] Subpath
+STEP: validating api versions
+Oct 26 16:39:54.524: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 api-versions'
+Oct 26 16:39:54.701: INFO: stderr: ""
+Oct 26 16:39:54.701: INFO: stdout: "admissionregistration.k8s.io/v1\nadmissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps.openshift.io/v1\napps/v1\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nauthorization.openshift.io/v1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\nbuild.openshift.io/v1\ncertificates.k8s.io/v1beta1\ncloudcredential.openshift.io/v1\nconfig.openshift.io/v1\nconsole.openshift.io/v1\ncoordination.k8s.io/v1\ncoordination.k8s.io/v1beta1\ncrd.projectcalico.org/v1\ndiscovery.k8s.io/v1beta1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nibm.com/v1alpha1\nimage.openshift.io/v1\nimageregistry.operator.openshift.io/v1\ningress.operator.openshift.io/v1\nk8s.cni.cncf.io/v1\nmetal3.io/v1alpha1\nmetrics.k8s.io/v1beta1\nmigration.k8s.io/v1alpha1\nmonitoring.coreos.com/v1\nnetwork.operator.openshift.io/v1\nnetworking.k8s.io/v1\nnetworking.k8s.io/v1beta1\nnode.k8s.io/v1beta1\noauth.openshift.io/v1\noperator.openshift.io/v1\noperator.openshift.io/v1alpha1\noperator.tigera.io/v1\noperators.coreos.com/v1\noperators.coreos.com/v1alpha1\noperators.coreos.com/v1alpha2\noperators.coreos.com/v2\npackages.operators.coreos.com/v1\npolicy/v1beta1\nproject.openshift.io/v1\nquota.openshift.io/v1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nroute.openshift.io/v1\nsamples.operator.openshift.io/v1\nscheduling.k8s.io/v1\nscheduling.k8s.io/v1beta1\nsecurity.openshift.io/v1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\ntemplate.openshift.io/v1\ntuned.openshift.io/v1\nuser.openshift.io/v1\nv1\nwhereabouts.cni.cncf.io/v1alpha1\n"
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:25:08.202: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "subpath-9844" for this suite.
-
-• [SLOW TEST:25.502 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
-  Atomic writer volumes
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
-    should support subpaths with secret pod [LinuxOnly] [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance]","total":280,"completed":127,"skipped":2093,"failed":0}
-SSSSSSSS
+Oct 26 16:39:54.701: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-8834" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions  [Conformance]","total":280,"completed":133,"skipped":2010,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  works for CRD preserving unknown fields at the schema root [Conformance]
+  removes definition from spec when one version gets changed to not be served [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:25:08.276: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 16:39:54.739: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename crd-publish-openapi
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] works for CRD preserving unknown fields at the schema root [Conformance]
+[It] removes definition from spec when one version gets changed to not be served [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:25:08.635: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: client-side validation (kubectl create and apply) allows request with any unknown properties
-Sep 21 16:25:17.246: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-2474 create -f -'
-Sep 21 16:25:18.090: INFO: stderr: ""
-Sep 21 16:25:18.090: INFO: stdout: "e2e-test-crd-publish-openapi-7732-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n"
-Sep 21 16:25:18.091: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-2474 delete e2e-test-crd-publish-openapi-7732-crds test-cr'
-Sep 21 16:25:18.405: INFO: stderr: ""
-Sep 21 16:25:18.405: INFO: stdout: "e2e-test-crd-publish-openapi-7732-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n"
-Sep 21 16:25:18.405: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-2474 apply -f -'
-Sep 21 16:25:18.754: INFO: stderr: ""
-Sep 21 16:25:18.754: INFO: stdout: "e2e-test-crd-publish-openapi-7732-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n"
-Sep 21 16:25:18.754: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-2474 delete e2e-test-crd-publish-openapi-7732-crds test-cr'
-Sep 21 16:25:19.059: INFO: stderr: ""
-Sep 21 16:25:19.059: INFO: stdout: "e2e-test-crd-publish-openapi-7732-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n"
-STEP: kubectl explain works to explain CR
-Sep 21 16:25:19.059: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 explain e2e-test-crd-publish-openapi-7732-crds'
-Sep 21 16:25:19.330: INFO: stderr: ""
-Sep 21 16:25:19.330: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-7732-crd\nVERSION:  crd-publish-openapi-test-unknown-at-root.example.com/v1\n\nDESCRIPTION:\n     \n"
+STEP: set up a multi version CRD
+Oct 26 16:39:54.901: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: mark a version not serverd
+STEP: check the unserved version gets removed
+STEP: check the other version is not changed
 [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:25:26.835: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-2474" for this suite.
+Oct 26 16:40:38.879: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-8196" for this suite.
 
-• [SLOW TEST:18.635 seconds]
+• [SLOW TEST:44.179 seconds]
 [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for CRD preserving unknown fields at the schema root [Conformance]
+  removes definition from spec when one version gets changed to not be served [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance]","total":280,"completed":128,"skipped":2101,"failed":0}
-SSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance]","total":280,"completed":134,"skipped":2035,"failed":0}
+SSSSSSSSS
 ------------------------------
-[k8s.io] Probing container 
-  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Probing container
+[BeforeEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:25:26.911: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-probe
+Oct 26 16:40:38.918: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:25:53.484: INFO: Container started at 2020-09-21 16:25:28 +0000 UTC, pod became ready at 2020-09-21 16:25:52 +0000 UTC
-[AfterEach] [k8s.io] Probing container
+STEP: Creating configMap with name projected-configmap-test-volume-8dc47252-0c8e-4e9f-9ba3-c39a72ba55bd
+STEP: Creating a pod to test consume configMaps
+Oct 26 16:40:39.171: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-b72e2b29-c978-43eb-adb3-a83c1f91052c" in namespace "projected-5031" to be "success or failure"
+Oct 26 16:40:39.187: INFO: Pod "pod-projected-configmaps-b72e2b29-c978-43eb-adb3-a83c1f91052c": Phase="Pending", Reason="", readiness=false. Elapsed: 15.513845ms
+Oct 26 16:40:41.203: INFO: Pod "pod-projected-configmaps-b72e2b29-c978-43eb-adb3-a83c1f91052c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.032203599s
+STEP: Saw pod success
+Oct 26 16:40:41.203: INFO: Pod "pod-projected-configmaps-b72e2b29-c978-43eb-adb3-a83c1f91052c" satisfied condition "success or failure"
+Oct 26 16:40:41.220: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-configmaps-b72e2b29-c978-43eb-adb3-a83c1f91052c container projected-configmap-volume-test: 
+STEP: delete the pod
+Oct 26 16:40:41.302: INFO: Waiting for pod pod-projected-configmaps-b72e2b29-c978-43eb-adb3-a83c1f91052c to disappear
+Oct 26 16:40:41.317: INFO: Pod pod-projected-configmaps-b72e2b29-c978-43eb-adb3-a83c1f91052c no longer exists
+[AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:25:53.485: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-7074" for this suite.
-
-• [SLOW TEST:26.643 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [k8s.io] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]","total":280,"completed":129,"skipped":2112,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:40:41.317: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-5031" for this suite.
+•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","total":280,"completed":135,"skipped":2044,"failed":0}
+SSSSSSSS
 ------------------------------
-[k8s.io] KubeletManagedEtcHosts 
-  should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
+[k8s.io] Docker Containers 
+  should use the image defaults if command and args are blank [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] KubeletManagedEtcHosts
+[BeforeEach] [k8s.io] Docker Containers
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:25:53.555: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts
+Oct 26 16:40:41.354: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename containers
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
+[It] should use the image defaults if command and args are blank [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Setting up the test
-STEP: Creating hostNetwork=false pod
-STEP: Creating hostNetwork=true pod
-STEP: Running the test
-STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false
-Sep 21 16:26:00.538: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7080 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:26:00.538: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:26:00.779: INFO: Exec stderr: ""
-Sep 21 16:26:00.779: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7080 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:26:00.779: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:26:01.056: INFO: Exec stderr: ""
-Sep 21 16:26:01.056: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7080 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:26:01.056: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:26:01.442: INFO: Exec stderr: ""
-Sep 21 16:26:01.442: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7080 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:26:01.442: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:26:02.056: INFO: Exec stderr: ""
-STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount
-Sep 21 16:26:02.056: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7080 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:26:02.056: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:26:02.367: INFO: Exec stderr: ""
-Sep 21 16:26:02.367: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7080 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:26:02.367: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:26:02.582: INFO: Exec stderr: ""
-STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true
-Sep 21 16:26:02.582: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7080 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:26:02.582: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:26:02.831: INFO: Exec stderr: ""
-Sep 21 16:26:02.831: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7080 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:26:02.831: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:26:03.111: INFO: Exec stderr: ""
-Sep 21 16:26:03.111: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7080 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:26:03.111: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:26:03.432: INFO: Exec stderr: ""
-Sep 21 16:26:03.432: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7080 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:26:03.432: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:26:03.640: INFO: Exec stderr: ""
-[AfterEach] [k8s.io] KubeletManagedEtcHosts
+[AfterEach] [k8s.io] Docker Containers
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:26:03.640: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-kubelet-etc-hosts-7080" for this suite.
-
-• [SLOW TEST:10.172 seconds]
-[k8s.io] KubeletManagedEtcHosts
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Oct 26 16:40:45.623: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "containers-4592" for this suite.
+•{"msg":"PASSED [k8s.io] Docker Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance]","total":280,"completed":136,"skipped":2052,"failed":0}
+SSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-{"msg":"PASSED [k8s.io] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":130,"skipped":2138,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[sig-storage] Projected downwardAPI 
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 16:40:45.666: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Oct 26 16:40:45.878: INFO: Waiting up to 5m0s for pod "downwardapi-volume-ac64ab2d-0033-4d89-81fd-8b1dba890c10" in namespace "projected-9105" to be "success or failure"
+Oct 26 16:40:45.901: INFO: Pod "downwardapi-volume-ac64ab2d-0033-4d89-81fd-8b1dba890c10": Phase="Pending", Reason="", readiness=false. Elapsed: 23.757605ms
+Oct 26 16:40:47.918: INFO: Pod "downwardapi-volume-ac64ab2d-0033-4d89-81fd-8b1dba890c10": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.040033777s
+STEP: Saw pod success
+Oct 26 16:40:47.918: INFO: Pod "downwardapi-volume-ac64ab2d-0033-4d89-81fd-8b1dba890c10" satisfied condition "success or failure"
+Oct 26 16:40:47.936: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-ac64ab2d-0033-4d89-81fd-8b1dba890c10 container client-container: 
+STEP: delete the pod
+Oct 26 16:40:48.030: INFO: Waiting for pod downwardapi-volume-ac64ab2d-0033-4d89-81fd-8b1dba890c10 to disappear
+Oct 26 16:40:48.052: INFO: Pod downwardapi-volume-ac64ab2d-0033-4d89-81fd-8b1dba890c10 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 16:40:48.052: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9105" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance]","total":280,"completed":137,"skipped":2073,"failed":0}
+SSSSS
 ------------------------------
-[k8s.io] Probing container 
-  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  should perform canary updates and phased rolling updates of template modifications [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Probing container
+[BeforeEach] [sig-apps] StatefulSet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:26:03.727: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-probe
+Oct 26 16:40:48.094: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename statefulset
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
+STEP: Creating service test in namespace statefulset-7442
+[It] should perform canary updates and phased rolling updates of template modifications [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod busybox-6c0ec7d9-d004-4ba2-8602-db8dc6fe76a6 in namespace container-probe-2071
-Sep 21 16:26:06.195: INFO: Started pod busybox-6c0ec7d9-d004-4ba2-8602-db8dc6fe76a6 in namespace container-probe-2071
-STEP: checking the pod's current state and verifying that restartCount is present
-Sep 21 16:26:06.222: INFO: Initial restart count of pod busybox-6c0ec7d9-d004-4ba2-8602-db8dc6fe76a6 is 0
-STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
+STEP: Creating a new StatefulSet
+Oct 26 16:40:48.312: INFO: Found 0 stateful pods, waiting for 3
+Oct 26 16:40:58.331: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
+Oct 26 16:40:58.331: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
+Oct 26 16:40:58.331: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Updating stateful set template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine
+Oct 26 16:40:58.450: INFO: Updating stateful set ss2
+STEP: Creating a new revision
+STEP: Not applying an update when the partition is greater than the number of replicas
+STEP: Performing a canary update
+Oct 26 16:41:08.569: INFO: Updating stateful set ss2
+Oct 26 16:41:08.603: INFO: Waiting for Pod statefulset-7442/ss2-2 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
+STEP: Restoring Pods to the correct revision when they are deleted
+Oct 26 16:41:18.768: INFO: Found 2 stateful pods, waiting for 3
+Oct 26 16:41:28.786: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
+Oct 26 16:41:28.786: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
+Oct 26 16:41:28.786: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Performing a phased rolling update
+Oct 26 16:41:28.869: INFO: Updating stateful set ss2
+Oct 26 16:41:28.906: INFO: Waiting for Pod statefulset-7442/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
+Oct 26 16:41:38.942: INFO: Waiting for Pod statefulset-7442/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
+Oct 26 16:41:48.992: INFO: Updating stateful set ss2
+Oct 26 16:41:49.033: INFO: Waiting for StatefulSet statefulset-7442/ss2 to complete update
+Oct 26 16:41:49.033: INFO: Waiting for Pod statefulset-7442/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
+Oct 26 16:41:59.069: INFO: Deleting all statefulset in ns statefulset-7442
+Oct 26 16:41:59.089: INFO: Scaling statefulset ss2 to 0
+Oct 26 16:42:29.170: INFO: Waiting for statefulset status.replicas updated to 0
+Oct 26 16:42:29.203: INFO: Deleting statefulset ss2
+[AfterEach] [sig-apps] StatefulSet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:30:07.778: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-2071" for this suite.
+Oct 26 16:42:29.289: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-7442" for this suite.
 
-• [SLOW TEST:244.119 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:101.235 seconds]
+[sig-apps] StatefulSet
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+    should perform canary updates and phased rolling updates of template modifications [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":280,"completed":131,"skipped":2170,"failed":0}
+{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance]","total":280,"completed":138,"skipped":2078,"failed":0}
 SSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should mutate custom resource with different stored version [Conformance]
+[sig-api-machinery] Watchers 
+  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+[BeforeEach] [sig-api-machinery] Watchers
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:30:07.850: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename webhook
+Oct 26 16:42:29.332: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename watch
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Sep 21 16:30:09.085: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Sep 21 16:30:11.157: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736302609, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736302609, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736302609, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736302609, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Sep 21 16:30:14.289: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should mutate custom resource with different stored version [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:30:14.316: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Registering the mutating webhook for custom resource e2e-test-webhook-9986-crds.webhook.example.com via the AdmissionRegistration API
-STEP: Creating a custom resource while v1 is storage version
-STEP: Patching Custom Resource Definition to set v2 as storage
-STEP: Patching the custom resource while v2 is storage version
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:30:15.878: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-6369" for this suite.
-STEP: Destroying namespace "webhook-6369-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
-
-• [SLOW TEST:8.717 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should mutate custom resource with different stored version [Conformance]
+[It] should be able to restart watching from the last resource version observed by the previous watch [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating a watch on configmaps
+STEP: creating a new configmap
+STEP: modifying the configmap once
+STEP: closing the watch once it receives two notifications
+Oct 26 16:42:29.541: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed  watch-7736 /api/v1/namespaces/watch-7736/configmaps/e2e-watch-test-watch-closed 77ea94a4-e093-4cdb-8369-1833cf49dd3f 70270 0 2020-10-26 16:42:29 +0000 UTC   map[watch-this-configmap:watch-closed-and-restarted] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Oct 26 16:42:29.541: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed  watch-7736 /api/v1/namespaces/watch-7736/configmaps/e2e-watch-test-watch-closed 77ea94a4-e093-4cdb-8369-1833cf49dd3f 70273 0 2020-10-26 16:42:29 +0000 UTC   map[watch-this-configmap:watch-closed-and-restarted] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+STEP: modifying the configmap a second time, while the watch is closed
+STEP: creating a new watch on configmaps from the last resource version observed by the first watch
+STEP: deleting the configmap
+STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed
+Oct 26 16:42:29.608: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed  watch-7736 /api/v1/namespaces/watch-7736/configmaps/e2e-watch-test-watch-closed 77ea94a4-e093-4cdb-8369-1833cf49dd3f 70275 0 2020-10-26 16:42:29 +0000 UTC   map[watch-this-configmap:watch-closed-and-restarted] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Oct 26 16:42:29.609: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed  watch-7736 /api/v1/namespaces/watch-7736/configmaps/e2e-watch-test-watch-closed 77ea94a4-e093-4cdb-8369-1833cf49dd3f 70278 0 2020-10-26 16:42:29 +0000 UTC   map[watch-this-configmap:watch-closed-and-restarted] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 16:42:29.609: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-7736" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance]","total":280,"completed":139,"skipped":2095,"failed":0}
+SSSSSSSSSSSS
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance]","total":280,"completed":132,"skipped":2187,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]
+[sig-api-machinery] Namespaces [Serial] 
+  should ensure that all pods are removed when a namespace is deleted [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+[BeforeEach] [sig-api-machinery] Namespaces [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:30:16.569: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename sched-pred
+Oct 26 16:42:29.649: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename namespaces
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
-Sep 21 16:30:17.098: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-Sep 21 16:30:17.200: INFO: Waiting for terminating namespaces to be deleted...
-Sep 21 16:30:17.268: INFO: 
-Logging pods the kubelet thinks is on node 10.241.51.147 before test
-Sep 21 16:30:32.340: INFO: Unable to retrieve kubelet pods for node 10.241.51.147: the server is currently unable to handle the request (get nodes 10.241.51.147:10250)
-Sep 21 16:30:32.340: INFO: 
-Logging pods the kubelet thinks is on node 10.241.51.149 before test
-Sep 21 16:30:32.590: INFO: ibm-file-plugin-66fdcfc4c9-tm6qk from kube-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.590: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
-Sep 21 16:30:32.590: INFO: packageserver-65fd8fc8fc-2mnzj from openshift-operator-lifecycle-manager started at 2020-09-21 16:14:19 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.590: INFO: 	Container packageserver ready: true, restart count 0
-Sep 21 16:30:32.590: INFO: cluster-samples-operator-5cfb985bf-n7nxd from openshift-cluster-samples-operator started at 2020-09-21 14:09:54 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.590: INFO: 	Container cluster-samples-operator ready: true, restart count 0
-Sep 21 16:30:32.590: INFO: 	Container cluster-samples-operator-watch ready: true, restart count 0
-Sep 21 16:30:32.590: INFO: redhat-marketplace-7b9bcdf96b-p4khw from openshift-marketplace started at 2020-09-21 14:09:34 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.590: INFO: 	Container redhat-marketplace ready: true, restart count 0
-Sep 21 16:30:32.590: INFO: openshift-service-catalog-apiserver-operator-6988647bc8-t9z7w from openshift-service-catalog-apiserver-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.590: INFO: 	Container operator ready: true, restart count 1
-Sep 21 16:30:32.590: INFO: ibmcloud-block-storage-plugin-68d5c65db9-7q8bv from kube-system started at 2020-09-21 14:07:36 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.590: INFO: 	Container ibmcloud-block-storage-plugin-container ready: true, restart count 0
-Sep 21 16:30:32.590: INFO: openshift-state-metrics-678b6c786f-698gn from openshift-monitoring started at 2020-09-21 14:08:22 +0000 UTC (3 container statuses recorded)
-Sep 21 16:30:32.590: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container openshift-state-metrics ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: prometheus-adapter-5f78ddc679-vb486 from openshift-monitoring started at 2020-09-21 14:08:28 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container prometheus-adapter ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: multus-hprgb from openshift-multus started at 2020-09-21 14:06:32 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container kube-multus ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: calico-typha-6fc9ff8b5f-gzhgw from calico-system started at 2020-09-21 14:07:18 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container calico-typha ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: kube-state-metrics-7455b6c8-455w7 from openshift-monitoring started at 2020-09-21 14:08:22 +0000 UTC (3 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container kube-state-metrics ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: network-operator-f456cfdbb-4m6pj from openshift-network-operator started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container network-operator ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: cluster-monitoring-operator-75f4ddd85f-579q6 from openshift-monitoring started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container cluster-monitoring-operator ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: kube-storage-version-migrator-operator-86d8fc6986-vl4fl from openshift-kube-storage-version-migrator-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container kube-storage-version-migrator-operator ready: true, restart count 1
-Sep 21 16:30:32.591: INFO: calico-kube-controllers-79d75767dd-xkvb6 from calico-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container calico-kube-controllers ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: multus-admission-controller-gkchr from openshift-multus started at 2020-09-21 14:07:34 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container multus-admission-controller ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: prometheus-adapter-5f78ddc679-9hj7w from openshift-monitoring started at 2020-09-21 14:08:28 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container prometheus-adapter ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: community-operators-6566687cd-smnd2 from openshift-marketplace started at 2020-09-21 14:09:35 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container community-operators ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: ibm-storage-watcher-7956f954f-26xx6 from kube-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: service-ca-bd8866b9b-qzggv from openshift-service-ca started at 2020-09-21 14:07:50 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container service-ca-controller ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: certified-operators-5cfd68758d-q6c6n from openshift-marketplace started at 2020-09-21 16:14:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container certified-operators ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: ibmcloud-block-storage-driver-pldbm from kube-system started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: dns-operator-7886f4f4f4-rk7dv from openshift-dns-operator started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container dns-operator ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: prometheus-k8s-1 from openshift-monitoring started at 2020-09-21 16:14:17 +0000 UTC (7 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container prometheus ready: true, restart count 1
-Sep 21 16:30:32.591: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container prometheus-proxy ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container thanos-sidecar ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: catalog-operator-69d4f5c59f-lqjgz from openshift-operator-lifecycle-manager started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container catalog-operator ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-lgbbw from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container systemd-logs ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: openshift-kube-proxy-9svk2 from openshift-kube-proxy started at 2020-09-21 14:06:38 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container kube-proxy ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: service-ca-operator-9f84dfdb5-wdhvg from openshift-service-ca-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container operator ready: true, restart count 1
-Sep 21 16:30:32.591: INFO: ingress-operator-57546dfcd9-2sg5v from openshift-ingress-operator started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container ingress-operator ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: router-default-84dfcdc9b5-tk98j from openshift-ingress started at 2020-09-21 14:09:06 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container router ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: dns-default-958xc from openshift-dns started at 2020-09-21 14:09:05 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container dns ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container dns-node-resolver ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: alertmanager-main-0 from openshift-monitoring started at 2020-09-21 14:14:12 +0000 UTC (3 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container alertmanager ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container alertmanager-proxy ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container config-reloader ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: downloads-7989988989-9rhh4 from openshift-console started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container download-server ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: tuned-gpfhj from openshift-cluster-node-tuning-operator started at 2020-09-21 14:08:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container tuned ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: node-exporter-jnrzm from openshift-monitoring started at 2020-09-21 14:08:24 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container node-exporter ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-44pnn from ibm-system started at 2020-09-21 14:12:56 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container ibm-cloud-provider-ip-169-48-211-85 ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: ibm-keepalived-watcher-r4tn7 from kube-system started at 2020-09-21 14:06:04 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: marketplace-operator-56c6694c8f-x9rcq from openshift-marketplace started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container marketplace-operator ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: downloads-7989988989-cjvjw from openshift-console started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container download-server ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: migrator-6c94c8c775-vxsnp from openshift-kube-storage-version-migrator started at 2020-09-21 14:07:55 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container migrator ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: tigera-operator-679798d94d-hdlvj from tigera-operator started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container tigera-operator ready: true, restart count 2
-Sep 21 16:30:32.591: INFO: cluster-image-registry-operator-695bf78ffc-95l5m from openshift-image-registry started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container cluster-image-registry-operator ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container cluster-image-registry-operator-watch ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: console-operator-58fd84b95c-dc6vq from openshift-console-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container console-operator ready: true, restart count 1
-Sep 21 16:30:32.591: INFO: cluster-storage-operator-5bcf8d9f8f-dls4n from openshift-cluster-storage-operator started at 2020-09-21 14:07:36 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container cluster-storage-operator ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: image-registry-775dd777d7-5qtdd from openshift-image-registry started at 2020-09-21 14:11:01 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container registry ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: ibm-master-proxy-static-10.241.51.149 from kube-system started at 2020-09-21 14:06:02 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 	Container pause ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: cluster-node-tuning-operator-bdbf98554-22kv8 from openshift-cluster-node-tuning-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container cluster-node-tuning-operator ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: redhat-operators-54dcdd9697-pcwgz from openshift-marketplace started at 2020-09-21 14:09:34 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container redhat-operators ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: registry-pvc-permissions-rlfxm from openshift-image-registry started at 2020-09-21 14:11:01 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container pvc-permissions ready: false, restart count 0
-Sep 21 16:30:32.591: INFO: console-8549dd6c4f-8vk99 from openshift-console started at 2020-09-21 16:14:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container console ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: node-ca-bxmsf from openshift-image-registry started at 2020-09-21 14:09:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container node-ca ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: prometheus-operator-8d96bcc68-ct2jw from openshift-monitoring started at 2020-09-21 16:14:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container prometheus-operator ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: calico-node-9gqht from calico-system started at 2020-09-21 14:07:18 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container calico-node ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: openshift-service-catalog-controller-manager-operator-b9ccczlcs from openshift-service-catalog-controller-manager-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container operator ready: true, restart count 1
-Sep 21 16:30:32.591: INFO: olm-operator-b5f47bf98-8d8cw from openshift-operator-lifecycle-manager started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.591: INFO: 	Container olm-operator ready: true, restart count 0
-Sep 21 16:30:32.591: INFO: 
-Logging pods the kubelet thinks is on node 10.241.51.150 before test
-Sep 21 16:30:32.753: INFO: packageserver-65fd8fc8fc-qlpmv from openshift-operator-lifecycle-manager started at 2020-09-21 16:14:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.754: INFO: 	Container packageserver ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: node-ca-ql97q from openshift-image-registry started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.754: INFO: 	Container node-ca ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: tuned-5n2xm from openshift-cluster-node-tuning-operator started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.754: INFO: 	Container tuned ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-xzz27 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.754: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: 	Container systemd-logs ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: ibm-keepalived-watcher-bntgc from kube-system started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.754: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: alertmanager-main-2 from openshift-monitoring started at 2020-09-21 16:14:17 +0000 UTC (3 container statuses recorded)
-Sep 21 16:30:32.754: INFO: 	Container alertmanager ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: 	Container alertmanager-proxy ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: 	Container config-reloader ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: console-8549dd6c4f-b5jnn from openshift-console started at 2020-09-21 14:18:13 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.754: INFO: 	Container console ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: node-exporter-flb5l from openshift-monitoring started at 2020-09-21 14:12:21 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.754: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: 	Container node-exporter ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: calico-node-dn4jv from calico-system started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.754: INFO: 	Container calico-node ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: grafana-c745ff78c-9pgfx from openshift-monitoring started at 2020-09-21 14:13:42 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.754: INFO: 	Container grafana ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: 	Container grafana-proxy ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-ghw5q from ibm-system started at 2020-09-21 16:07:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.754: INFO: 	Container ibm-cloud-provider-ip-169-48-211-85 ready: true, restart count 0
-Sep 21 16:30:32.754: INFO: dns-default-z8v9j from openshift-dns started at 2020-09-21 14:12:21 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.755: INFO: 	Container dns ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: 	Container dns-node-resolver ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: calico-typha-6fc9ff8b5f-cbhcz from calico-system started at 2020-09-21 14:13:14 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.755: INFO: 	Container calico-typha ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: multus-admission-controller-snglx from openshift-multus started at 2020-09-21 14:13:31 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.755: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: 	Container multus-admission-controller ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: alertmanager-main-1 from openshift-monitoring started at 2020-09-21 14:14:02 +0000 UTC (3 container statuses recorded)
-Sep 21 16:30:32.755: INFO: 	Container alertmanager ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: 	Container alertmanager-proxy ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: 	Container config-reloader ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: router-default-84dfcdc9b5-mt4b7 from openshift-ingress started at 2020-09-21 16:07:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.755: INFO: 	Container router ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: multus-n5qc8 from openshift-multus started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.755: INFO: 	Container kube-multus ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: openshift-kube-proxy-9bv2v from openshift-kube-proxy started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.755: INFO: 	Container kube-proxy ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: prometheus-k8s-0 from openshift-monitoring started at 2020-09-21 14:15:10 +0000 UTC (7 container statuses recorded)
-Sep 21 16:30:32.755: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: 	Container prometheus ready: true, restart count 1
-Sep 21 16:30:32.755: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: 	Container prometheus-proxy ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: 	Container thanos-sidecar ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: ibm-master-proxy-static-10.241.51.150 from kube-system started at 2020-09-21 14:12:19 +0000 UTC (2 container statuses recorded)
-Sep 21 16:30:32.755: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: 	Container pause ready: true, restart count 0
-Sep 21 16:30:32.755: INFO: telemeter-client-68bdb7f795-nnhfl from openshift-monitoring started at 2020-09-21 14:13:34 +0000 UTC (3 container statuses recorded)
-Sep 21 16:30:32.756: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:30:32.756: INFO: 	Container reload ready: true, restart count 0
-Sep 21 16:30:32.756: INFO: 	Container telemeter-client ready: true, restart count 0
-Sep 21 16:30:32.756: INFO: vpn-64d8d4987c-lxt95 from kube-system started at 2020-09-21 16:07:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.756: INFO: 	Container vpn ready: true, restart count 0
-Sep 21 16:30:32.756: INFO: thanos-querier-8f4c5c746-xv2gj from openshift-monitoring started at 2020-09-21 16:14:05 +0000 UTC (4 container statuses recorded)
-Sep 21 16:30:32.756: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:30:32.756: INFO: 	Container oauth-proxy ready: true, restart count 0
-Sep 21 16:30:32.756: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:30:32.756: INFO: 	Container thanos-querier ready: true, restart count 0
-Sep 21 16:30:32.756: INFO: ibmcloud-block-storage-driver-jf9s4 from kube-system started at 2020-09-21 14:12:26 +0000 UTC (1 container statuses recorded)
-Sep 21 16:30:32.756: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
-Sep 21 16:30:32.756: INFO: thanos-querier-8f4c5c746-v982g from openshift-monitoring started at 2020-09-21 14:14:42 +0000 UTC (4 container statuses recorded)
-Sep 21 16:30:32.756: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:30:32.756: INFO: 	Container oauth-proxy ready: true, restart count 0
-Sep 21 16:30:32.756: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:30:32.756: INFO: 	Container thanos-querier ready: true, restart count 0
-[It] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]
+[It] should ensure that all pods are removed when a namespace is deleted [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Trying to launch a pod without a label to get a node which can launch it.
-STEP: Explicitly delete pod here to free the resource it takes.
-STEP: Trying to apply a random label on the found node.
-STEP: verifying the node has the label kubernetes.io/e2e-ef4d7acf-1b5c-47a9-bac0-51a96b0d8de9 95
-STEP: Trying to create a pod(pod4) with hostport 54322 and hostIP 0.0.0.0(empty string here) and expect scheduled
-STEP: Trying to create another pod(pod5) with hostport 54322 but hostIP 127.0.0.1 on the node which pod4 resides and expect not scheduled
-STEP: removing the label kubernetes.io/e2e-ef4d7acf-1b5c-47a9-bac0-51a96b0d8de9 off the node 10.241.51.147
-STEP: verifying the node doesn't have the label kubernetes.io/e2e-ef4d7acf-1b5c-47a9-bac0-51a96b0d8de9
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+STEP: Creating a test namespace
+STEP: Waiting for a default service account to be provisioned in namespace
+STEP: Creating a pod in the namespace
+STEP: Waiting for the pod to have running status
+STEP: Deleting the namespace
+STEP: Waiting for the namespace to be removed.
+STEP: Recreating the namespace
+STEP: Verifying there are no pods in the namespace
+[AfterEach] [sig-api-machinery] Namespaces [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:35:39.802: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "sched-pred-3020" for this suite.
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
+Oct 26 16:42:48.267: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "namespaces-170" for this suite.
+STEP: Destroying namespace "nsdeletetest-1192" for this suite.
+Oct 26 16:42:48.323: INFO: Namespace nsdeletetest-1192 was already deleted
+STEP: Destroying namespace "nsdeletetest-4582" for this suite.
 
-• [SLOW TEST:323.298 seconds]
-[sig-scheduling] SchedulerPredicates [Serial]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
-  validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]
+• [SLOW TEST:18.694 seconds]
+[sig-api-machinery] Namespaces [Serial]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should ensure that all pods are removed when a namespace is deleted [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]","total":280,"completed":133,"skipped":2283,"failed":0}
-SSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance]","total":280,"completed":140,"skipped":2107,"failed":0}
+SSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl run deployment 
-  should create a deployment from an image  [Conformance]
+[k8s.io] Container Runtime blackbox test on terminated container 
+  should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [k8s.io] Container Runtime
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:35:39.869: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubectl
+Oct 26 16:42:48.343: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename container-runtime
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[BeforeEach] Kubectl run deployment
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1733
-[It] should create a deployment from an image  [Conformance]
+[It] should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: running the image docker.io/library/httpd:2.4.38-alpine
-Sep 21 16:35:40.125: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 run e2e-test-httpd-deployment --image=docker.io/library/httpd:2.4.38-alpine --generator=deployment/apps.v1 --namespace=kubectl-7353'
-Sep 21 16:35:40.507: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-Sep 21 16:35:40.507: INFO: stdout: "deployment.apps/e2e-test-httpd-deployment created\n"
-STEP: verifying the deployment e2e-test-httpd-deployment was created
-STEP: verifying the pod controlled by deployment e2e-test-httpd-deployment was created
-[AfterEach] Kubectl run deployment
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1738
-Sep 21 16:35:42.578: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete deployment e2e-test-httpd-deployment --namespace=kubectl-7353'
-Sep 21 16:35:42.782: INFO: stderr: ""
-Sep 21 16:35:42.782: INFO: stdout: "deployment.apps \"e2e-test-httpd-deployment\" deleted\n"
-[AfterEach] [sig-cli] Kubectl client
+STEP: create the container
+STEP: wait for the container to reach Succeeded
+STEP: get the container status
+STEP: the container should be terminated
+STEP: the termination message should be set
+Oct 26 16:42:50.639: INFO: Expected: &{OK} to match Container's Termination Message: OK --
+STEP: delete the container
+[AfterEach] [k8s.io] Container Runtime
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:35:42.782: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-7353" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run deployment should create a deployment from an image  [Conformance]","total":280,"completed":134,"skipped":2303,"failed":0}
-SSSSS
+Oct 26 16:42:50.709: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-runtime-1238" for this suite.
+•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":280,"completed":141,"skipped":2116,"failed":0}
+SSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-storage] Projected downwardAPI 
-  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:35:42.852: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 16:42:50.745: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:35:43.256: INFO: Waiting up to 5m0s for pod "downwardapi-volume-84deec03-3d7c-45c1-9e22-990b69e369cf" in namespace "projected-920" to be "success or failure"
-Sep 21 16:35:43.277: INFO: Pod "downwardapi-volume-84deec03-3d7c-45c1-9e22-990b69e369cf": Phase="Pending", Reason="", readiness=false. Elapsed: 21.220635ms
-Sep 21 16:35:45.302: INFO: Pod "downwardapi-volume-84deec03-3d7c-45c1-9e22-990b69e369cf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.045798477s
-Sep 21 16:35:47.325: INFO: Pod "downwardapi-volume-84deec03-3d7c-45c1-9e22-990b69e369cf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.069142413s
+Oct 26 16:42:50.945: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f376c8ca-02b6-4d6f-ad6c-fcd6af91d4a8" in namespace "projected-9648" to be "success or failure"
+Oct 26 16:42:50.967: INFO: Pod "downwardapi-volume-f376c8ca-02b6-4d6f-ad6c-fcd6af91d4a8": Phase="Pending", Reason="", readiness=false. Elapsed: 22.602144ms
+Oct 26 16:42:52.983: INFO: Pod "downwardapi-volume-f376c8ca-02b6-4d6f-ad6c-fcd6af91d4a8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.038517976s
+Oct 26 16:42:55.004: INFO: Pod "downwardapi-volume-f376c8ca-02b6-4d6f-ad6c-fcd6af91d4a8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.059179261s
 STEP: Saw pod success
-Sep 21 16:35:47.325: INFO: Pod "downwardapi-volume-84deec03-3d7c-45c1-9e22-990b69e369cf" satisfied condition "success or failure"
-Sep 21 16:35:47.443: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-84deec03-3d7c-45c1-9e22-990b69e369cf container client-container: 
+Oct 26 16:42:55.004: INFO: Pod "downwardapi-volume-f376c8ca-02b6-4d6f-ad6c-fcd6af91d4a8" satisfied condition "success or failure"
+Oct 26 16:42:55.024: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-f376c8ca-02b6-4d6f-ad6c-fcd6af91d4a8 container client-container: 
 STEP: delete the pod
-Sep 21 16:35:47.758: INFO: Waiting for pod downwardapi-volume-84deec03-3d7c-45c1-9e22-990b69e369cf to disappear
-Sep 21 16:35:47.786: INFO: Pod downwardapi-volume-84deec03-3d7c-45c1-9e22-990b69e369cf no longer exists
+Oct 26 16:42:55.149: INFO: Waiting for pod downwardapi-volume-f376c8ca-02b6-4d6f-ad6c-fcd6af91d4a8 to disappear
+Oct 26 16:42:55.165: INFO: Pod downwardapi-volume-f376c8ca-02b6-4d6f-ad6c-fcd6af91d4a8 no longer exists
 [AfterEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:35:47.786: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-920" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":135,"skipped":2308,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:42:55.166: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9648" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":142,"skipped":2133,"failed":0}
+SSSSSS
 ------------------------------
-[k8s.io] Container Runtime blackbox test when starting a container that exits 
-  should run with the expected status [NodeConformance] [Conformance]
+[sig-apps] Daemon set [Serial] 
+  should run and stop complex daemon [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Runtime
+[BeforeEach] [sig-apps] Daemon set [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:35:47.846: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-runtime
+Oct 26 16:42:55.204: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename daemonsets
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should run with the expected status [NodeConformance] [Conformance]
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
+[It] should run and stop complex daemon [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount'
-STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase'
-STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition
-STEP: Container 'terminate-cmd-rpa': should get the expected 'State'
-STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance]
-STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount'
-STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase'
-STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition
-STEP: Container 'terminate-cmd-rpof': should get the expected 'State'
-STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance]
-STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount'
-STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase'
-STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition
-STEP: Container 'terminate-cmd-rpn': should get the expected 'State'
-STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance]
-[AfterEach] [k8s.io] Container Runtime
+Oct 26 16:42:55.466: INFO: Creating daemon "daemon-set" with a node selector
+STEP: Initially, daemon pods should not be running on any nodes.
+Oct 26 16:42:55.530: INFO: Number of nodes with available pods: 0
+Oct 26 16:42:55.530: INFO: Number of running nodes: 0, number of available pods: 0
+STEP: Change node label to blue, check that daemon pod is launched.
+Oct 26 16:42:55.691: INFO: Number of nodes with available pods: 0
+Oct 26 16:42:55.691: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:42:56.708: INFO: Number of nodes with available pods: 0
+Oct 26 16:42:56.708: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:42:57.708: INFO: Number of nodes with available pods: 0
+Oct 26 16:42:57.708: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:42:58.708: INFO: Number of nodes with available pods: 1
+Oct 26 16:42:58.708: INFO: Number of running nodes: 1, number of available pods: 1
+STEP: Update the node label to green, and wait for daemons to be unscheduled
+Oct 26 16:42:58.802: INFO: Number of nodes with available pods: 1
+Oct 26 16:42:58.802: INFO: Number of running nodes: 0, number of available pods: 1
+Oct 26 16:42:59.822: INFO: Number of nodes with available pods: 0
+Oct 26 16:42:59.822: INFO: Number of running nodes: 0, number of available pods: 0
+STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate
+Oct 26 16:42:59.860: INFO: Number of nodes with available pods: 0
+Oct 26 16:42:59.860: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:43:00.877: INFO: Number of nodes with available pods: 0
+Oct 26 16:43:00.877: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:43:01.884: INFO: Number of nodes with available pods: 0
+Oct 26 16:43:01.884: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:43:02.888: INFO: Number of nodes with available pods: 0
+Oct 26 16:43:02.888: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:43:03.876: INFO: Number of nodes with available pods: 0
+Oct 26 16:43:03.876: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:43:04.877: INFO: Number of nodes with available pods: 1
+Oct 26 16:43:04.877: INFO: Number of running nodes: 1, number of available pods: 1
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-9412, will wait for the garbage collector to delete the pods
+Oct 26 16:43:05.024: INFO: Deleting DaemonSet.extensions daemon-set took: 39.365765ms
+Oct 26 16:43:05.125: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.272626ms
+Oct 26 16:43:18.344: INFO: Number of nodes with available pods: 0
+Oct 26 16:43:18.344: INFO: Number of running nodes: 0, number of available pods: 0
+Oct 26 16:43:18.360: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-9412/daemonsets","resourceVersion":"70992"},"items":null}
+
+Oct 26 16:43:18.377: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-9412/pods","resourceVersion":"70992"},"items":null}
+
+[AfterEach] [sig-apps] Daemon set [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:36:12.993: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-runtime-2548" for this suite.
+Oct 26 16:43:18.460: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "daemonsets-9412" for this suite.
 
-• [SLOW TEST:25.218 seconds]
-[k8s.io] Container Runtime
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  blackbox test
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38
-    when starting a container that exits
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:39
-      should run with the expected status [NodeConformance] [Conformance]
-      /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:23.295 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should run and stop complex daemon [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance]","total":280,"completed":136,"skipped":2364,"failed":0}
+{"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance]","total":280,"completed":143,"skipped":2139,"failed":0}
 SSSS
 ------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+[sig-apps] ReplicationController 
+  should surface a failure condition on a common issue like exceeded quota [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
+[BeforeEach] [sig-apps] ReplicationController
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:36:13.066: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename secrets
+Oct 26 16:43:18.499: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename replication-controller
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+[It] should surface a failure condition on a common issue like exceeded quota [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-map-f499ab68-da10-41f2-925a-f18ebd3c6a8a
-STEP: Creating a pod to test consume secrets
-Sep 21 16:36:13.625: INFO: Waiting up to 5m0s for pod "pod-secrets-c3d65113-0b9d-4cb4-8f9d-424b30488690" in namespace "secrets-2" to be "success or failure"
-Sep 21 16:36:13.651: INFO: Pod "pod-secrets-c3d65113-0b9d-4cb4-8f9d-424b30488690": Phase="Pending", Reason="", readiness=false. Elapsed: 25.976595ms
-Sep 21 16:36:15.678: INFO: Pod "pod-secrets-c3d65113-0b9d-4cb4-8f9d-424b30488690": Phase="Pending", Reason="", readiness=false. Elapsed: 2.053727896s
-Sep 21 16:36:17.703: INFO: Pod "pod-secrets-c3d65113-0b9d-4cb4-8f9d-424b30488690": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.078073105s
-STEP: Saw pod success
-Sep 21 16:36:17.703: INFO: Pod "pod-secrets-c3d65113-0b9d-4cb4-8f9d-424b30488690" satisfied condition "success or failure"
-Sep 21 16:36:17.728: INFO: Trying to get logs from node 10.241.51.147 pod pod-secrets-c3d65113-0b9d-4cb4-8f9d-424b30488690 container secret-volume-test: 
-STEP: delete the pod
-Sep 21 16:36:17.906: INFO: Waiting for pod pod-secrets-c3d65113-0b9d-4cb4-8f9d-424b30488690 to disappear
-Sep 21 16:36:17.928: INFO: Pod pod-secrets-c3d65113-0b9d-4cb4-8f9d-424b30488690 no longer exists
-[AfterEach] [sig-storage] Secrets
+Oct 26 16:43:18.639: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace
+STEP: Creating rc "condition-test" that asks for more than the allowed pod quota
+STEP: Checking rc "condition-test" has the desired failure condition set
+STEP: Scaling down rc "condition-test" to satisfy pod quota
+Oct 26 16:43:19.765: INFO: Updating replication controller "condition-test"
+STEP: Checking rc "condition-test" has no failure condition set
+[AfterEach] [sig-apps] ReplicationController
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:36:17.928: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-2" for this suite.
-•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":137,"skipped":2368,"failed":0}
-SS
+Oct 26 16:43:20.796: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replication-controller-4049" for this suite.
+•{"msg":"PASSED [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance]","total":280,"completed":144,"skipped":2143,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+[sig-storage] Secrets 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [sig-storage] Secrets
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:36:17.993: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename emptydir
+Oct 26 16:43:20.845: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename secrets
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0777 on node default medium
-Sep 21 16:36:18.400: INFO: Waiting up to 5m0s for pod "pod-38096c7d-a46a-4c57-9194-6f1874fa1f41" in namespace "emptydir-6185" to be "success or failure"
-Sep 21 16:36:18.431: INFO: Pod "pod-38096c7d-a46a-4c57-9194-6f1874fa1f41": Phase="Pending", Reason="", readiness=false. Elapsed: 31.062421ms
-Sep 21 16:36:20.466: INFO: Pod "pod-38096c7d-a46a-4c57-9194-6f1874fa1f41": Phase="Pending", Reason="", readiness=false. Elapsed: 2.065664962s
-Sep 21 16:36:22.490: INFO: Pod "pod-38096c7d-a46a-4c57-9194-6f1874fa1f41": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.090178859s
+STEP: Creating secret with name secret-test-map-21e7a39d-12c3-4fb3-9a53-f28804d344d4
+STEP: Creating a pod to test consume secrets
+Oct 26 16:43:21.088: INFO: Waiting up to 5m0s for pod "pod-secrets-16fc96c0-ee40-4a22-86cf-ac64bb2e5255" in namespace "secrets-4542" to be "success or failure"
+Oct 26 16:43:21.106: INFO: Pod "pod-secrets-16fc96c0-ee40-4a22-86cf-ac64bb2e5255": Phase="Pending", Reason="", readiness=false. Elapsed: 17.891088ms
+Oct 26 16:43:23.123: INFO: Pod "pod-secrets-16fc96c0-ee40-4a22-86cf-ac64bb2e5255": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034860916s
+Oct 26 16:43:25.142: INFO: Pod "pod-secrets-16fc96c0-ee40-4a22-86cf-ac64bb2e5255": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.053302509s
 STEP: Saw pod success
-Sep 21 16:36:22.490: INFO: Pod "pod-38096c7d-a46a-4c57-9194-6f1874fa1f41" satisfied condition "success or failure"
-Sep 21 16:36:22.516: INFO: Trying to get logs from node 10.241.51.147 pod pod-38096c7d-a46a-4c57-9194-6f1874fa1f41 container test-container: 
+Oct 26 16:43:25.142: INFO: Pod "pod-secrets-16fc96c0-ee40-4a22-86cf-ac64bb2e5255" satisfied condition "success or failure"
+Oct 26 16:43:25.159: INFO: Trying to get logs from node 10.72.119.74 pod pod-secrets-16fc96c0-ee40-4a22-86cf-ac64bb2e5255 container secret-volume-test: 
 STEP: delete the pod
-Sep 21 16:36:22.672: INFO: Waiting for pod pod-38096c7d-a46a-4c57-9194-6f1874fa1f41 to disappear
-Sep 21 16:36:22.695: INFO: Pod pod-38096c7d-a46a-4c57-9194-6f1874fa1f41 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+Oct 26 16:43:25.252: INFO: Waiting for pod pod-secrets-16fc96c0-ee40-4a22-86cf-ac64bb2e5255 to disappear
+Oct 26 16:43:25.268: INFO: Pod pod-secrets-16fc96c0-ee40-4a22-86cf-ac64bb2e5255 no longer exists
+[AfterEach] [sig-storage] Secrets
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:36:22.696: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-6185" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":138,"skipped":2370,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:43:25.268: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-4542" for this suite.
+•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":145,"skipped":2171,"failed":0}
+SSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] 
-  should be able to convert a non homogeneous list of CRs [Conformance]
+[k8s.io] Pods 
+  should be updated [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+[BeforeEach] [k8s.io] Pods
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:36:22.767: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename crd-webhook
+Oct 26 16:43:25.321: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename pods
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:125
-STEP: Setting up server cert
-STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication
-STEP: Deploying the custom resource conversion webhook pod
-STEP: Wait for the deployment to be ready
-Sep 21 16:36:23.971: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736302983, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736302983, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736302983, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736302983, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-78dcf5dd84\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:36:25.994: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736302983, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736302983, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736302983, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736302983, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-78dcf5dd84\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Sep 21 16:36:29.094: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1
-[It] should be able to convert a non homogeneous list of CRs [Conformance]
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
+[It] should be updated [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:36:29.126: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Creating a v1 custom resource
-STEP: Create a v2 custom resource
-STEP: List CRs in v1
-STEP: List CRs in v2
-[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+STEP: updating the pod
+Oct 26 16:43:28.170: INFO: Successfully updated pod "pod-update-77ec990f-8ed0-4128-aac8-8672fce7a20f"
+STEP: verifying the updated pod is in kubernetes
+Oct 26 16:43:28.204: INFO: Pod update OK
+[AfterEach] [k8s.io] Pods
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:36:31.627: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-webhook-7796" for this suite.
-[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:136
-
-• [SLOW TEST:9.443 seconds]
-[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should be able to convert a non homogeneous list of CRs [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance]","total":280,"completed":139,"skipped":2470,"failed":0}
-SSS
+Oct 26 16:43:28.204: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-1885" for this suite.
+•{"msg":"PASSED [k8s.io] Pods should be updated [NodeConformance] [Conformance]","total":280,"completed":146,"skipped":2186,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-network] Services 
-  should provide secure master service  [Conformance]
+[sig-apps] ReplicationController 
+  should adopt matching pods on creation [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Services
+[BeforeEach] [sig-apps] ReplicationController
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:36:32.210: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename services
+Oct 26 16:43:28.256: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename replication-controller
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should provide secure master service  [Conformance]
+[It] should adopt matching pods on creation [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [sig-network] Services
+STEP: Given a Pod with a 'name' label pod-adoption is created
+STEP: When a replication controller with a matching selector is created
+STEP: Then the orphan pod is adopted
+[AfterEach] [sig-apps] ReplicationController
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:36:32.559: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-4925" for this suite.
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
-•{"msg":"PASSED [sig-network] Services should provide secure master service  [Conformance]","total":280,"completed":140,"skipped":2473,"failed":0}
-S
+Oct 26 16:43:31.583: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replication-controller-1745" for this suite.
+•{"msg":"PASSED [sig-apps] ReplicationController should adopt matching pods on creation [Conformance]","total":280,"completed":147,"skipped":2233,"failed":0}
+SSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-auth] ServiceAccounts 
-  should allow opting out of API token automount  [Conformance]
+[sig-storage] EmptyDir volumes 
+  should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-auth] ServiceAccounts
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:36:32.644: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename svcaccounts
+Oct 26 16:43:31.625: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should allow opting out of API token automount  [Conformance]
+[It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: getting the auto-created API token
-Sep 21 16:36:33.641: INFO: created pod pod-service-account-defaultsa
-Sep 21 16:36:33.641: INFO: pod pod-service-account-defaultsa service account token volume mount: true
-Sep 21 16:36:33.718: INFO: created pod pod-service-account-mountsa
-Sep 21 16:36:33.718: INFO: pod pod-service-account-mountsa service account token volume mount: true
-Sep 21 16:36:33.787: INFO: created pod pod-service-account-nomountsa
-Sep 21 16:36:33.787: INFO: pod pod-service-account-nomountsa service account token volume mount: false
-Sep 21 16:36:33.850: INFO: created pod pod-service-account-defaultsa-mountspec
-Sep 21 16:36:33.850: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true
-Sep 21 16:36:33.926: INFO: created pod pod-service-account-mountsa-mountspec
-Sep 21 16:36:33.926: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true
-Sep 21 16:36:34.001: INFO: created pod pod-service-account-nomountsa-mountspec
-Sep 21 16:36:34.001: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true
-Sep 21 16:36:34.079: INFO: created pod pod-service-account-defaultsa-nomountspec
-Sep 21 16:36:34.079: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false
-Sep 21 16:36:34.161: INFO: created pod pod-service-account-mountsa-nomountspec
-Sep 21 16:36:34.161: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false
-Sep 21 16:36:34.269: INFO: created pod pod-service-account-nomountsa-nomountspec
-Sep 21 16:36:34.269: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false
-[AfterEach] [sig-auth] ServiceAccounts
+STEP: Creating a pod to test emptydir 0666 on node default medium
+Oct 26 16:43:32.831: INFO: Waiting up to 5m0s for pod "pod-d910fd2b-ab87-495f-8994-17890335cb55" in namespace "emptydir-8626" to be "success or failure"
+Oct 26 16:43:32.850: INFO: Pod "pod-d910fd2b-ab87-495f-8994-17890335cb55": Phase="Pending", Reason="", readiness=false. Elapsed: 19.088606ms
+Oct 26 16:43:34.907: INFO: Pod "pod-d910fd2b-ab87-495f-8994-17890335cb55": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.075316575s
+STEP: Saw pod success
+Oct 26 16:43:34.907: INFO: Pod "pod-d910fd2b-ab87-495f-8994-17890335cb55" satisfied condition "success or failure"
+Oct 26 16:43:34.925: INFO: Trying to get logs from node 10.72.119.74 pod pod-d910fd2b-ab87-495f-8994-17890335cb55 container test-container: 
+STEP: delete the pod
+Oct 26 16:43:35.020: INFO: Waiting for pod pod-d910fd2b-ab87-495f-8994-17890335cb55 to disappear
+Oct 26 16:43:35.038: INFO: Pod pod-d910fd2b-ab87-495f-8994-17890335cb55 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:36:34.269: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "svcaccounts-7786" for this suite.
-•{"msg":"PASSED [sig-auth] ServiceAccounts should allow opting out of API token automount  [Conformance]","total":280,"completed":141,"skipped":2474,"failed":0}
-SSSSSSSSSSSSSSS
+Oct 26 16:43:35.038: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-8626" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":148,"skipped":2254,"failed":0}
+SSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition 
+  listing custom resource definition objects works  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:36:34.375: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename emptydir
+Oct 26 16:43:35.083: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename custom-resource-definition
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+[It] listing custom resource definition objects works  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir volume type on tmpfs
-Sep 21 16:36:34.782: INFO: Waiting up to 5m0s for pod "pod-86b3f1b5-a974-4381-92e3-2a3471f16af6" in namespace "emptydir-7756" to be "success or failure"
-Sep 21 16:36:34.811: INFO: Pod "pod-86b3f1b5-a974-4381-92e3-2a3471f16af6": Phase="Pending", Reason="", readiness=false. Elapsed: 28.991165ms
-Sep 21 16:36:36.853: INFO: Pod "pod-86b3f1b5-a974-4381-92e3-2a3471f16af6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.070573706s
-STEP: Saw pod success
-Sep 21 16:36:36.853: INFO: Pod "pod-86b3f1b5-a974-4381-92e3-2a3471f16af6" satisfied condition "success or failure"
-Sep 21 16:36:36.881: INFO: Trying to get logs from node 10.241.51.147 pod pod-86b3f1b5-a974-4381-92e3-2a3471f16af6 container test-container: 
-STEP: delete the pod
-Sep 21 16:36:37.032: INFO: Waiting for pod pod-86b3f1b5-a974-4381-92e3-2a3471f16af6 to disappear
-Sep 21 16:36:37.061: INFO: Pod pod-86b3f1b5-a974-4381-92e3-2a3471f16af6 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+Oct 26 16:43:35.217: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:36:37.061: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-7756" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":142,"skipped":2489,"failed":0}
-SSSSSSSS
+Oct 26 16:43:42.814: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "custom-resource-definition-4370" for this suite.
+
+• [SLOW TEST:7.779 seconds]
+[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  Simple CustomResourceDefinition
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:47
+    listing custom resource definition objects works  [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] Probing container 
-  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works  [Conformance]","total":280,"completed":149,"skipped":2263,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
+  should have an terminated reason [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Probing container
+[BeforeEach] [k8s.io] Kubelet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:36:37.142: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-probe
+Oct 26 16:43:42.862: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubelet-test
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[BeforeEach] when scheduling a busybox command that always fails in a pod
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
+[It] should have an terminated reason [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod liveness-b7791d6e-389f-44a6-86f4-45a0754f854b in namespace container-probe-3287
-Sep 21 16:36:41.616: INFO: Started pod liveness-b7791d6e-389f-44a6-86f4-45a0754f854b in namespace container-probe-3287
-STEP: checking the pod's current state and verifying that restartCount is present
-Sep 21 16:36:41.641: INFO: Initial restart count of pod liveness-b7791d6e-389f-44a6-86f4-45a0754f854b is 0
-Sep 21 16:37:01.927: INFO: Restart count of pod container-probe-3287/liveness-b7791d6e-389f-44a6-86f4-45a0754f854b is now 1 (20.28603335s elapsed)
-STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
+[AfterEach] [k8s.io] Kubelet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:37:02.028: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-3287" for this suite.
+Oct 26 16:43:47.116: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-9656" for this suite.
+•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance]","total":280,"completed":150,"skipped":2285,"failed":0}
 
-• [SLOW TEST:24.948 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":280,"completed":143,"skipped":2497,"failed":0}
-[sig-api-machinery] ResourceQuota 
-  should verify ResourceQuota with best effort scope. [Conformance]
+[sig-api-machinery] Watchers 
+  should receive events on concurrent watches in same order [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
+[BeforeEach] [sig-api-machinery] Watchers
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:37:02.091: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename resourcequota
+Oct 26 16:43:47.161: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename watch
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should verify ResourceQuota with best effort scope. [Conformance]
+[It] should receive events on concurrent watches in same order [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a ResourceQuota with best effort scope
-STEP: Ensuring ResourceQuota status is calculated
-STEP: Creating a ResourceQuota with not best effort scope
-STEP: Ensuring ResourceQuota status is calculated
-STEP: Creating a best-effort pod
-STEP: Ensuring resource quota with best effort scope captures the pod usage
-STEP: Ensuring resource quota with not best effort ignored the pod usage
-STEP: Deleting the pod
-STEP: Ensuring resource quota status released the pod usage
-STEP: Creating a not best-effort pod
-STEP: Ensuring resource quota with not best effort scope captures the pod usage
-STEP: Ensuring resource quota with best effort scope ignored the pod usage
-STEP: Deleting the pod
-STEP: Ensuring resource quota status released the pod usage
-[AfterEach] [sig-api-machinery] ResourceQuota
+STEP: starting a background goroutine to produce watch events
+STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order
+[AfterEach] [sig-api-machinery] Watchers
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:37:19.373: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-559" for this suite.
-
-• [SLOW TEST:17.428 seconds]
-[sig-api-machinery] ResourceQuota
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should verify ResourceQuota with best effort scope. [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance]","total":280,"completed":144,"skipped":2497,"failed":0}
-SSSSSSSSS
+Oct 26 16:43:52.062: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-2216" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance]","total":280,"completed":151,"skipped":2285,"failed":0}
+SSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  binary data should be reflected in volume [NodeConformance] [Conformance]
+[sig-network] DNS 
+  should support configurable pod DNS nameservers [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
+[BeforeEach] [sig-network] DNS
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:37:19.518: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename configmap
+Oct 26 16:43:52.152: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename dns
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] binary data should be reflected in volume [NodeConformance] [Conformance]
+[It] should support configurable pod DNS nameservers [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:37:19.809: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
-STEP: Creating configMap with name configmap-test-upd-2d2c4104-df53-4931-a3ae-247852278115
-STEP: Creating the pod
-STEP: Waiting for pod with text data
-STEP: Waiting for pod with binary data
-[AfterEach] [sig-storage] ConfigMap
+STEP: Creating a pod with dnsPolicy=None and customized dnsConfig...
+Oct 26 16:43:52.368: INFO: Created pod &Pod{ObjectMeta:{dns-5018  dns-5018 /api/v1/namespaces/dns-5018/pods/dns-5018 5e1a5799-09d6-4c55-bc3e-fdc924f63874 71897 0 2020-10-26 16:43:52 +0000 UTC   map[] map[openshift.io/scc:anyuid] [] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-z8sz5,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-z8sz5,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[pause],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-z8sz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:None,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:&SELinuxOptions{User:,Role:,Type:,Level:s0:c57,c9,},RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:&PodDNSConfig{Nameservers:[1.1.1.1],Searches:[resolv.conf.local],Options:[]PodDNSConfigOption{},},ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+STEP: Verifying customized DNS suffix list is configured on pod...
+Oct 26 16:43:56.402: INFO: ExecWithOptions {Command:[/agnhost dns-suffix] Namespace:dns-5018 PodName:dns-5018 ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:43:56.402: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Verifying customized DNS server is configured on pod...
+Oct 26 16:43:56.680: INFO: ExecWithOptions {Command:[/agnhost dns-server-list] Namespace:dns-5018 PodName:dns-5018 ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:43:56.680: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:43:56.913: INFO: Deleting pod dns-5018...
+[AfterEach] [sig-network] DNS
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:37:22.209: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-8553" for this suite.
-•{"msg":"PASSED [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":145,"skipped":2506,"failed":0}
-SSS
+Oct 26 16:43:56.965: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-5018" for this suite.
+•{"msg":"PASSED [sig-network] DNS should support configurable pod DNS nameservers [Conformance]","total":280,"completed":152,"skipped":2290,"failed":0}
+SSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should update annotations on modification [NodeConformance] [Conformance]
+[sig-cli] Kubectl client Proxy server 
+  should support --unix-socket=/path  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:37:22.276: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
+Oct 26 16:43:57.003: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubectl
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should update annotations on modification [NodeConformance] [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
+[It] should support --unix-socket=/path  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating the pod
-Sep 21 16:37:27.426: INFO: Successfully updated pod "annotationupdate08557f86-2264-4862-9b21-fdd4adafec54"
-[AfterEach] [sig-storage] Downward API volume
+STEP: Starting the proxy
+Oct 26 16:43:57.149: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-871182394 proxy --unix-socket=/tmp/kubectl-proxy-unix355593385/test'
+STEP: retrieving proxy /api/ output
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:37:29.570: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-2310" for this suite.
+Oct 26 16:43:57.253: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-5350" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support --unix-socket=/path  [Conformance]","total":280,"completed":153,"skipped":2306,"failed":0}
 
-• [SLOW TEST:7.453 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:35
-  should update annotations on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance]","total":280,"completed":146,"skipped":2509,"failed":0}
-S
 ------------------------------
-[sig-storage] Downward API volume 
-  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+[sig-storage] Projected downwardAPI 
+  should provide container's cpu request [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
+[BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:37:29.729: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
+Oct 26 16:43:57.295: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should provide container's cpu request [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:37:30.293: INFO: Waiting up to 5m0s for pod "downwardapi-volume-0bf04299-c1ce-40e7-afea-283ca806c47b" in namespace "downward-api-8667" to be "success or failure"
-Sep 21 16:37:30.318: INFO: Pod "downwardapi-volume-0bf04299-c1ce-40e7-afea-283ca806c47b": Phase="Pending", Reason="", readiness=false. Elapsed: 25.030904ms
-Sep 21 16:37:32.343: INFO: Pod "downwardapi-volume-0bf04299-c1ce-40e7-afea-283ca806c47b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.049835194s
-Sep 21 16:37:34.369: INFO: Pod "downwardapi-volume-0bf04299-c1ce-40e7-afea-283ca806c47b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.07603359s
+Oct 26 16:43:57.502: INFO: Waiting up to 5m0s for pod "downwardapi-volume-3f3b67dd-0f96-49b8-a8b1-62c6b69fd6d4" in namespace "projected-5664" to be "success or failure"
+Oct 26 16:43:57.517: INFO: Pod "downwardapi-volume-3f3b67dd-0f96-49b8-a8b1-62c6b69fd6d4": Phase="Pending", Reason="", readiness=false. Elapsed: 15.10524ms
+Oct 26 16:43:59.536: INFO: Pod "downwardapi-volume-3f3b67dd-0f96-49b8-a8b1-62c6b69fd6d4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034166347s
+Oct 26 16:44:01.556: INFO: Pod "downwardapi-volume-3f3b67dd-0f96-49b8-a8b1-62c6b69fd6d4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.053989514s
 STEP: Saw pod success
-Sep 21 16:37:34.370: INFO: Pod "downwardapi-volume-0bf04299-c1ce-40e7-afea-283ca806c47b" satisfied condition "success or failure"
-Sep 21 16:37:34.395: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-0bf04299-c1ce-40e7-afea-283ca806c47b container client-container: 
+Oct 26 16:44:01.556: INFO: Pod "downwardapi-volume-3f3b67dd-0f96-49b8-a8b1-62c6b69fd6d4" satisfied condition "success or failure"
+Oct 26 16:44:01.573: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-3f3b67dd-0f96-49b8-a8b1-62c6b69fd6d4 container client-container: 
 STEP: delete the pod
-Sep 21 16:37:34.572: INFO: Waiting for pod downwardapi-volume-0bf04299-c1ce-40e7-afea-283ca806c47b to disappear
-Sep 21 16:37:34.596: INFO: Pod downwardapi-volume-0bf04299-c1ce-40e7-afea-283ca806c47b no longer exists
-[AfterEach] [sig-storage] Downward API volume
+Oct 26 16:44:01.672: INFO: Waiting for pod downwardapi-volume-3f3b67dd-0f96-49b8-a8b1-62c6b69fd6d4 to disappear
+Oct 26 16:44:01.690: INFO: Pod downwardapi-volume-3f3b67dd-0f96-49b8-a8b1-62c6b69fd6d4 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:37:34.596: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-8667" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":147,"skipped":2510,"failed":0}
-SSSSSSS
+Oct 26 16:44:01.691: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-5664" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance]","total":280,"completed":154,"skipped":2306,"failed":0}
+SSSSSSSSSS
 ------------------------------
-[sig-apps] ReplicaSet 
-  should serve a basic image on each replica with a public image  [Conformance]
+[sig-storage] EmptyDir volumes 
+  should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] ReplicaSet
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:37:34.660: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename replicaset
+Oct 26 16:44:01.731: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should serve a basic image on each replica with a public image  [Conformance]
+[It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:37:35.011: INFO: Creating ReplicaSet my-hostname-basic-e4c8f663-3643-4daa-8bf2-16f2d1771d21
-Sep 21 16:37:35.109: INFO: Pod name my-hostname-basic-e4c8f663-3643-4daa-8bf2-16f2d1771d21: Found 0 pods out of 1
-Sep 21 16:37:40.135: INFO: Pod name my-hostname-basic-e4c8f663-3643-4daa-8bf2-16f2d1771d21: Found 1 pods out of 1
-Sep 21 16:37:40.136: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-e4c8f663-3643-4daa-8bf2-16f2d1771d21" is running
-Sep 21 16:37:40.160: INFO: Pod "my-hostname-basic-e4c8f663-3643-4daa-8bf2-16f2d1771d21-scnrn" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-09-21 16:37:35 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-09-21 16:37:37 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-09-21 16:37:37 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-09-21 16:37:35 +0000 UTC Reason: Message:}])
-Sep 21 16:37:40.160: INFO: Trying to dial the pod
-Sep 21 16:37:45.269: INFO: Controller my-hostname-basic-e4c8f663-3643-4daa-8bf2-16f2d1771d21: Got expected result from replica 1 [my-hostname-basic-e4c8f663-3643-4daa-8bf2-16f2d1771d21-scnrn]: "my-hostname-basic-e4c8f663-3643-4daa-8bf2-16f2d1771d21-scnrn", 1 of 1 required successes so far
-[AfterEach] [sig-apps] ReplicaSet
+STEP: Creating a pod to test emptydir 0777 on node default medium
+Oct 26 16:44:01.925: INFO: Waiting up to 5m0s for pod "pod-64acf5d0-5e24-4a01-8154-e219fedf2899" in namespace "emptydir-6129" to be "success or failure"
+Oct 26 16:44:01.940: INFO: Pod "pod-64acf5d0-5e24-4a01-8154-e219fedf2899": Phase="Pending", Reason="", readiness=false. Elapsed: 15.286664ms
+Oct 26 16:44:03.956: INFO: Pod "pod-64acf5d0-5e24-4a01-8154-e219fedf2899": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.03123594s
+STEP: Saw pod success
+Oct 26 16:44:03.956: INFO: Pod "pod-64acf5d0-5e24-4a01-8154-e219fedf2899" satisfied condition "success or failure"
+Oct 26 16:44:03.973: INFO: Trying to get logs from node 10.72.119.74 pod pod-64acf5d0-5e24-4a01-8154-e219fedf2899 container test-container: 
+STEP: delete the pod
+Oct 26 16:44:04.077: INFO: Waiting for pod pod-64acf5d0-5e24-4a01-8154-e219fedf2899 to disappear
+Oct 26 16:44:04.098: INFO: Pod pod-64acf5d0-5e24-4a01-8154-e219fedf2899 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:37:45.269: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replicaset-6497" for this suite.
-
-• [SLOW TEST:10.672 seconds]
-[sig-apps] ReplicaSet
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] ReplicaSet should serve a basic image on each replica with a public image  [Conformance]","total":280,"completed":148,"skipped":2517,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:44:04.098: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-6129" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":155,"skipped":2316,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-node] Downward API 
-  should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
+[sig-storage] Projected secret 
+  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] Downward API
+[BeforeEach] [sig-storage] Projected secret
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:37:45.332: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
+Oct 26 16:44:04.143: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward api env vars
-Sep 21 16:37:45.780: INFO: Waiting up to 5m0s for pod "downward-api-6283506b-21e0-4a4c-b815-5992d3b23938" in namespace "downward-api-3390" to be "success or failure"
-Sep 21 16:37:45.801: INFO: Pod "downward-api-6283506b-21e0-4a4c-b815-5992d3b23938": Phase="Pending", Reason="", readiness=false. Elapsed: 21.242183ms
-Sep 21 16:37:47.831: INFO: Pod "downward-api-6283506b-21e0-4a4c-b815-5992d3b23938": Phase="Pending", Reason="", readiness=false. Elapsed: 2.05112862s
-Sep 21 16:37:49.856: INFO: Pod "downward-api-6283506b-21e0-4a4c-b815-5992d3b23938": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.076474418s
+STEP: Creating projection with secret that has name projected-secret-test-map-e84dcd92-5667-4c67-a5fd-67b31f4d4eb5
+STEP: Creating a pod to test consume secrets
+Oct 26 16:44:04.368: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-f3c86ae4-a0e6-41f8-abcd-ffaf2ef2beef" in namespace "projected-1547" to be "success or failure"
+Oct 26 16:44:04.385: INFO: Pod "pod-projected-secrets-f3c86ae4-a0e6-41f8-abcd-ffaf2ef2beef": Phase="Pending", Reason="", readiness=false. Elapsed: 16.688197ms
+Oct 26 16:44:06.401: INFO: Pod "pod-projected-secrets-f3c86ae4-a0e6-41f8-abcd-ffaf2ef2beef": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033352085s
+Oct 26 16:44:08.418: INFO: Pod "pod-projected-secrets-f3c86ae4-a0e6-41f8-abcd-ffaf2ef2beef": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.05015294s
 STEP: Saw pod success
-Sep 21 16:37:49.856: INFO: Pod "downward-api-6283506b-21e0-4a4c-b815-5992d3b23938" satisfied condition "success or failure"
-Sep 21 16:37:49.882: INFO: Trying to get logs from node 10.241.51.147 pod downward-api-6283506b-21e0-4a4c-b815-5992d3b23938 container dapi-container: 
+Oct 26 16:44:08.418: INFO: Pod "pod-projected-secrets-f3c86ae4-a0e6-41f8-abcd-ffaf2ef2beef" satisfied condition "success or failure"
+Oct 26 16:44:08.433: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-secrets-f3c86ae4-a0e6-41f8-abcd-ffaf2ef2beef container projected-secret-volume-test: 
 STEP: delete the pod
-Sep 21 16:37:50.023: INFO: Waiting for pod downward-api-6283506b-21e0-4a4c-b815-5992d3b23938 to disappear
-Sep 21 16:37:50.047: INFO: Pod downward-api-6283506b-21e0-4a4c-b815-5992d3b23938 no longer exists
-[AfterEach] [sig-node] Downward API
+Oct 26 16:44:08.522: INFO: Waiting for pod pod-projected-secrets-f3c86ae4-a0e6-41f8-abcd-ffaf2ef2beef to disappear
+Oct 26 16:44:08.538: INFO: Pod pod-projected-secrets-f3c86ae4-a0e6-41f8-abcd-ffaf2ef2beef no longer exists
+[AfterEach] [sig-storage] Projected secret
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:37:50.047: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-3390" for this suite.
-•{"msg":"PASSED [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]","total":280,"completed":149,"skipped":2546,"failed":0}
-SSSSSSSSSSSSSSSS
+Oct 26 16:44:08.538: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-1547" for this suite.
+•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":156,"skipped":2368,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-network] Networking Granular Checks: Pods 
-  should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+[sig-storage] Projected downwardAPI 
+  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Networking
+[BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:37:50.127: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename pod-network-test
+Oct 26 16:44:08.583: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Performing setup for networking test in namespace pod-network-test-7749
-STEP: creating a selector
-STEP: Creating the service pods in kubernetes
-Sep 21 16:37:50.510: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
-STEP: Creating test pods
-Sep 21 16:38:15.894: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.30.237.74 8081 | grep -v '^\s*$'] Namespace:pod-network-test-7749 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:38:15.894: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:38:17.176: INFO: Found all expected endpoints: [netserver-0]
-Sep 21 16:38:17.204: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.30.139.48 8081 | grep -v '^\s*$'] Namespace:pod-network-test-7749 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:38:17.204: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:38:18.536: INFO: Found all expected endpoints: [netserver-1]
-Sep 21 16:38:18.561: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.30.81.33 8081 | grep -v '^\s*$'] Namespace:pod-network-test-7749 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:38:18.561: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:38:19.853: INFO: Found all expected endpoints: [netserver-2]
-[AfterEach] [sig-network] Networking
+STEP: Creating a pod to test downward API volume plugin
+Oct 26 16:44:08.775: INFO: Waiting up to 5m0s for pod "downwardapi-volume-64db1c9b-7c56-4838-8db5-ae918b198804" in namespace "projected-785" to be "success or failure"
+Oct 26 16:44:08.791: INFO: Pod "downwardapi-volume-64db1c9b-7c56-4838-8db5-ae918b198804": Phase="Pending", Reason="", readiness=false. Elapsed: 16.011111ms
+Oct 26 16:44:10.808: INFO: Pod "downwardapi-volume-64db1c9b-7c56-4838-8db5-ae918b198804": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.033175541s
+STEP: Saw pod success
+Oct 26 16:44:10.808: INFO: Pod "downwardapi-volume-64db1c9b-7c56-4838-8db5-ae918b198804" satisfied condition "success or failure"
+Oct 26 16:44:10.825: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-64db1c9b-7c56-4838-8db5-ae918b198804 container client-container: 
+STEP: delete the pod
+Oct 26 16:44:10.925: INFO: Waiting for pod downwardapi-volume-64db1c9b-7c56-4838-8db5-ae918b198804 to disappear
+Oct 26 16:44:10.941: INFO: Pod downwardapi-volume-64db1c9b-7c56-4838-8db5-ae918b198804 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:38:19.853: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pod-network-test-7749" for this suite.
-
-• [SLOW TEST:29.795 seconds]
-[sig-network] Networking
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
-  Granular Checks: Pods
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
-    should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Oct 26 16:44:10.941: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-785" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":157,"skipped":2393,"failed":0}
+SSSSSSSS
 ------------------------------
-{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":150,"skipped":2562,"failed":0}
-[sig-api-machinery] Watchers 
-  should observe add, update, and delete watch notifications on configmaps [Conformance]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  works for multiple CRDs of same group but different versions [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Watchers
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:38:19.921: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename watch
+Oct 26 16:44:10.997: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename crd-publish-openapi
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should observe add, update, and delete watch notifications on configmaps [Conformance]
+[It] works for multiple CRDs of same group but different versions [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a watch on configmaps with label A
-STEP: creating a watch on configmaps with label B
-STEP: creating a watch on configmaps with label A or B
-STEP: creating a configmap with label A and ensuring the correct watchers observe the notification
-Sep 21 16:38:20.251: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-7336 /api/v1/namespaces/watch-7336/configmaps/e2e-watch-test-configmap-a 50716890-87f5-47f9-bcb0-536a01829d6e 74430 0 2020-09-21 16:38:20 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Sep 21 16:38:20.251: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-7336 /api/v1/namespaces/watch-7336/configmaps/e2e-watch-test-configmap-a 50716890-87f5-47f9-bcb0-536a01829d6e 74430 0 2020-09-21 16:38:20 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-STEP: modifying configmap A and ensuring the correct watchers observe the notification
-Sep 21 16:38:30.295: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-7336 /api/v1/namespaces/watch-7336/configmaps/e2e-watch-test-configmap-a 50716890-87f5-47f9-bcb0-536a01829d6e 74544 0 2020-09-21 16:38:20 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-Sep 21 16:38:30.295: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-7336 /api/v1/namespaces/watch-7336/configmaps/e2e-watch-test-configmap-a 50716890-87f5-47f9-bcb0-536a01829d6e 74544 0 2020-09-21 16:38:20 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-STEP: modifying configmap A again and ensuring the correct watchers observe the notification
-Sep 21 16:38:40.342: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-7336 /api/v1/namespaces/watch-7336/configmaps/e2e-watch-test-configmap-a 50716890-87f5-47f9-bcb0-536a01829d6e 74592 0 2020-09-21 16:38:20 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Sep 21 16:38:40.342: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-7336 /api/v1/namespaces/watch-7336/configmaps/e2e-watch-test-configmap-a 50716890-87f5-47f9-bcb0-536a01829d6e 74592 0 2020-09-21 16:38:20 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-STEP: deleting configmap A and ensuring the correct watchers observe the notification
-Sep 21 16:38:50.682: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-7336 /api/v1/namespaces/watch-7336/configmaps/e2e-watch-test-configmap-a 50716890-87f5-47f9-bcb0-536a01829d6e 74636 0 2020-09-21 16:38:20 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Sep 21 16:38:50.683: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-7336 /api/v1/namespaces/watch-7336/configmaps/e2e-watch-test-configmap-a 50716890-87f5-47f9-bcb0-536a01829d6e 74636 0 2020-09-21 16:38:20 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-STEP: creating a configmap with label B and ensuring the correct watchers observe the notification
-Sep 21 16:39:00.724: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b  watch-7336 /api/v1/namespaces/watch-7336/configmaps/e2e-watch-test-configmap-b 4595b2ff-2d5b-4b20-923c-01007a901e60 74683 0 2020-09-21 16:39:00 +0000 UTC   map[watch-this-configmap:multiple-watchers-B] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Sep 21 16:39:00.724: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b  watch-7336 /api/v1/namespaces/watch-7336/configmaps/e2e-watch-test-configmap-b 4595b2ff-2d5b-4b20-923c-01007a901e60 74683 0 2020-09-21 16:39:00 +0000 UTC   map[watch-this-configmap:multiple-watchers-B] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-STEP: deleting configmap B and ensuring the correct watchers observe the notification
-Sep 21 16:39:10.763: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b  watch-7336 /api/v1/namespaces/watch-7336/configmaps/e2e-watch-test-configmap-b 4595b2ff-2d5b-4b20-923c-01007a901e60 74722 0 2020-09-21 16:39:00 +0000 UTC   map[watch-this-configmap:multiple-watchers-B] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Sep 21 16:39:10.763: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b  watch-7336 /api/v1/namespaces/watch-7336/configmaps/e2e-watch-test-configmap-b 4595b2ff-2d5b-4b20-923c-01007a901e60 74722 0 2020-09-21 16:39:00 +0000 UTC   map[watch-this-configmap:multiple-watchers-B] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-[AfterEach] [sig-api-machinery] Watchers
+STEP: CRs in the same group but different versions (one multiversion CRD) show up in OpenAPI documentation
+Oct 26 16:44:11.147: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: CRs in the same group but different versions (two CRDs) show up in OpenAPI documentation
+Oct 26 16:44:41.365: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:44:49.385: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:39:20.763: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "watch-7336" for this suite.
+Oct 26 16:45:21.916: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-5889" for this suite.
 
-• [SLOW TEST:60.919 seconds]
-[sig-api-machinery] Watchers
+• [SLOW TEST:70.957 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should observe add, update, and delete watch notifications on configmaps [Conformance]
+  works for multiple CRDs of same group but different versions [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance]","total":280,"completed":151,"skipped":2562,"failed":0}
-SSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance]","total":280,"completed":158,"skipped":2401,"failed":0}
+SSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] ResourceQuota 
-  should create a ResourceQuota and capture the life of a replication controller. [Conformance]
+[sig-api-machinery] Secrets 
+  should be consumable via the environment [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
+[BeforeEach] [sig-api-machinery] Secrets
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:39:20.840: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename resourcequota
+Oct 26 16:45:21.954: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename secrets
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should create a ResourceQuota and capture the life of a replication controller. [Conformance]
+[It] should be consumable via the environment [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Counting existing ResourceQuota
-STEP: Creating a ResourceQuota
-STEP: Ensuring resource quota status is calculated
-STEP: Creating a ReplicationController
-STEP: Ensuring resource quota status captures replication controller creation
-STEP: Deleting a ReplicationController
-STEP: Ensuring resource quota status released usage
-[AfterEach] [sig-api-machinery] ResourceQuota
+STEP: creating secret secrets-6036/secret-test-ee1af6b7-7cb9-4547-91b3-ec74a65d9b56
+STEP: Creating a pod to test consume secrets
+Oct 26 16:45:22.203: INFO: Waiting up to 5m0s for pod "pod-configmaps-f6b6b926-9d8b-4aa9-ac97-f2aeb66127cd" in namespace "secrets-6036" to be "success or failure"
+Oct 26 16:45:22.222: INFO: Pod "pod-configmaps-f6b6b926-9d8b-4aa9-ac97-f2aeb66127cd": Phase="Pending", Reason="", readiness=false. Elapsed: 16.138268ms
+Oct 26 16:45:24.239: INFO: Pod "pod-configmaps-f6b6b926-9d8b-4aa9-ac97-f2aeb66127cd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.033114185s
+STEP: Saw pod success
+Oct 26 16:45:24.239: INFO: Pod "pod-configmaps-f6b6b926-9d8b-4aa9-ac97-f2aeb66127cd" satisfied condition "success or failure"
+Oct 26 16:45:24.255: INFO: Trying to get logs from node 10.72.119.74 pod pod-configmaps-f6b6b926-9d8b-4aa9-ac97-f2aeb66127cd container env-test: 
+STEP: delete the pod
+Oct 26 16:45:24.347: INFO: Waiting for pod pod-configmaps-f6b6b926-9d8b-4aa9-ac97-f2aeb66127cd to disappear
+Oct 26 16:45:24.362: INFO: Pod pod-configmaps-f6b6b926-9d8b-4aa9-ac97-f2aeb66127cd no longer exists
+[AfterEach] [sig-api-machinery] Secrets
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:39:32.536: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-1852" for this suite.
-
-• [SLOW TEST:11.770 seconds]
-[sig-api-machinery] ResourceQuota
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should create a ResourceQuota and capture the life of a replication controller. [Conformance]
+Oct 26 16:45:24.362: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-6036" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Secrets should be consumable via the environment [NodeConformance] [Conformance]","total":280,"completed":159,"skipped":2420,"failed":0}
+SSSS
+------------------------------
+[sig-api-machinery] Watchers 
+  should be able to start watching from a specific resource version [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 16:45:24.396: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename watch
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be able to start watching from a specific resource version [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating a new configmap
+STEP: modifying the configmap once
+STEP: modifying the configmap a second time
+STEP: deleting the configmap
+STEP: creating a watch on configmaps from the resource version returned by the first update
+STEP: Expecting to observe notifications for all changes to the configmap after the first update
+Oct 26 16:45:24.700: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version  watch-6898 /api/v1/namespaces/watch-6898/configmaps/e2e-watch-test-resource-version 4dcf137b-3281-464a-9125-20d64ac86507 72901 0 2020-10-26 16:45:24 +0000 UTC   map[watch-this-configmap:from-resource-version] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Oct 26 16:45:24.701: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version  watch-6898 /api/v1/namespaces/watch-6898/configmaps/e2e-watch-test-resource-version 4dcf137b-3281-464a-9125-20d64ac86507 72902 0 2020-10-26 16:45:24 +0000 UTC   map[watch-this-configmap:from-resource-version] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 16:45:24.701: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-6898" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance]","total":280,"completed":160,"skipped":2424,"failed":0}
+SSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance]","total":280,"completed":152,"skipped":2577,"failed":0}
-SSSSS
-------------------------------
-[sig-storage] Downward API volume 
-  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+[sig-storage] EmptyDir volumes 
+  should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:39:32.610: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
+Oct 26 16:45:24.742: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+[It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:39:33.802: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fd22f29e-eae8-4e34-9c02-bd6ba5d7a79f" in namespace "downward-api-155" to be "success or failure"
-Sep 21 16:39:33.847: INFO: Pod "downwardapi-volume-fd22f29e-eae8-4e34-9c02-bd6ba5d7a79f": Phase="Pending", Reason="", readiness=false. Elapsed: 45.464029ms
-Sep 21 16:39:35.876: INFO: Pod "downwardapi-volume-fd22f29e-eae8-4e34-9c02-bd6ba5d7a79f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.074252045s
-Sep 21 16:39:37.902: INFO: Pod "downwardapi-volume-fd22f29e-eae8-4e34-9c02-bd6ba5d7a79f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.100493926s
+STEP: Creating a pod to test emptydir 0666 on tmpfs
+Oct 26 16:45:24.961: INFO: Waiting up to 5m0s for pod "pod-2d442678-8a59-4be8-8384-5fb68616c0a4" in namespace "emptydir-8163" to be "success or failure"
+Oct 26 16:45:24.988: INFO: Pod "pod-2d442678-8a59-4be8-8384-5fb68616c0a4": Phase="Pending", Reason="", readiness=false. Elapsed: 26.819912ms
+Oct 26 16:45:27.011: INFO: Pod "pod-2d442678-8a59-4be8-8384-5fb68616c0a4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.04952596s
 STEP: Saw pod success
-Sep 21 16:39:37.902: INFO: Pod "downwardapi-volume-fd22f29e-eae8-4e34-9c02-bd6ba5d7a79f" satisfied condition "success or failure"
-Sep 21 16:39:37.930: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-fd22f29e-eae8-4e34-9c02-bd6ba5d7a79f container client-container: 
+Oct 26 16:45:27.011: INFO: Pod "pod-2d442678-8a59-4be8-8384-5fb68616c0a4" satisfied condition "success or failure"
+Oct 26 16:45:27.027: INFO: Trying to get logs from node 10.72.119.74 pod pod-2d442678-8a59-4be8-8384-5fb68616c0a4 container test-container: 
 STEP: delete the pod
-Sep 21 16:39:38.169: INFO: Waiting for pod downwardapi-volume-fd22f29e-eae8-4e34-9c02-bd6ba5d7a79f to disappear
-Sep 21 16:39:38.199: INFO: Pod downwardapi-volume-fd22f29e-eae8-4e34-9c02-bd6ba5d7a79f no longer exists
-[AfterEach] [sig-storage] Downward API volume
+Oct 26 16:45:27.121: INFO: Waiting for pod pod-2d442678-8a59-4be8-8384-5fb68616c0a4 to disappear
+Oct 26 16:45:27.136: INFO: Pod pod-2d442678-8a59-4be8-8384-5fb68616c0a4 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:39:38.199: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-155" for this suite.
-
-• [SLOW TEST:5.686 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:35
-  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":153,"skipped":2582,"failed":0}
+Oct 26 16:45:27.137: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-8163" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":161,"skipped":2445,"failed":0}
 SSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
+[sig-node] Downward API 
+  should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
+[BeforeEach] [sig-node] Downward API
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:39:38.296: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename configmap
+Oct 26 16:45:27.176: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename downward-api
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
+[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name configmap-test-volume-map-beb41a6c-a804-4f12-bf21-851d86b9ad01
-STEP: Creating a pod to test consume configMaps
-Sep 21 16:39:38.862: INFO: Waiting up to 5m0s for pod "pod-configmaps-93c7ca53-4c0a-4743-96f4-86ace4a003ac" in namespace "configmap-1516" to be "success or failure"
-Sep 21 16:39:38.887: INFO: Pod "pod-configmaps-93c7ca53-4c0a-4743-96f4-86ace4a003ac": Phase="Pending", Reason="", readiness=false. Elapsed: 25.090867ms
-Sep 21 16:39:40.912: INFO: Pod "pod-configmaps-93c7ca53-4c0a-4743-96f4-86ace4a003ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.049707055s
-Sep 21 16:39:42.937: INFO: Pod "pod-configmaps-93c7ca53-4c0a-4743-96f4-86ace4a003ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.075357645s
+STEP: Creating a pod to test downward api env vars
+Oct 26 16:45:27.372: INFO: Waiting up to 5m0s for pod "downward-api-e9397d3d-739c-40ba-b710-b63f9736675f" in namespace "downward-api-8899" to be "success or failure"
+Oct 26 16:45:27.388: INFO: Pod "downward-api-e9397d3d-739c-40ba-b710-b63f9736675f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.364416ms
+Oct 26 16:45:29.405: INFO: Pod "downward-api-e9397d3d-739c-40ba-b710-b63f9736675f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.032444968s
+Oct 26 16:45:31.426: INFO: Pod "downward-api-e9397d3d-739c-40ba-b710-b63f9736675f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.054107049s
 STEP: Saw pod success
-Sep 21 16:39:42.938: INFO: Pod "pod-configmaps-93c7ca53-4c0a-4743-96f4-86ace4a003ac" satisfied condition "success or failure"
-Sep 21 16:39:42.960: INFO: Trying to get logs from node 10.241.51.147 pod pod-configmaps-93c7ca53-4c0a-4743-96f4-86ace4a003ac container configmap-volume-test: 
+Oct 26 16:45:31.427: INFO: Pod "downward-api-e9397d3d-739c-40ba-b710-b63f9736675f" satisfied condition "success or failure"
+Oct 26 16:45:31.445: INFO: Trying to get logs from node 10.72.119.74 pod downward-api-e9397d3d-739c-40ba-b710-b63f9736675f container dapi-container: 
 STEP: delete the pod
-Sep 21 16:39:43.097: INFO: Waiting for pod pod-configmaps-93c7ca53-4c0a-4743-96f4-86ace4a003ac to disappear
-Sep 21 16:39:43.119: INFO: Pod pod-configmaps-93c7ca53-4c0a-4743-96f4-86ace4a003ac no longer exists
-[AfterEach] [sig-storage] ConfigMap
+Oct 26 16:45:31.538: INFO: Waiting for pod downward-api-e9397d3d-739c-40ba-b710-b63f9736675f to disappear
+Oct 26 16:45:31.554: INFO: Pod downward-api-e9397d3d-739c-40ba-b710-b63f9736675f no longer exists
+[AfterEach] [sig-node] Downward API
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:39:43.119: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-1516" for this suite.
-•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":154,"skipped":2604,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:45:31.555: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-8899" for this suite.
+•{"msg":"PASSED [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]","total":280,"completed":162,"skipped":2467,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Probing container 
-  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+[sig-network] DNS 
+  should provide DNS for pods for Hostname [LinuxOnly] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Probing container
+[BeforeEach] [sig-network] DNS
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:39:43.187: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-probe
+Oct 26 16:45:31.595: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename dns
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+[It] should provide DNS for pods for Hostname [LinuxOnly] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod busybox-5ff7a26f-36d1-49b4-9bc6-3f2cc80cd0c5 in namespace container-probe-6862
-Sep 21 16:39:45.678: INFO: Started pod busybox-5ff7a26f-36d1-49b4-9bc6-3f2cc80cd0c5 in namespace container-probe-6862
-STEP: checking the pod's current state and verifying that restartCount is present
-Sep 21 16:39:45.706: INFO: Initial restart count of pod busybox-5ff7a26f-36d1-49b4-9bc6-3f2cc80cd0c5 is 0
-Sep 21 16:40:40.911: INFO: Restart count of pod container-probe-6862/busybox-5ff7a26f-36d1-49b4-9bc6-3f2cc80cd0c5 is now 1 (55.205036462s elapsed)
+STEP: Creating a test headless service
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-181.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-2.dns-test-service-2.dns-181.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/wheezy_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-181.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
+
+STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-181.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-2.dns-test-service-2.dns-181.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/jessie_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-181.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+
+STEP: creating a pod to probe DNS
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Oct 26 16:45:36.064: INFO: DNS probes using dns-181/dns-test-cc0567eb-1a8c-48fc-9621-c8902190780b succeeded
+
 STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
+STEP: deleting the test headless service
+[AfterEach] [sig-network] DNS
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:40:41.012: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-6862" for this suite.
-
-• [SLOW TEST:57.899 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [k8s.io] Probing container should be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":280,"completed":155,"skipped":2631,"failed":0}
-SSSSSSSSSSSSS
+Oct 26 16:45:36.181: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-181" for this suite.
+•{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Hostname [LinuxOnly] [Conformance]","total":280,"completed":163,"skipped":2508,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should mutate custom resource [Conformance]
+[k8s.io] InitContainer [NodeConformance] 
+  should invoke init containers on a RestartAlways pod [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:40:41.086: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename webhook
+Oct 26 16:45:36.223: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename init-container
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Sep 21 16:40:43.007: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303242, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303242, loc:(*time.Location)(0x7db5bc0)}}, Reason:"NewReplicaSetCreated", Message:"Created new replica set \"sample-webhook-deployment-5f65f8c764\""}, v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303242, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303242, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Sep 21 16:40:46.451: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should mutate custom resource [Conformance]
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
+[It] should invoke init containers on a RestartAlways pod [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:40:46.488: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Registering the mutating webhook for custom resource e2e-test-webhook-1645-crds.webhook.example.com via the AdmissionRegistration API
-STEP: Creating a custom resource that should be mutated by the webhook
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+STEP: creating the pod
+Oct 26 16:45:36.412: INFO: PodSpec: initContainers in spec.initContainers
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:40:48.009: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-4080" for this suite.
-STEP: Destroying namespace "webhook-4080-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
-
-• [SLOW TEST:7.433 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should mutate custom resource [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance]","total":280,"completed":156,"skipped":2644,"failed":0}
-SSSSSSSSSSSSS
+Oct 26 16:45:41.049: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "init-container-5747" for this suite.
+•{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance]","total":280,"completed":164,"skipped":2562,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl replace 
-  should update a single-container pod's image  [Conformance]
+[sig-storage] EmptyDir volumes 
+  volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:40:48.520: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubectl
+Oct 26 16:45:41.089: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[BeforeEach] Kubectl replace
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1897
-[It] should update a single-container pod's image  [Conformance]
+[It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: running the image docker.io/library/httpd:2.4.38-alpine
-Sep 21 16:40:48.864: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 run e2e-test-httpd-pod --generator=run-pod/v1 --image=docker.io/library/httpd:2.4.38-alpine --labels=run=e2e-test-httpd-pod --namespace=kubectl-2977'
-Sep 21 16:40:49.194: INFO: stderr: ""
-Sep 21 16:40:49.194: INFO: stdout: "pod/e2e-test-httpd-pod created\n"
-STEP: verifying the pod e2e-test-httpd-pod is running
-STEP: verifying the pod e2e-test-httpd-pod was created
-Sep 21 16:40:54.245: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pod e2e-test-httpd-pod --namespace=kubectl-2977 -o json'
-Sep 21 16:40:54.361: INFO: stderr: ""
-Sep 21 16:40:54.361: INFO: stdout: "{\n    \"apiVersion\": \"v1\",\n    \"kind\": \"Pod\",\n    \"metadata\": {\n        \"annotations\": {\n            \"cni.projectcalico.org/podIP\": \"172.30.237.93/32\",\n            \"cni.projectcalico.org/podIPs\": \"172.30.237.93/32\",\n            \"k8s.v1.cni.cncf.io/networks-status\": \"[{\\n    \\\"name\\\": \\\"k8s-pod-network\\\",\\n    \\\"ips\\\": [\\n        \\\"172.30.237.93\\\"\\n    ],\\n    \\\"dns\\\": {}\\n}]\",\n            \"openshift.io/scc\": \"anyuid\"\n        },\n        \"creationTimestamp\": \"2020-09-21T16:40:49Z\",\n        \"labels\": {\n            \"run\": \"e2e-test-httpd-pod\"\n        },\n        \"name\": \"e2e-test-httpd-pod\",\n        \"namespace\": \"kubectl-2977\",\n        \"resourceVersion\": \"75629\",\n        \"selfLink\": \"/api/v1/namespaces/kubectl-2977/pods/e2e-test-httpd-pod\",\n        \"uid\": \"1884901e-0771-49c9-921f-a838b5c238ba\"\n    },\n    \"spec\": {\n        \"containers\": [\n            {\n                \"image\": \"docker.io/library/httpd:2.4.38-alpine\",\n                \"imagePullPolicy\": \"IfNotPresent\",\n                \"name\": \"e2e-test-httpd-pod\",\n                \"resources\": {},\n                \"securityContext\": {\n                    \"capabilities\": {\n                        \"drop\": [\n                            \"MKNOD\"\n                        ]\n                    }\n                },\n                \"terminationMessagePath\": \"/dev/termination-log\",\n                \"terminationMessagePolicy\": \"File\",\n                \"volumeMounts\": [\n                    {\n                        \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n                        \"name\": \"default-token-4mdts\",\n                        \"readOnly\": true\n                    }\n                ]\n            }\n        ],\n        \"dnsPolicy\": \"ClusterFirst\",\n        \"enableServiceLinks\": true,\n        \"nodeName\": \"10.241.51.147\",\n        \"priority\": 0,\n        \"restartPolicy\": \"Always\",\n        \"schedulerName\": \"default-scheduler\",\n        \"securityContext\": {\n            \"seLinuxOptions\": {\n                \"level\": \"s0:c58,c42\"\n            }\n        },\n        \"serviceAccount\": \"default\",\n        \"serviceAccountName\": \"default\",\n        \"terminationGracePeriodSeconds\": 30,\n        \"tolerations\": [\n            {\n                \"effect\": \"NoExecute\",\n                \"key\": \"node.kubernetes.io/not-ready\",\n                \"operator\": \"Exists\",\n                \"tolerationSeconds\": 300\n            },\n            {\n                \"effect\": \"NoExecute\",\n                \"key\": \"node.kubernetes.io/unreachable\",\n                \"operator\": \"Exists\",\n                \"tolerationSeconds\": 300\n            }\n        ],\n        \"volumes\": [\n            {\n                \"name\": \"default-token-4mdts\",\n                \"secret\": {\n                    \"defaultMode\": 420,\n                    \"secretName\": \"default-token-4mdts\"\n                }\n            }\n        ]\n    },\n    \"status\": {\n        \"conditions\": [\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2020-09-21T16:40:49Z\",\n                \"status\": \"True\",\n                \"type\": \"Initialized\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2020-09-21T16:40:51Z\",\n                \"status\": \"True\",\n                \"type\": \"Ready\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2020-09-21T16:40:51Z\",\n                \"status\": \"True\",\n                \"type\": \"ContainersReady\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2020-09-21T16:40:49Z\",\n                \"status\": \"True\",\n                \"type\": \"PodScheduled\"\n            }\n        ],\n        \"containerStatuses\": [\n            {\n                \"containerID\": \"cri-o://03e0ba05b4bf28afeb763fd36847a6dd3566a5a2e92b37c64475e9487ce08a40\",\n                \"image\": \"docker.io/library/httpd:2.4.38-alpine\",\n                \"imageID\": \"docker.io/library/httpd@sha256:6feb0ea7b0967367da66e8d58ba813fde32bdb92f63bfc21a9e170d211539db4\",\n                \"lastState\": {},\n                \"name\": \"e2e-test-httpd-pod\",\n                \"ready\": true,\n                \"restartCount\": 0,\n                \"started\": true,\n                \"state\": {\n                    \"running\": {\n                        \"startedAt\": \"2020-09-21T16:40:50Z\"\n                    }\n                }\n            }\n        ],\n        \"hostIP\": \"10.241.51.147\",\n        \"phase\": \"Running\",\n        \"podIP\": \"172.30.237.93\",\n        \"podIPs\": [\n            {\n                \"ip\": \"172.30.237.93\"\n            }\n        ],\n        \"qosClass\": \"BestEffort\",\n        \"startTime\": \"2020-09-21T16:40:49Z\"\n    }\n}\n"
-STEP: replace the image in the pod
-Sep 21 16:40:54.362: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 replace -f - --namespace=kubectl-2977'
-Sep 21 16:40:55.027: INFO: stderr: ""
-Sep 21 16:40:55.027: INFO: stdout: "pod/e2e-test-httpd-pod replaced\n"
-STEP: verifying the pod e2e-test-httpd-pod has the right image docker.io/library/busybox:1.29
-[AfterEach] Kubectl replace
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1902
-Sep 21 16:40:55.055: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete pods e2e-test-httpd-pod --namespace=kubectl-2977'
-Sep 21 16:40:57.550: INFO: stderr: ""
-Sep 21 16:40:57.550: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n"
-[AfterEach] [sig-cli] Kubectl client
+STEP: Creating a pod to test emptydir volume type on node default medium
+Oct 26 16:45:41.300: INFO: Waiting up to 5m0s for pod "pod-929f779f-da4b-4ec2-926f-dc1d5e02c2bf" in namespace "emptydir-817" to be "success or failure"
+Oct 26 16:45:41.316: INFO: Pod "pod-929f779f-da4b-4ec2-926f-dc1d5e02c2bf": Phase="Pending", Reason="", readiness=false. Elapsed: 15.74523ms
+Oct 26 16:45:43.334: INFO: Pod "pod-929f779f-da4b-4ec2-926f-dc1d5e02c2bf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.034441072s
+STEP: Saw pod success
+Oct 26 16:45:43.335: INFO: Pod "pod-929f779f-da4b-4ec2-926f-dc1d5e02c2bf" satisfied condition "success or failure"
+Oct 26 16:45:43.350: INFO: Trying to get logs from node 10.72.119.74 pod pod-929f779f-da4b-4ec2-926f-dc1d5e02c2bf container test-container: 
+STEP: delete the pod
+Oct 26 16:45:43.436: INFO: Waiting for pod pod-929f779f-da4b-4ec2-926f-dc1d5e02c2bf to disappear
+Oct 26 16:45:43.451: INFO: Pod pod-929f779f-da4b-4ec2-926f-dc1d5e02c2bf no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:40:57.550: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-2977" for this suite.
+Oct 26 16:45:43.451: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-817" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":165,"skipped":2586,"failed":0}
 
-• [SLOW TEST:9.107 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Kubectl replace
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1893
-    should update a single-container pod's image  [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Kubectl replace should update a single-container pod's image  [Conformance]","total":280,"completed":157,"skipped":2657,"failed":0}
-SSSS
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should be able to deny attaching pod [Conformance]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  works for CRD preserving unknown fields at the schema root [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:40:57.627: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename webhook
+Oct 26 16:45:43.487: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename crd-publish-openapi
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Sep 21 16:40:58.881: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303258, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303258, loc:(*time.Location)(0x7db5bc0)}}, Reason:"NewReplicaSetCreated", Message:"Created new replica set \"sample-webhook-deployment-5f65f8c764\""}, v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303258, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303258, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:41:00.918: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303258, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303258, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303258, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303258, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Sep 21 16:41:04.005: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should be able to deny attaching pod [Conformance]
+[It] works for CRD preserving unknown fields at the schema root [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering the webhook via the AdmissionRegistration API
-STEP: create a pod
-STEP: 'kubectl attach' the pod, should be denied by the webhook
-Sep 21 16:41:08.507: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 attach --namespace=webhook-3188 to-be-attached-pod -i -c=container1'
-Sep 21 16:41:08.733: INFO: rc: 1
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+Oct 26 16:45:43.643: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: client-side validation (kubectl create and apply) allows request with any unknown properties
+Oct 26 16:45:51.906: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-4642 create -f -'
+Oct 26 16:46:00.255: INFO: stderr: ""
+Oct 26 16:46:00.255: INFO: stdout: "e2e-test-crd-publish-openapi-4917-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n"
+Oct 26 16:46:00.256: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-4642 delete e2e-test-crd-publish-openapi-4917-crds test-cr'
+Oct 26 16:46:00.493: INFO: stderr: ""
+Oct 26 16:46:00.493: INFO: stdout: "e2e-test-crd-publish-openapi-4917-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n"
+Oct 26 16:46:00.494: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-4642 apply -f -'
+Oct 26 16:46:01.185: INFO: stderr: ""
+Oct 26 16:46:01.185: INFO: stdout: "e2e-test-crd-publish-openapi-4917-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n"
+Oct 26 16:46:01.185: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-4642 delete e2e-test-crd-publish-openapi-4917-crds test-cr'
+Oct 26 16:46:01.365: INFO: stderr: ""
+Oct 26 16:46:01.365: INFO: stdout: "e2e-test-crd-publish-openapi-4917-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n"
+STEP: kubectl explain works to explain CR
+Oct 26 16:46:01.365: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 explain e2e-test-crd-publish-openapi-4917-crds'
+Oct 26 16:46:01.710: INFO: stderr: ""
+Oct 26 16:46:01.710: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-4917-crd\nVERSION:  crd-publish-openapi-test-unknown-at-root.example.com/v1\n\nDESCRIPTION:\n     \n"
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:41:08.834: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-3188" for this suite.
-STEP: Destroying namespace "webhook-3188-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+Oct 26 16:46:09.708: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-4642" for this suite.
 
-• [SLOW TEST:11.778 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+• [SLOW TEST:26.259 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should be able to deny attaching pod [Conformance]
+  works for CRD preserving unknown fields at the schema root [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance]","total":280,"completed":158,"skipped":2661,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance]","total":280,"completed":166,"skipped":2586,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition 
-  creating/deleting custom resource definition objects works  [Conformance]
+[sig-storage] Downward API volume 
+  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+[BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:41:09.406: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename custom-resource-definition
+Oct 26 16:46:09.747: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename downward-api
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] creating/deleting custom resource definition objects works  [Conformance]
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:41:09.741: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+STEP: Creating a pod to test downward API volume plugin
+Oct 26 16:46:09.932: INFO: Waiting up to 5m0s for pod "downwardapi-volume-ed8a8599-5506-4fc6-bebd-95d1a7f68c10" in namespace "downward-api-1970" to be "success or failure"
+Oct 26 16:46:09.947: INFO: Pod "downwardapi-volume-ed8a8599-5506-4fc6-bebd-95d1a7f68c10": Phase="Pending", Reason="", readiness=false. Elapsed: 14.201723ms
+Oct 26 16:46:11.963: INFO: Pod "downwardapi-volume-ed8a8599-5506-4fc6-bebd-95d1a7f68c10": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030129666s
+Oct 26 16:46:13.981: INFO: Pod "downwardapi-volume-ed8a8599-5506-4fc6-bebd-95d1a7f68c10": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.048221276s
+STEP: Saw pod success
+Oct 26 16:46:13.981: INFO: Pod "downwardapi-volume-ed8a8599-5506-4fc6-bebd-95d1a7f68c10" satisfied condition "success or failure"
+Oct 26 16:46:14.000: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-ed8a8599-5506-4fc6-bebd-95d1a7f68c10 container client-container: 
+STEP: delete the pod
+Oct 26 16:46:14.097: INFO: Waiting for pod downwardapi-volume-ed8a8599-5506-4fc6-bebd-95d1a7f68c10 to disappear
+Oct 26 16:46:14.113: INFO: Pod downwardapi-volume-ed8a8599-5506-4fc6-bebd-95d1a7f68c10 no longer exists
+[AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:41:10.414: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "custom-resource-definition-977" for this suite.
-•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works  [Conformance]","total":280,"completed":159,"skipped":2687,"failed":0}
-SSSSS
+Oct 26 16:46:14.113: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-1970" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":167,"skipped":2622,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should be able to deny pod and configmap creation [Conformance]
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with downward pod [LinuxOnly] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+[BeforeEach] [sig-storage] Subpath
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:41:10.646: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename webhook
+Oct 26 16:46:14.156: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename subpath
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Sep 21 16:41:12.131: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303272, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303272, loc:(*time.Location)(0x7db5bc0)}}, Reason:"NewReplicaSetCreated", Message:"Created new replica set \"sample-webhook-deployment-5f65f8c764\""}, v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303272, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303272, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:41:14.170: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303272, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303272, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303272, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303272, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Sep 21 16:41:17.238: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should be able to deny pod and configmap creation [Conformance]
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
+STEP: Setting up data
+[It] should support subpaths with downward pod [LinuxOnly] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering the webhook via the AdmissionRegistration API
-STEP: create a pod that should be denied by the webhook
-STEP: create a pod that causes the webhook to hang
-STEP: create a configmap that should be denied by the webhook
-STEP: create a configmap that should be admitted by the webhook
-STEP: update (PUT) the admitted configmap to a non-compliant one should be rejected by the webhook
-STEP: update (PATCH) the admitted configmap to a non-compliant one should be rejected by the webhook
-STEP: create a namespace that bypass the webhook
-STEP: create a configmap that violates the webhook policy but is in a whitelisted namespace
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+STEP: Creating pod pod-subpath-test-downwardapi-hkwd
+STEP: Creating a pod to test atomic-volume-subpath
+Oct 26 16:46:14.392: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-hkwd" in namespace "subpath-4693" to be "success or failure"
+Oct 26 16:46:14.411: INFO: Pod "pod-subpath-test-downwardapi-hkwd": Phase="Pending", Reason="", readiness=false. Elapsed: 19.472225ms
+Oct 26 16:46:16.431: INFO: Pod "pod-subpath-test-downwardapi-hkwd": Phase="Running", Reason="", readiness=true. Elapsed: 2.039691076s
+Oct 26 16:46:18.453: INFO: Pod "pod-subpath-test-downwardapi-hkwd": Phase="Running", Reason="", readiness=true. Elapsed: 4.061136898s
+Oct 26 16:46:20.469: INFO: Pod "pod-subpath-test-downwardapi-hkwd": Phase="Running", Reason="", readiness=true. Elapsed: 6.077551987s
+Oct 26 16:46:22.486: INFO: Pod "pod-subpath-test-downwardapi-hkwd": Phase="Running", Reason="", readiness=true. Elapsed: 8.093991632s
+Oct 26 16:46:24.507: INFO: Pod "pod-subpath-test-downwardapi-hkwd": Phase="Running", Reason="", readiness=true. Elapsed: 10.115674198s
+Oct 26 16:46:26.523: INFO: Pod "pod-subpath-test-downwardapi-hkwd": Phase="Running", Reason="", readiness=true. Elapsed: 12.131572525s
+Oct 26 16:46:28.545: INFO: Pod "pod-subpath-test-downwardapi-hkwd": Phase="Running", Reason="", readiness=true. Elapsed: 14.153132728s
+Oct 26 16:46:30.561: INFO: Pod "pod-subpath-test-downwardapi-hkwd": Phase="Running", Reason="", readiness=true. Elapsed: 16.169474587s
+Oct 26 16:46:32.579: INFO: Pod "pod-subpath-test-downwardapi-hkwd": Phase="Running", Reason="", readiness=true. Elapsed: 18.187116157s
+Oct 26 16:46:34.594: INFO: Pod "pod-subpath-test-downwardapi-hkwd": Phase="Running", Reason="", readiness=true. Elapsed: 20.202469507s
+Oct 26 16:46:36.613: INFO: Pod "pod-subpath-test-downwardapi-hkwd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.220871166s
+STEP: Saw pod success
+Oct 26 16:46:36.613: INFO: Pod "pod-subpath-test-downwardapi-hkwd" satisfied condition "success or failure"
+Oct 26 16:46:36.648: INFO: Trying to get logs from node 10.72.119.74 pod pod-subpath-test-downwardapi-hkwd container test-container-subpath-downwardapi-hkwd: 
+STEP: delete the pod
+Oct 26 16:46:36.771: INFO: Waiting for pod pod-subpath-test-downwardapi-hkwd to disappear
+Oct 26 16:46:36.787: INFO: Pod pod-subpath-test-downwardapi-hkwd no longer exists
+STEP: Deleting pod pod-subpath-test-downwardapi-hkwd
+Oct 26 16:46:36.787: INFO: Deleting pod "pod-subpath-test-downwardapi-hkwd" in namespace "subpath-4693"
+[AfterEach] [sig-storage] Subpath
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:41:28.186: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-8993" for this suite.
-STEP: Destroying namespace "webhook-8993-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+Oct 26 16:46:36.804: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-4693" for this suite.
 
-• [SLOW TEST:18.121 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should be able to deny pod and configmap creation [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:22.689 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
+  Atomic writer volumes
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
+    should support subpaths with downward pod [LinuxOnly] [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance]","total":280,"completed":160,"skipped":2692,"failed":0}
-SSSSSSSSSSS
+{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [LinuxOnly] [Conformance]","total":280,"completed":168,"skipped":2646,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]
+[sig-apps] Job 
+  should delete a job [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] StatefulSet
+[BeforeEach] [sig-apps] Job
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:41:28.767: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename statefulset
+Oct 26 16:46:36.850: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename job
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
-STEP: Creating service test in namespace statefulset-4716
-[It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]
+[It] should delete a job [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Initializing watcher for selector baz=blah,foo=bar
-STEP: Creating stateful set ss in namespace statefulset-4716
-STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-4716
-Sep 21 16:41:29.116: INFO: Found 0 stateful pods, waiting for 1
-Sep 21 16:41:39.143: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod
-Sep 21 16:41:39.169: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-4716 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Sep 21 16:41:39.500: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Sep 21 16:41:39.500: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Sep 21 16:41:39.500: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Sep 21 16:41:39.524: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
-Sep 21 16:41:49.550: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
-Sep 21 16:41:49.550: INFO: Waiting for statefulset status.replicas updated to 0
-Sep 21 16:41:49.644: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999997798s
-Sep 21 16:41:50.668: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.978578365s
-Sep 21 16:41:51.692: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.954166515s
-Sep 21 16:41:52.716: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.930027405s
-Sep 21 16:41:53.742: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.906729035s
-Sep 21 16:41:54.780: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.880276238s
-Sep 21 16:41:55.805: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.842308946s
-Sep 21 16:41:56.833: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.817836624s
-Sep 21 16:41:57.852: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.789728288s
-Sep 21 16:41:58.882: INFO: Verifying statefulset ss doesn't scale past 1 for another 770.064036ms
-STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-4716
-Sep 21 16:41:59.907: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-4716 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Sep 21 16:42:00.402: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Sep 21 16:42:00.402: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Sep 21 16:42:00.402: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Sep 21 16:42:00.427: INFO: Found 1 stateful pods, waiting for 3
-Sep 21 16:42:10.452: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-Sep 21 16:42:10.452: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
-Sep 21 16:42:10.452: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Verifying that stateful set ss was scaled up in order
-STEP: Scale down will halt with unhealthy stateful pod
-Sep 21 16:42:10.491: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-4716 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Sep 21 16:42:10.849: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Sep 21 16:42:10.849: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Sep 21 16:42:10.849: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Sep 21 16:42:10.849: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-4716 ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Sep 21 16:42:11.221: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Sep 21 16:42:11.221: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Sep 21 16:42:11.221: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Sep 21 16:42:11.221: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-4716 ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Sep 21 16:42:11.683: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Sep 21 16:42:11.683: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Sep 21 16:42:11.683: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Sep 21 16:42:11.683: INFO: Waiting for statefulset status.replicas updated to 0
-Sep 21 16:42:11.704: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2
-Sep 21 16:42:21.968: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
-Sep 21 16:42:21.969: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
-Sep 21 16:42:21.969: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
-Sep 21 16:42:22.065: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999997907s
-Sep 21 16:42:23.089: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.96587269s
-Sep 21 16:42:24.121: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.941729643s
-Sep 21 16:42:25.141: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.909858041s
-Sep 21 16:42:26.173: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.889175808s
-Sep 21 16:42:27.207: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.85707834s
-Sep 21 16:42:28.247: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.823723378s
-Sep 21 16:42:29.283: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.783296238s
-Sep 21 16:42:30.335: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.747783398s
-Sep 21 16:42:31.362: INFO: Verifying statefulset ss doesn't scale past 3 for another 695.771868ms
-STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-4716
-Sep 21 16:42:32.402: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-4716 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Sep 21 16:42:32.804: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Sep 21 16:42:32.804: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Sep 21 16:42:32.804: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Sep 21 16:42:32.804: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-4716 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Sep 21 16:42:33.237: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Sep 21 16:42:33.237: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Sep 21 16:42:33.237: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Sep 21 16:42:33.237: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=statefulset-4716 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Sep 21 16:42:33.648: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Sep 21 16:42:33.648: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Sep 21 16:42:33.648: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Sep 21 16:42:33.648: INFO: Scaling statefulset ss to 0
-STEP: Verifying that stateful set ss was scaled down in reverse order
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
-Sep 21 16:43:03.752: INFO: Deleting all statefulset in ns statefulset-4716
-Sep 21 16:43:03.873: INFO: Scaling statefulset ss to 0
-Sep 21 16:43:03.952: INFO: Waiting for statefulset status.replicas updated to 0
-Sep 21 16:43:03.971: INFO: Deleting statefulset ss
-[AfterEach] [sig-apps] StatefulSet
+STEP: Creating a job
+STEP: Ensuring active pods == parallelism
+STEP: delete a job
+STEP: deleting Job.batch foo in namespace job-4809, will wait for the garbage collector to delete the pods
+Oct 26 16:46:41.142: INFO: Deleting Job.batch foo took: 43.204431ms
+Oct 26 16:46:41.543: INFO: Terminating Job.batch foo pods took: 400.411838ms
+STEP: Ensuring job was deleted
+[AfterEach] [sig-apps] Job
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:43:04.118: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-4716" for this suite.
+Oct 26 16:47:16.860: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "job-4809" for this suite.
 
-• [SLOW TEST:95.456 seconds]
-[sig-apps] StatefulSet
+• [SLOW TEST:40.046 seconds]
+[sig-apps] Job
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-    Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should delete a job [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]","total":280,"completed":161,"skipped":2703,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-apps] Job should delete a job [Conformance]","total":280,"completed":169,"skipped":2683,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should honor timeout [Conformance]
+  should mutate custom resource with pruning [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:43:04.223: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 16:47:16.900: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
@@ -7912,1657 +7961,788 @@ STEP: Setting up server cert
 STEP: Create role binding to let webhook read extension-apiserver-authentication
 STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Sep 21 16:43:05.210: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created
-Sep 21 16:43:07.297: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303385, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303385, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303385, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303385, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 16:47:18.339: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 16:47:20.385: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327638, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327638, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327638, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327638, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Sep 21 16:43:10.416: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should honor timeout [Conformance]
+Oct 26 16:47:23.436: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should mutate custom resource with pruning [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Setting timeout (1s) shorter than webhook latency (5s)
-STEP: Registering slow webhook via the AdmissionRegistration API
-STEP: Request fails when timeout (1s) is shorter than slow webhook latency (5s)
-STEP: Having no error when timeout is shorter than webhook latency and failure policy is ignore
-STEP: Registering slow webhook via the AdmissionRegistration API
-STEP: Having no error when timeout is longer than webhook latency
-STEP: Registering slow webhook via the AdmissionRegistration API
-STEP: Having no error when timeout is empty (defaulted to 10s in v1)
-STEP: Registering slow webhook via the AdmissionRegistration API
+Oct 26 16:47:23.464: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Registering the mutating webhook for custom resource e2e-test-webhook-8476-crds.webhook.example.com via the AdmissionRegistration API
+STEP: Creating a custom resource that should be mutated by the webhook
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:43:23.479: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-1596" for this suite.
-STEP: Destroying namespace "webhook-1596-markers" for this suite.
+Oct 26 16:47:24.786: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-7227" for this suite.
+STEP: Destroying namespace "webhook-7227-markers" for this suite.
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:19.800 seconds]
+• [SLOW TEST:8.167 seconds]
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should honor timeout [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance]","total":280,"completed":162,"skipped":2750,"failed":0}
-SSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
-  should execute poststart http hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:43:24.024: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-lifecycle-hook
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] when create a pod with lifecycle hook
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
-STEP: create the container to handle the HTTPGet hook request.
-[It] should execute poststart http hook properly [NodeConformance] [Conformance]
+  should mutate custom resource with pruning [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the pod with lifecycle hook
-STEP: check poststart hook
-STEP: delete the pod with lifecycle hook
-Sep 21 16:43:32.854: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Sep 21 16:43:32.880: INFO: Pod pod-with-poststart-http-hook still exists
-Sep 21 16:43:34.880: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Sep 21 16:43:34.903: INFO: Pod pod-with-poststart-http-hook still exists
-Sep 21 16:43:36.880: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Sep 21 16:43:36.904: INFO: Pod pod-with-poststart-http-hook still exists
-Sep 21 16:43:38.880: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Sep 21 16:43:38.904: INFO: Pod pod-with-poststart-http-hook no longer exists
-[AfterEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:43:38.904: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-lifecycle-hook-7308" for this suite.
-
-• [SLOW TEST:14.956 seconds]
-[k8s.io] Container Lifecycle Hook
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  when create a pod with lifecycle hook
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
-    should execute poststart http hook properly [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance]","total":280,"completed":163,"skipped":2771,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Garbage collector 
-  should delete RS created by deployment when not orphaning [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:43:38.980: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename gc
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should delete RS created by deployment when not orphaning [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the deployment
-STEP: Wait for the Deployment to create new ReplicaSet
-STEP: delete the deployment
-STEP: wait for all rs to be garbage collected
-STEP: expected 0 pods, got 2 pods
-STEP: Gathering metrics
-Sep 21 16:43:40.114: INFO: For apiserver_request_total:
-For apiserver_request_latency_seconds:
-For apiserver_init_events_total:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
-
-W0921 16:43:40.114543      24 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:43:40.114: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-2798" for this suite.
-•{"msg":"PASSED [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance]","total":280,"completed":164,"skipped":2795,"failed":0}
-SSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance]","total":280,"completed":170,"skipped":2706,"failed":0}
+SSSSSS
 ------------------------------
-[sig-api-machinery] Watchers 
-  should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  works for CRD without validation schema [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Watchers
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:43:40.207: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename watch
+Oct 26 16:47:25.068: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename crd-publish-openapi
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
+[It] works for CRD without validation schema [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a watch on configmaps with a certain label
-STEP: creating a new configmap
-STEP: modifying the configmap once
-STEP: changing the label value of the configmap
-STEP: Expecting to observe a delete notification for the watched object
-Sep 21 16:43:40.597: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-8141 /api/v1/namespaces/watch-8141/configmaps/e2e-watch-test-label-changed a3c6c173-59af-46de-97b5-1b462a3d0b85 77427 0 2020-09-21 16:43:40 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Sep 21 16:43:40.597: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-8141 /api/v1/namespaces/watch-8141/configmaps/e2e-watch-test-label-changed a3c6c173-59af-46de-97b5-1b462a3d0b85 77430 0 2020-09-21 16:43:40 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-Sep 21 16:43:40.597: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-8141 /api/v1/namespaces/watch-8141/configmaps/e2e-watch-test-label-changed a3c6c173-59af-46de-97b5-1b462a3d0b85 77432 0 2020-09-21 16:43:40 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-STEP: modifying the configmap a second time
-STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements
-STEP: changing the label value of the configmap back
-STEP: modifying the configmap a third time
-STEP: deleting the configmap
-STEP: Expecting to observe an add notification for the watched object when the label value was restored
-Sep 21 16:43:50.761: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-8141 /api/v1/namespaces/watch-8141/configmaps/e2e-watch-test-label-changed a3c6c173-59af-46de-97b5-1b462a3d0b85 77565 0 2020-09-21 16:43:40 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Sep 21 16:43:50.762: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-8141 /api/v1/namespaces/watch-8141/configmaps/e2e-watch-test-label-changed a3c6c173-59af-46de-97b5-1b462a3d0b85 77566 0 2020-09-21 16:43:40 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
-Sep 21 16:43:50.762: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-8141 /api/v1/namespaces/watch-8141/configmaps/e2e-watch-test-label-changed a3c6c173-59af-46de-97b5-1b462a3d0b85 77567 0 2020-09-21 16:43:40 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
-[AfterEach] [sig-api-machinery] Watchers
+Oct 26 16:47:25.217: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: client-side validation (kubectl create and apply) allows request with any unknown properties
+Oct 26 16:47:35.677: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-1577 create -f -'
+Oct 26 16:47:36.555: INFO: stderr: ""
+Oct 26 16:47:36.556: INFO: stdout: "e2e-test-crd-publish-openapi-9711-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n"
+Oct 26 16:47:36.556: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-1577 delete e2e-test-crd-publish-openapi-9711-crds test-cr'
+Oct 26 16:47:36.750: INFO: stderr: ""
+Oct 26 16:47:36.750: INFO: stdout: "e2e-test-crd-publish-openapi-9711-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n"
+Oct 26 16:47:36.750: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-1577 apply -f -'
+Oct 26 16:47:37.127: INFO: stderr: ""
+Oct 26 16:47:37.128: INFO: stdout: "e2e-test-crd-publish-openapi-9711-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n"
+Oct 26 16:47:37.128: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=crd-publish-openapi-1577 delete e2e-test-crd-publish-openapi-9711-crds test-cr'
+Oct 26 16:47:37.357: INFO: stderr: ""
+Oct 26 16:47:37.357: INFO: stdout: "e2e-test-crd-publish-openapi-9711-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n"
+STEP: kubectl explain works to explain CR without validation schema
+Oct 26 16:47:37.357: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 explain e2e-test-crd-publish-openapi-9711-crds'
+Oct 26 16:47:37.899: INFO: stderr: ""
+Oct 26 16:47:37.899: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-9711-crd\nVERSION:  crd-publish-openapi-test-empty.example.com/v1\n\nDESCRIPTION:\n     \n"
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:43:50.762: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "watch-8141" for this suite.
+Oct 26 16:47:46.483: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-1577" for this suite.
 
-• [SLOW TEST:10.625 seconds]
-[sig-api-machinery] Watchers
+• [SLOW TEST:21.453 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
+  works for CRD without validation schema [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance]","total":280,"completed":165,"skipped":2804,"failed":0}
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance]","total":280,"completed":171,"skipped":2712,"failed":0}
 SSSSSS
 ------------------------------
-[k8s.io] Kubelet when scheduling a read only busybox container 
-  should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:43:50.833: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubelet-test
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
-[It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:43:53.551: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubelet-test-2946" for this suite.
-•{"msg":"PASSED [k8s.io] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":166,"skipped":2810,"failed":0}
-SS
-------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:43:53.833: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename secrets
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-e8e4e48c-e097-4b53-acb6-b39285e41098
-STEP: Creating a pod to test consume secrets
-Sep 21 16:43:54.251: INFO: Waiting up to 5m0s for pod "pod-secrets-a36d0758-61b6-40d8-bcb9-bfe099078f3e" in namespace "secrets-8219" to be "success or failure"
-Sep 21 16:43:54.276: INFO: Pod "pod-secrets-a36d0758-61b6-40d8-bcb9-bfe099078f3e": Phase="Pending", Reason="", readiness=false. Elapsed: 24.277134ms
-Sep 21 16:43:56.303: INFO: Pod "pod-secrets-a36d0758-61b6-40d8-bcb9-bfe099078f3e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.051798461s
-Sep 21 16:43:58.327: INFO: Pod "pod-secrets-a36d0758-61b6-40d8-bcb9-bfe099078f3e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.07541651s
-STEP: Saw pod success
-Sep 21 16:43:58.327: INFO: Pod "pod-secrets-a36d0758-61b6-40d8-bcb9-bfe099078f3e" satisfied condition "success or failure"
-Sep 21 16:43:58.355: INFO: Trying to get logs from node 10.241.51.147 pod pod-secrets-a36d0758-61b6-40d8-bcb9-bfe099078f3e container secret-volume-test: 
-STEP: delete the pod
-Sep 21 16:43:58.761: INFO: Waiting for pod pod-secrets-a36d0758-61b6-40d8-bcb9-bfe099078f3e to disappear
-Sep 21 16:43:58.789: INFO: Pod pod-secrets-a36d0758-61b6-40d8-bcb9-bfe099078f3e no longer exists
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:43:58.789: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-8219" for this suite.
-
-• [SLOW TEST:5.022 seconds]
-[sig-storage] Secrets
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
-  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":167,"skipped":2812,"failed":0}
-SSSSSSSSSS
-------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide container's memory request [NodeConformance] [Conformance]
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
+[BeforeEach] [sig-storage] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:43:58.855: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 16:47:46.521: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename configmap
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should provide container's memory request [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:43:59.259: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2dc76a12-d6e6-44a0-8869-f952ac4bb074" in namespace "projected-9141" to be "success or failure"
-Sep 21 16:43:59.300: INFO: Pod "downwardapi-volume-2dc76a12-d6e6-44a0-8869-f952ac4bb074": Phase="Pending", Reason="", readiness=false. Elapsed: 40.482276ms
-Sep 21 16:44:01.328: INFO: Pod "downwardapi-volume-2dc76a12-d6e6-44a0-8869-f952ac4bb074": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.068727285s
+STEP: Creating configMap with name configmap-test-volume-map-a15a0a86-21b7-44ea-8f73-f3788bc6adda
+STEP: Creating a pod to test consume configMaps
+Oct 26 16:47:46.747: INFO: Waiting up to 5m0s for pod "pod-configmaps-592ebd15-bf5b-4ce3-9329-443ddf7e935f" in namespace "configmap-4300" to be "success or failure"
+Oct 26 16:47:46.774: INFO: Pod "pod-configmaps-592ebd15-bf5b-4ce3-9329-443ddf7e935f": Phase="Pending", Reason="", readiness=false. Elapsed: 26.237923ms
+Oct 26 16:47:48.794: INFO: Pod "pod-configmaps-592ebd15-bf5b-4ce3-9329-443ddf7e935f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.047050719s
 STEP: Saw pod success
-Sep 21 16:44:01.328: INFO: Pod "downwardapi-volume-2dc76a12-d6e6-44a0-8869-f952ac4bb074" satisfied condition "success or failure"
-Sep 21 16:44:01.352: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-2dc76a12-d6e6-44a0-8869-f952ac4bb074 container client-container: 
+Oct 26 16:47:48.794: INFO: Pod "pod-configmaps-592ebd15-bf5b-4ce3-9329-443ddf7e935f" satisfied condition "success or failure"
+Oct 26 16:47:48.810: INFO: Trying to get logs from node 10.72.119.74 pod pod-configmaps-592ebd15-bf5b-4ce3-9329-443ddf7e935f container configmap-volume-test: 
 STEP: delete the pod
-Sep 21 16:44:01.689: INFO: Waiting for pod downwardapi-volume-2dc76a12-d6e6-44a0-8869-f952ac4bb074 to disappear
-Sep 21 16:44:01.713: INFO: Pod downwardapi-volume-2dc76a12-d6e6-44a0-8869-f952ac4bb074 no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
+Oct 26 16:47:48.903: INFO: Waiting for pod pod-configmaps-592ebd15-bf5b-4ce3-9329-443ddf7e935f to disappear
+Oct 26 16:47:48.921: INFO: Pod pod-configmaps-592ebd15-bf5b-4ce3-9329-443ddf7e935f no longer exists
+[AfterEach] [sig-storage] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:44:01.713: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-9141" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance]","total":280,"completed":168,"skipped":2822,"failed":0}
-SSSSSSSSSSSSSSSSS
+Oct 26 16:47:48.921: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-4300" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":172,"skipped":2718,"failed":0}
+SS
 ------------------------------
-[sig-cli] Kubectl client Update Demo 
-  should do a rolling update of a replication controller  [Conformance]
+[sig-network] Networking Granular Checks: Pods 
+  should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-network] Networking
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:44:01.837: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubectl
+Oct 26 16:47:48.959: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename pod-network-test
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[BeforeEach] Update Demo
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:330
-[It] should do a rolling update of a replication controller  [Conformance]
+[It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the initial replication controller
-Sep 21 16:44:02.114: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-1643'
-Sep 21 16:44:02.708: INFO: stderr: ""
-Sep 21 16:44:02.708: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Sep 21 16:44:02.708: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1643'
-Sep 21 16:44:02.871: INFO: stderr: ""
-Sep 21 16:44:02.871: INFO: stdout: "update-demo-nautilus-2phx7 update-demo-nautilus-nwkpx "
-Sep 21 16:44:02.871: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-2phx7 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1643'
-Sep 21 16:44:02.998: INFO: stderr: ""
-Sep 21 16:44:02.998: INFO: stdout: ""
-Sep 21 16:44:02.998: INFO: update-demo-nautilus-2phx7 is created but not running
-Sep 21 16:44:07.998: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1643'
-Sep 21 16:44:08.139: INFO: stderr: ""
-Sep 21 16:44:08.139: INFO: stdout: "update-demo-nautilus-2phx7 update-demo-nautilus-nwkpx "
-Sep 21 16:44:08.139: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-2phx7 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1643'
-Sep 21 16:44:08.272: INFO: stderr: ""
-Sep 21 16:44:08.272: INFO: stdout: "true"
-Sep 21 16:44:08.273: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-2phx7 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1643'
-Sep 21 16:44:08.392: INFO: stderr: ""
-Sep 21 16:44:08.392: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Sep 21 16:44:08.392: INFO: validating pod update-demo-nautilus-2phx7
-Sep 21 16:44:08.443: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Sep 21 16:44:08.443: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Sep 21 16:44:08.443: INFO: update-demo-nautilus-2phx7 is verified up and running
-Sep 21 16:44:08.443: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-nwkpx -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1643'
-Sep 21 16:44:08.575: INFO: stderr: ""
-Sep 21 16:44:08.575: INFO: stdout: "true"
-Sep 21 16:44:08.575: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-nwkpx -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1643'
-Sep 21 16:44:08.740: INFO: stderr: ""
-Sep 21 16:44:08.740: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Sep 21 16:44:08.740: INFO: validating pod update-demo-nautilus-nwkpx
-Sep 21 16:44:08.801: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Sep 21 16:44:08.801: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Sep 21 16:44:08.801: INFO: update-demo-nautilus-nwkpx is verified up and running
-STEP: rolling-update to new replication controller
-Sep 21 16:44:08.805: INFO: scanned /root for discovery docs: 
-Sep 21 16:44:08.805: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 rolling-update update-demo-nautilus --update-period=1s -f - --namespace=kubectl-1643'
-Sep 21 16:44:32.012: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n"
-Sep 21 16:44:32.012: INFO: stdout: "Created update-demo-kitten\nScaling up update-demo-kitten from 0 to 2, scaling down update-demo-nautilus from 2 to 0 (keep 2 pods available, don't exceed 3 pods)\nScaling update-demo-kitten up to 1\nScaling update-demo-nautilus down to 1\nScaling update-demo-kitten up to 2\nScaling update-demo-nautilus down to 0\nUpdate succeeded. Deleting old controller: update-demo-nautilus\nRenaming update-demo-kitten to update-demo-nautilus\nreplicationcontroller/update-demo-nautilus rolling updated\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Sep 21 16:44:32.012: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1643'
-Sep 21 16:44:32.146: INFO: stderr: ""
-Sep 21 16:44:32.146: INFO: stdout: "update-demo-kitten-cvg2h update-demo-kitten-m6wsh "
-Sep 21 16:44:32.147: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-kitten-cvg2h -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1643'
-Sep 21 16:44:32.281: INFO: stderr: ""
-Sep 21 16:44:32.281: INFO: stdout: "true"
-Sep 21 16:44:32.281: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-kitten-cvg2h -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1643'
-Sep 21 16:44:32.426: INFO: stderr: ""
-Sep 21 16:44:32.426: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
-Sep 21 16:44:32.426: INFO: validating pod update-demo-kitten-cvg2h
-Sep 21 16:44:32.482: INFO: got data: {
-  "image": "kitten.jpg"
-}
-
-Sep 21 16:44:32.482: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
-Sep 21 16:44:32.482: INFO: update-demo-kitten-cvg2h is verified up and running
-Sep 21 16:44:32.482: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-kitten-m6wsh -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1643'
-Sep 21 16:44:32.639: INFO: stderr: ""
-Sep 21 16:44:32.639: INFO: stdout: "true"
-Sep 21 16:44:32.640: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-kitten-m6wsh -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1643'
-Sep 21 16:44:32.775: INFO: stderr: ""
-Sep 21 16:44:32.775: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
-Sep 21 16:44:32.775: INFO: validating pod update-demo-kitten-m6wsh
-Sep 21 16:44:32.831: INFO: got data: {
-  "image": "kitten.jpg"
-}
-
-Sep 21 16:44:32.831: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
-Sep 21 16:44:32.831: INFO: update-demo-kitten-m6wsh is verified up and running
-[AfterEach] [sig-cli] Kubectl client
+STEP: Performing setup for networking test in namespace pod-network-test-8336
+STEP: creating a selector
+STEP: Creating the service pods in kubernetes
+Oct 26 16:47:49.082: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+STEP: Creating test pods
+Oct 26 16:48:11.603: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.30.194.117 8081 | grep -v '^\s*$'] Namespace:pod-network-test-8336 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:48:11.603: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:48:12.850: INFO: Found all expected endpoints: [netserver-0]
+Oct 26 16:48:12.869: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.30.103.230 8081 | grep -v '^\s*$'] Namespace:pod-network-test-8336 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:48:12.869: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:48:14.111: INFO: Found all expected endpoints: [netserver-1]
+Oct 26 16:48:14.131: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.30.54.162 8081 | grep -v '^\s*$'] Namespace:pod-network-test-8336 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 16:48:14.131: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 16:48:15.396: INFO: Found all expected endpoints: [netserver-2]
+[AfterEach] [sig-network] Networking
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:44:32.831: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-1643" for this suite.
+Oct 26 16:48:15.396: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pod-network-test-8336" for this suite.
 
-• [SLOW TEST:31.061 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Update Demo
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:328
-    should do a rolling update of a replication controller  [Conformance]
+• [SLOW TEST:26.478 seconds]
+[sig-network] Networking
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
+  Granular Checks: Pods
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
+    should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
     /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Update Demo should do a rolling update of a replication controller  [Conformance]","total":280,"completed":169,"skipped":2839,"failed":0}
-SSSSSSSSSSS
+{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":173,"skipped":2720,"failed":0}
+SSS
 ------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]
+[sig-api-machinery] Garbage collector 
+  should delete RS created by deployment when not orphaning [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+[BeforeEach] [sig-api-machinery] Garbage collector
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:44:32.898: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename sched-pred
+Oct 26 16:48:15.437: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename gc
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
-Sep 21 16:44:33.216: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-Sep 21 16:44:33.345: INFO: Waiting for terminating namespaces to be deleted...
-Sep 21 16:44:33.374: INFO: 
-Logging pods the kubelet thinks is on node 10.241.51.147 before test
-Sep 21 16:44:33.568: INFO: tuned-6qvw2 from openshift-cluster-node-tuning-operator started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.568: INFO: 	Container tuned ready: true, restart count 0
-Sep 21 16:44:33.568: INFO: calico-typha-6fc9ff8b5f-9pmb4 from calico-system started at 2020-09-21 14:13:14 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.568: INFO: 	Container calico-typha ready: true, restart count 0
-Sep 21 16:44:33.568: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-shgq6 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.568: INFO: 	Container sonobuoy-worker ready: true, restart count 1
-Sep 21 16:44:33.568: INFO: 	Container systemd-logs ready: true, restart count 0
-Sep 21 16:44:33.568: INFO: packageserver-75876789fb-w6q6s from openshift-operator-lifecycle-manager started at 2020-09-21 16:30:37 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.568: INFO: 	Container packageserver ready: true, restart count 0
-Sep 21 16:44:33.568: INFO: ibm-keepalived-watcher-h4dbn from kube-system started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.568: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Sep 21 16:44:33.568: INFO: calico-node-t2n87 from calico-system started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.569: INFO: 	Container calico-node ready: true, restart count 0
-Sep 21 16:44:33.569: INFO: ibmcloud-block-storage-driver-8r6lc from kube-system started at 2020-09-21 14:12:04 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.569: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
-Sep 21 16:44:33.569: INFO: openshift-kube-proxy-tvbvw from openshift-kube-proxy started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.569: INFO: 	Container kube-proxy ready: true, restart count 0
-Sep 21 16:44:33.569: INFO: multus-admission-controller-7rkpv from openshift-multus started at 2020-09-21 16:14:37 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.569: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.569: INFO: 	Container multus-admission-controller ready: true, restart count 0
-Sep 21 16:44:33.569: INFO: multus-7xt8r from openshift-multus started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.569: INFO: 	Container kube-multus ready: true, restart count 0
-Sep 21 16:44:33.569: INFO: busybox-readonly-fsfb5ad76e-2f9e-4e94-bcb2-e9de78f3006d from kubelet-test-2946 started at 2020-09-21 16:43:51 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.569: INFO: 	Container busybox-readonly-fsfb5ad76e-2f9e-4e94-bcb2-e9de78f3006d ready: false, restart count 0
-Sep 21 16:44:33.569: INFO: ibm-master-proxy-static-10.241.51.147 from kube-system started at 2020-09-21 14:11:47 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.569: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
-Sep 21 16:44:33.569: INFO: 	Container pause ready: true, restart count 0
-Sep 21 16:44:33.569: INFO: sonobuoy-e2e-job-36e1517c33ed4f81 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.570: INFO: 	Container e2e ready: true, restart count 0
-Sep 21 16:44:33.570: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Sep 21 16:44:33.570: INFO: node-exporter-46mwx from openshift-monitoring started at 2020-09-21 14:11:49 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.570: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.570: INFO: 	Container node-exporter ready: true, restart count 0
-Sep 21 16:44:33.570: INFO: node-ca-6r9nd from openshift-image-registry started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.570: INFO: 	Container node-ca ready: true, restart count 0
-Sep 21 16:44:33.570: INFO: dns-default-f6vtl from openshift-dns started at 2020-09-21 14:11:49 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.570: INFO: 	Container dns ready: true, restart count 0
-Sep 21 16:44:33.570: INFO: 	Container dns-node-resolver ready: true, restart count 0
-Sep 21 16:44:33.570: INFO: update-demo-kitten-cvg2h from kubectl-1643 started at 2020-09-21 16:44:09 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.570: INFO: 	Container update-demo ready: true, restart count 0
-Sep 21 16:44:33.570: INFO: sonobuoy from sonobuoy started at 2020-09-21 15:39:44 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.570: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-Sep 21 16:44:33.570: INFO: 
-Logging pods the kubelet thinks is on node 10.241.51.149 before test
-Sep 21 16:44:33.802: INFO: olm-operator-b5f47bf98-8d8cw from openshift-operator-lifecycle-manager started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.802: INFO: 	Container olm-operator ready: true, restart count 0
-Sep 21 16:44:33.802: INFO: console-8549dd6c4f-8vk99 from openshift-console started at 2020-09-21 16:14:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.802: INFO: 	Container console ready: true, restart count 0
-Sep 21 16:44:33.802: INFO: node-ca-bxmsf from openshift-image-registry started at 2020-09-21 14:09:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.802: INFO: 	Container node-ca ready: true, restart count 0
-Sep 21 16:44:33.802: INFO: prometheus-operator-8d96bcc68-ct2jw from openshift-monitoring started at 2020-09-21 16:14:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.802: INFO: 	Container prometheus-operator ready: true, restart count 0
-Sep 21 16:44:33.803: INFO: calico-node-9gqht from calico-system started at 2020-09-21 14:07:18 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.803: INFO: 	Container calico-node ready: true, restart count 0
-Sep 21 16:44:33.803: INFO: openshift-service-catalog-controller-manager-operator-b9ccczlcs from openshift-service-catalog-controller-manager-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.803: INFO: 	Container operator ready: true, restart count 1
-Sep 21 16:44:33.803: INFO: cluster-samples-operator-5cfb985bf-n7nxd from openshift-cluster-samples-operator started at 2020-09-21 14:09:54 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.803: INFO: 	Container cluster-samples-operator ready: true, restart count 0
-Sep 21 16:44:33.803: INFO: 	Container cluster-samples-operator-watch ready: true, restart count 0
-Sep 21 16:44:33.803: INFO: ibm-file-plugin-66fdcfc4c9-tm6qk from kube-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.803: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
-Sep 21 16:44:33.803: INFO: openshift-state-metrics-678b6c786f-698gn from openshift-monitoring started at 2020-09-21 14:08:22 +0000 UTC (3 container statuses recorded)
-Sep 21 16:44:33.803: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
-Sep 21 16:44:33.803: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
-Sep 21 16:44:33.803: INFO: 	Container openshift-state-metrics ready: true, restart count 0
-Sep 21 16:44:33.803: INFO: redhat-marketplace-7b9bcdf96b-p4khw from openshift-marketplace started at 2020-09-21 14:09:34 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.803: INFO: 	Container redhat-marketplace ready: true, restart count 0
-Sep 21 16:44:33.803: INFO: openshift-service-catalog-apiserver-operator-6988647bc8-t9z7w from openshift-service-catalog-apiserver-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.804: INFO: 	Container operator ready: true, restart count 1
-Sep 21 16:44:33.804: INFO: ibmcloud-block-storage-plugin-68d5c65db9-7q8bv from kube-system started at 2020-09-21 14:07:36 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.804: INFO: 	Container ibmcloud-block-storage-plugin-container ready: true, restart count 0
-Sep 21 16:44:33.804: INFO: kube-state-metrics-7455b6c8-455w7 from openshift-monitoring started at 2020-09-21 14:08:22 +0000 UTC (3 container statuses recorded)
-Sep 21 16:44:33.804: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
-Sep 21 16:44:33.804: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
-Sep 21 16:44:33.804: INFO: 	Container kube-state-metrics ready: true, restart count 0
-Sep 21 16:44:33.804: INFO: prometheus-adapter-5f78ddc679-vb486 from openshift-monitoring started at 2020-09-21 14:08:28 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.804: INFO: 	Container prometheus-adapter ready: true, restart count 0
-Sep 21 16:44:33.804: INFO: multus-hprgb from openshift-multus started at 2020-09-21 14:06:32 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.804: INFO: 	Container kube-multus ready: true, restart count 0
-Sep 21 16:44:33.804: INFO: calico-typha-6fc9ff8b5f-gzhgw from calico-system started at 2020-09-21 14:07:18 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.804: INFO: 	Container calico-typha ready: true, restart count 0
-Sep 21 16:44:33.804: INFO: kube-storage-version-migrator-operator-86d8fc6986-vl4fl from openshift-kube-storage-version-migrator-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.804: INFO: 	Container kube-storage-version-migrator-operator ready: true, restart count 1
-Sep 21 16:44:33.804: INFO: network-operator-f456cfdbb-4m6pj from openshift-network-operator started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.804: INFO: 	Container network-operator ready: true, restart count 0
-Sep 21 16:44:33.804: INFO: cluster-monitoring-operator-75f4ddd85f-579q6 from openshift-monitoring started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.804: INFO: 	Container cluster-monitoring-operator ready: true, restart count 0
-Sep 21 16:44:33.804: INFO: prometheus-adapter-5f78ddc679-9hj7w from openshift-monitoring started at 2020-09-21 14:08:28 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.805: INFO: 	Container prometheus-adapter ready: true, restart count 0
-Sep 21 16:44:33.805: INFO: calico-kube-controllers-79d75767dd-xkvb6 from calico-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.805: INFO: 	Container calico-kube-controllers ready: true, restart count 0
-Sep 21 16:44:33.805: INFO: multus-admission-controller-gkchr from openshift-multus started at 2020-09-21 14:07:34 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.805: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.805: INFO: 	Container multus-admission-controller ready: true, restart count 0
-Sep 21 16:44:33.805: INFO: certified-operators-5cfd68758d-q6c6n from openshift-marketplace started at 2020-09-21 16:14:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.805: INFO: 	Container certified-operators ready: true, restart count 0
-Sep 21 16:44:33.805: INFO: community-operators-6566687cd-smnd2 from openshift-marketplace started at 2020-09-21 14:09:35 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.805: INFO: 	Container community-operators ready: true, restart count 0
-Sep 21 16:44:33.805: INFO: ibm-storage-watcher-7956f954f-26xx6 from kube-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.805: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
-Sep 21 16:44:33.805: INFO: service-ca-bd8866b9b-qzggv from openshift-service-ca started at 2020-09-21 14:07:50 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.805: INFO: 	Container service-ca-controller ready: true, restart count 0
-Sep 21 16:44:33.805: INFO: prometheus-k8s-1 from openshift-monitoring started at 2020-09-21 16:14:17 +0000 UTC (7 container statuses recorded)
-Sep 21 16:44:33.805: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.805: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:44:33.805: INFO: 	Container prometheus ready: true, restart count 1
-Sep 21 16:44:33.805: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
-Sep 21 16:44:33.805: INFO: 	Container prometheus-proxy ready: true, restart count 0
-Sep 21 16:44:33.806: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
-Sep 21 16:44:33.806: INFO: 	Container thanos-sidecar ready: true, restart count 0
-Sep 21 16:44:33.806: INFO: ibmcloud-block-storage-driver-pldbm from kube-system started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.806: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
-Sep 21 16:44:33.806: INFO: dns-operator-7886f4f4f4-rk7dv from openshift-dns-operator started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.806: INFO: 	Container dns-operator ready: true, restart count 0
-Sep 21 16:44:33.806: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.806: INFO: ingress-operator-57546dfcd9-2sg5v from openshift-ingress-operator started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.806: INFO: 	Container ingress-operator ready: true, restart count 0
-Sep 21 16:44:33.806: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.806: INFO: catalog-operator-69d4f5c59f-lqjgz from openshift-operator-lifecycle-manager started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.806: INFO: 	Container catalog-operator ready: true, restart count 0
-Sep 21 16:44:33.806: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-lgbbw from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.806: INFO: 	Container sonobuoy-worker ready: true, restart count 1
-Sep 21 16:44:33.806: INFO: 	Container systemd-logs ready: true, restart count 0
-Sep 21 16:44:33.806: INFO: openshift-kube-proxy-9svk2 from openshift-kube-proxy started at 2020-09-21 14:06:38 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.806: INFO: 	Container kube-proxy ready: true, restart count 0
-Sep 21 16:44:33.806: INFO: service-ca-operator-9f84dfdb5-wdhvg from openshift-service-ca-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.806: INFO: 	Container operator ready: true, restart count 1
-Sep 21 16:44:33.807: INFO: router-default-84dfcdc9b5-tk98j from openshift-ingress started at 2020-09-21 14:09:06 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.807: INFO: 	Container router ready: true, restart count 0
-Sep 21 16:44:33.807: INFO: node-exporter-jnrzm from openshift-monitoring started at 2020-09-21 14:08:24 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.807: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.807: INFO: 	Container node-exporter ready: true, restart count 0
-Sep 21 16:44:33.807: INFO: dns-default-958xc from openshift-dns started at 2020-09-21 14:09:05 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.807: INFO: 	Container dns ready: true, restart count 0
-Sep 21 16:44:33.807: INFO: 	Container dns-node-resolver ready: true, restart count 0
-Sep 21 16:44:33.807: INFO: alertmanager-main-0 from openshift-monitoring started at 2020-09-21 14:14:12 +0000 UTC (3 container statuses recorded)
-Sep 21 16:44:33.807: INFO: 	Container alertmanager ready: true, restart count 0
-Sep 21 16:44:33.807: INFO: 	Container alertmanager-proxy ready: true, restart count 0
-Sep 21 16:44:33.807: INFO: 	Container config-reloader ready: true, restart count 0
-Sep 21 16:44:33.807: INFO: downloads-7989988989-9rhh4 from openshift-console started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.807: INFO: 	Container download-server ready: true, restart count 0
-Sep 21 16:44:33.807: INFO: tuned-gpfhj from openshift-cluster-node-tuning-operator started at 2020-09-21 14:08:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.807: INFO: 	Container tuned ready: true, restart count 0
-Sep 21 16:44:33.807: INFO: downloads-7989988989-cjvjw from openshift-console started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.807: INFO: 	Container download-server ready: true, restart count 0
-Sep 21 16:44:33.807: INFO: ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-44pnn from ibm-system started at 2020-09-21 14:12:56 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.807: INFO: 	Container ibm-cloud-provider-ip-169-48-211-85 ready: true, restart count 0
-Sep 21 16:44:33.808: INFO: ibm-keepalived-watcher-r4tn7 from kube-system started at 2020-09-21 14:06:04 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.808: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Sep 21 16:44:33.808: INFO: marketplace-operator-56c6694c8f-x9rcq from openshift-marketplace started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.808: INFO: 	Container marketplace-operator ready: true, restart count 0
-Sep 21 16:44:33.808: INFO: console-operator-58fd84b95c-dc6vq from openshift-console-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.808: INFO: 	Container console-operator ready: true, restart count 1
-Sep 21 16:44:33.808: INFO: migrator-6c94c8c775-vxsnp from openshift-kube-storage-version-migrator started at 2020-09-21 14:07:55 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.808: INFO: 	Container migrator ready: true, restart count 0
-Sep 21 16:44:33.808: INFO: tigera-operator-679798d94d-hdlvj from tigera-operator started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.808: INFO: 	Container tigera-operator ready: true, restart count 2
-Sep 21 16:44:33.808: INFO: cluster-image-registry-operator-695bf78ffc-95l5m from openshift-image-registry started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.808: INFO: 	Container cluster-image-registry-operator ready: true, restart count 0
-Sep 21 16:44:33.808: INFO: 	Container cluster-image-registry-operator-watch ready: true, restart count 0
-Sep 21 16:44:33.808: INFO: cluster-storage-operator-5bcf8d9f8f-dls4n from openshift-cluster-storage-operator started at 2020-09-21 14:07:36 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.808: INFO: 	Container cluster-storage-operator ready: true, restart count 0
-Sep 21 16:44:33.808: INFO: image-registry-775dd777d7-5qtdd from openshift-image-registry started at 2020-09-21 14:11:01 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.808: INFO: 	Container registry ready: true, restart count 0
-Sep 21 16:44:33.808: INFO: ibm-master-proxy-static-10.241.51.149 from kube-system started at 2020-09-21 14:06:02 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.808: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
-Sep 21 16:44:33.808: INFO: 	Container pause ready: true, restart count 0
-Sep 21 16:44:33.809: INFO: registry-pvc-permissions-rlfxm from openshift-image-registry started at 2020-09-21 14:11:01 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.809: INFO: 	Container pvc-permissions ready: false, restart count 0
-Sep 21 16:44:33.809: INFO: cluster-node-tuning-operator-bdbf98554-22kv8 from openshift-cluster-node-tuning-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.809: INFO: 	Container cluster-node-tuning-operator ready: true, restart count 0
-Sep 21 16:44:33.809: INFO: redhat-operators-54dcdd9697-pcwgz from openshift-marketplace started at 2020-09-21 14:09:34 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.809: INFO: 	Container redhat-operators ready: true, restart count 0
-Sep 21 16:44:33.809: INFO: 
-Logging pods the kubelet thinks is on node 10.241.51.150 before test
-Sep 21 16:44:33.975: INFO: multus-admission-controller-snglx from openshift-multus started at 2020-09-21 14:13:31 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.975: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.975: INFO: 	Container multus-admission-controller ready: true, restart count 0
-Sep 21 16:44:33.975: INFO: alertmanager-main-1 from openshift-monitoring started at 2020-09-21 14:14:02 +0000 UTC (3 container statuses recorded)
-Sep 21 16:44:33.975: INFO: 	Container alertmanager ready: true, restart count 0
-Sep 21 16:44:33.975: INFO: 	Container alertmanager-proxy ready: true, restart count 0
-Sep 21 16:44:33.975: INFO: 	Container config-reloader ready: true, restart count 0
-Sep 21 16:44:33.975: INFO: router-default-84dfcdc9b5-mt4b7 from openshift-ingress started at 2020-09-21 16:07:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.975: INFO: 	Container router ready: true, restart count 0
-Sep 21 16:44:33.975: INFO: multus-n5qc8 from openshift-multus started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.976: INFO: 	Container kube-multus ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: openshift-kube-proxy-9bv2v from openshift-kube-proxy started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.976: INFO: 	Container kube-proxy ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: prometheus-k8s-0 from openshift-monitoring started at 2020-09-21 14:15:10 +0000 UTC (7 container statuses recorded)
-Sep 21 16:44:33.976: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: 	Container prometheus ready: true, restart count 1
-Sep 21 16:44:33.976: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: 	Container prometheus-proxy ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: 	Container thanos-sidecar ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: ibm-master-proxy-static-10.241.51.150 from kube-system started at 2020-09-21 14:12:19 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.976: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: 	Container pause ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: telemeter-client-68bdb7f795-nnhfl from openshift-monitoring started at 2020-09-21 14:13:34 +0000 UTC (3 container statuses recorded)
-Sep 21 16:44:33.976: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: 	Container reload ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: 	Container telemeter-client ready: true, restart count 0
-Sep 21 16:44:33.976: INFO: vpn-64d8d4987c-lxt95 from kube-system started at 2020-09-21 16:07:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.977: INFO: 	Container vpn ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: thanos-querier-8f4c5c746-xv2gj from openshift-monitoring started at 2020-09-21 16:14:05 +0000 UTC (4 container statuses recorded)
-Sep 21 16:44:33.977: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: 	Container oauth-proxy ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: 	Container thanos-querier ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: update-demo-kitten-m6wsh from kubectl-1643 started at 2020-09-21 16:44:16 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.977: INFO: 	Container update-demo ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: ibmcloud-block-storage-driver-jf9s4 from kube-system started at 2020-09-21 14:12:26 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.977: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: thanos-querier-8f4c5c746-v982g from openshift-monitoring started at 2020-09-21 14:14:42 +0000 UTC (4 container statuses recorded)
-Sep 21 16:44:33.977: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: 	Container oauth-proxy ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: 	Container thanos-querier ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: packageserver-75876789fb-h2vxx from openshift-operator-lifecycle-manager started at 2020-09-21 16:30:42 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.977: INFO: 	Container packageserver ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: node-ca-ql97q from openshift-image-registry started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.977: INFO: 	Container node-ca ready: true, restart count 0
-Sep 21 16:44:33.977: INFO: tuned-5n2xm from openshift-cluster-node-tuning-operator started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.978: INFO: 	Container tuned ready: true, restart count 0
-Sep 21 16:44:33.978: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-xzz27 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.978: INFO: 	Container sonobuoy-worker ready: true, restart count 1
-Sep 21 16:44:33.978: INFO: 	Container systemd-logs ready: true, restart count 0
-Sep 21 16:44:33.978: INFO: ibm-keepalived-watcher-bntgc from kube-system started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.978: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Sep 21 16:44:33.978: INFO: alertmanager-main-2 from openshift-monitoring started at 2020-09-21 16:14:17 +0000 UTC (3 container statuses recorded)
-Sep 21 16:44:33.978: INFO: 	Container alertmanager ready: true, restart count 0
-Sep 21 16:44:33.978: INFO: 	Container alertmanager-proxy ready: true, restart count 0
-Sep 21 16:44:33.978: INFO: 	Container config-reloader ready: true, restart count 0
-Sep 21 16:44:33.978: INFO: console-8549dd6c4f-b5jnn from openshift-console started at 2020-09-21 14:18:13 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.978: INFO: 	Container console ready: true, restart count 0
-Sep 21 16:44:33.978: INFO: node-exporter-flb5l from openshift-monitoring started at 2020-09-21 14:12:21 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.978: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:44:33.978: INFO: 	Container node-exporter ready: true, restart count 0
-Sep 21 16:44:33.978: INFO: calico-node-dn4jv from calico-system started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.978: INFO: 	Container calico-node ready: true, restart count 0
-Sep 21 16:44:33.978: INFO: grafana-c745ff78c-9pgfx from openshift-monitoring started at 2020-09-21 14:13:42 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.978: INFO: 	Container grafana ready: true, restart count 0
-Sep 21 16:44:33.978: INFO: 	Container grafana-proxy ready: true, restart count 0
-Sep 21 16:44:33.978: INFO: ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-ghw5q from ibm-system started at 2020-09-21 16:07:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.979: INFO: 	Container ibm-cloud-provider-ip-169-48-211-85 ready: true, restart count 0
-Sep 21 16:44:33.979: INFO: dns-default-z8v9j from openshift-dns started at 2020-09-21 14:12:21 +0000 UTC (2 container statuses recorded)
-Sep 21 16:44:33.979: INFO: 	Container dns ready: true, restart count 0
-Sep 21 16:44:33.979: INFO: 	Container dns-node-resolver ready: true, restart count 0
-Sep 21 16:44:33.979: INFO: calico-typha-6fc9ff8b5f-cbhcz from calico-system started at 2020-09-21 14:13:14 +0000 UTC (1 container statuses recorded)
-Sep 21 16:44:33.979: INFO: 	Container calico-typha ready: true, restart count 0
-[It] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]
+[It] should delete RS created by deployment when not orphaning [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Trying to launch a pod without a label to get a node which can launch it.
-STEP: Explicitly delete pod here to free the resource it takes.
-STEP: Trying to apply a random label on the found node.
-STEP: verifying the node has the label kubernetes.io/e2e-6419e453-84d6-4709-aebd-5b84931276f4 90
-STEP: Trying to create a pod(pod1) with hostport 54321 and hostIP 127.0.0.1 and expect scheduled
-STEP: Trying to create another pod(pod2) with hostport 54321 but hostIP 127.0.0.2 on the node which pod1 resides and expect scheduled
-STEP: Trying to create a third pod(pod3) with hostport 54321, hostIP 127.0.0.2 but use UDP protocol on the node which pod2 resides
-STEP: removing the label kubernetes.io/e2e-6419e453-84d6-4709-aebd-5b84931276f4 off the node 10.241.51.147
-STEP: verifying the node doesn't have the label kubernetes.io/e2e-6419e453-84d6-4709-aebd-5b84931276f4
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+STEP: create the deployment
+STEP: Wait for the Deployment to create new ReplicaSet
+STEP: delete the deployment
+STEP: wait for all rs to be garbage collected
+STEP: expected 0 pods, got 1 pods
+STEP: expected 0 rs, got 1 rs
+STEP: Gathering metrics
+Oct 26 16:48:16.355: INFO: For apiserver_request_total:
+For apiserver_request_latency_seconds:
+For apiserver_init_events_total:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
+
+[AfterEach] [sig-api-machinery] Garbage collector
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:44:49.067: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "sched-pred-1487" for this suite.
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
+W1026 16:48:16.355145      26 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Oct 26 16:48:16.355: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-4094" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance]","total":280,"completed":174,"skipped":2723,"failed":0}
 
-• [SLOW TEST:16.240 seconds]
-[sig-scheduling] SchedulerPredicates [Serial]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
-  validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]","total":280,"completed":170,"skipped":2850,"failed":0}
-SSSSSSSSS
 ------------------------------
-[sig-apps] Deployment 
-  deployment should support rollover [Conformance]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should be able to deny attaching pod [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Deployment
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:44:49.139: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename deployment
+Oct 26 16:48:16.394: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
-[It] deployment should support rollover [Conformance]
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Oct 26 16:48:17.539: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327697, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327697, loc:(*time.Location)(0x7db5bc0)}}, Reason:"NewReplicaSetCreated", Message:"Created new replica set \"sample-webhook-deployment-5f65f8c764\""}, v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327697, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327697, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}}, CollisionCount:(*int32)(nil)}
+Oct 26 16:48:19.558: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327697, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327697, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327697, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327697, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Oct 26 16:48:22.590: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should be able to deny attaching pod [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:44:49.602: INFO: Pod name rollover-pod: Found 0 pods out of 1
-Sep 21 16:44:54.629: INFO: Pod name rollover-pod: Found 1 pods out of 1
-STEP: ensuring each pod is running
-Sep 21 16:44:54.629: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready
-Sep 21 16:44:56.651: INFO: Creating deployment "test-rollover-deployment"
-Sep 21 16:44:56.704: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations
-Sep 21 16:44:58.753: INFO: Check revision of new replica set for deployment "test-rollover-deployment"
-Sep 21 16:44:58.811: INFO: Ensure that both replica sets have 1 created replica
-Sep 21 16:44:58.855: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update
-Sep 21 16:44:58.905: INFO: Updating deployment test-rollover-deployment
-Sep 21 16:44:58.905: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller
-Sep 21 16:45:01.051: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2
-Sep 21 16:45:01.116: INFO: Make sure deployment "test-rollover-deployment" is complete
-Sep 21 16:45:01.190: INFO: all replica sets need to contain the pod-template-hash label
-Sep 21 16:45:01.190: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303499, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:45:03.239: INFO: all replica sets need to contain the pod-template-hash label
-Sep 21 16:45:03.239: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303501, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:45:05.257: INFO: all replica sets need to contain the pod-template-hash label
-Sep 21 16:45:05.257: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303501, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:45:07.253: INFO: all replica sets need to contain the pod-template-hash label
-Sep 21 16:45:07.253: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303501, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:45:09.240: INFO: all replica sets need to contain the pod-template-hash label
-Sep 21 16:45:09.240: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303501, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:45:11.265: INFO: all replica sets need to contain the pod-template-hash label
-Sep 21 16:45:11.265: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303501, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303496, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:45:13.360: INFO: 
-Sep 21 16:45:13.361: INFO: Ensure that both old replica sets have no replicas
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
-Sep 21 16:45:13.435: INFO: Deployment "test-rollover-deployment":
-&Deployment{ObjectMeta:{test-rollover-deployment  deployment-1776 /apis/apps/v1/namespaces/deployment-1776/deployments/test-rollover-deployment ae7c7290-9375-4e8a-930e-7b31f11a6f09 78625 2 2020-09-21 16:44:56 +0000 UTC   map[name:rollover-pod] map[deployment.kubernetes.io/revision:2] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0091e4128  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-09-21 16:44:56 +0000 UTC,LastTransitionTime:2020-09-21 16:44:56 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rollover-deployment-574d6dfbff" has successfully progressed.,LastUpdateTime:2020-09-21 16:45:11 +0000 UTC,LastTransitionTime:2020-09-21 16:44:56 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},}
-
-Sep 21 16:45:13.456: INFO: New ReplicaSet "test-rollover-deployment-574d6dfbff" of Deployment "test-rollover-deployment":
-&ReplicaSet{ObjectMeta:{test-rollover-deployment-574d6dfbff  deployment-1776 /apis/apps/v1/namespaces/deployment-1776/replicasets/test-rollover-deployment-574d6dfbff c21b0c2a-8241-4c2f-a4af-b484f48aea86 78613 2 2020-09-21 16:44:58 +0000 UTC   map[name:rollover-pod pod-template-hash:574d6dfbff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-rollover-deployment ae7c7290-9375-4e8a-930e-7b31f11a6f09 0xc0091e45a7 0xc0091e45a8}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 574d6dfbff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod pod-template-hash:574d6dfbff] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0091e4618  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},}
-Sep 21 16:45:13.457: INFO: All old ReplicaSets of Deployment "test-rollover-deployment":
-Sep 21 16:45:13.457: INFO: &ReplicaSet{ObjectMeta:{test-rollover-controller  deployment-1776 /apis/apps/v1/namespaces/deployment-1776/replicasets/test-rollover-controller d306b092-97ec-46b3-a31c-9933dd87721f 78623 2 2020-09-21 16:44:49 +0000 UTC   map[name:rollover-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2] [{apps/v1 Deployment test-rollover-deployment ae7c7290-9375-4e8a-930e-7b31f11a6f09 0xc0091e44d7 0xc0091e44d8}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod pod:httpd] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc0091e4538  ClusterFirst map[]     false false false  PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
-Sep 21 16:45:13.457: INFO: &ReplicaSet{ObjectMeta:{test-rollover-deployment-f6c94f66c  deployment-1776 /apis/apps/v1/namespaces/deployment-1776/replicasets/test-rollover-deployment-f6c94f66c 3929684e-3de9-43e2-9a12-2df0fff51157 78541 2 2020-09-21 16:44:56 +0000 UTC   map[name:rollover-pod pod-template-hash:f6c94f66c] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-rollover-deployment ae7c7290-9375-4e8a-930e-7b31f11a6f09 0xc0091e4680 0xc0091e4681}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: f6c94f66c,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod pod-template-hash:f6c94f66c] map[] [] []  []} {[] [] [{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0091e46f8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
-Sep 21 16:45:13.483: INFO: Pod "test-rollover-deployment-574d6dfbff-gk2v5" is available:
-&Pod{ObjectMeta:{test-rollover-deployment-574d6dfbff-gk2v5 test-rollover-deployment-574d6dfbff- deployment-1776 /api/v1/namespaces/deployment-1776/pods/test-rollover-deployment-574d6dfbff-gk2v5 ea0a755c-8c48-487b-8a9d-cfffbf91865a 78563 0 2020-09-21 16:44:59 +0000 UTC   map[name:rollover-pod pod-template-hash:574d6dfbff] map[cni.projectcalico.org/podIP:172.30.237.89/32 cni.projectcalico.org/podIPs:172.30.237.89/32 k8s.v1.cni.cncf.io/networks-status:[{
-    "name": "k8s-pod-network",
-    "ips": [
-        "172.30.237.89"
-    ],
-    "dns": {}
-}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet test-rollover-deployment-574d6dfbff c21b0c2a-8241-4c2f-a4af-b484f48aea86 0xc00370c4c7 0xc00370c4c8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-tt9cc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-tt9cc,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-tt9cc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-6m9j9,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:44:59 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:45:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:45:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:44:59 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:172.30.237.89,StartTime:2020-09-21 16:44:59 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-09-21 16:45:00 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:gcr.io/kubernetes-e2e-test-images/agnhost@sha256:02b9e10ea1f7439a03d9bc3ee16d984172ca7eac1818792ad4d721c5c8f72ff4,ContainerID:cri-o://5603e7f289079ff49c75384f2777665449e4a02e5468f3e6eec22f22c2c10d12,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.237.89,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-[AfterEach] [sig-apps] Deployment
+STEP: Registering the webhook via the AdmissionRegistration API
+STEP: create a pod
+STEP: 'kubectl attach' the pod, should be denied by the webhook
+Oct 26 16:48:24.764: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 attach --namespace=webhook-3718 to-be-attached-pod -i -c=container1'
+Oct 26 16:48:24.997: INFO: rc: 1
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:45:13.483: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "deployment-1776" for this suite.
+Oct 26 16:48:25.039: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-3718" for this suite.
+STEP: Destroying namespace "webhook-3718-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:24.432 seconds]
-[sig-apps] Deployment
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  deployment should support rollover [Conformance]
+• [SLOW TEST:8.942 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should be able to deny attaching pod [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] Deployment deployment should support rollover [Conformance]","total":280,"completed":171,"skipped":2859,"failed":0}
-SSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance]","total":280,"completed":175,"skipped":2723,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-network] Services 
-  should serve a basic endpoint from pods  [Conformance]
+[k8s.io] Probing container 
+  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Services
+[BeforeEach] [k8s.io] Probing container
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:45:13.571: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename services
+Oct 26 16:48:25.336: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename container-probe
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should serve a basic endpoint from pods  [Conformance]
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating service endpoint-test2 in namespace services-5457
-STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-5457 to expose endpoints map[]
-Sep 21 16:45:14.304: INFO: successfully validated that service endpoint-test2 in namespace services-5457 exposes endpoints map[] (309.276386ms elapsed)
-STEP: Creating pod pod1 in namespace services-5457
-STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-5457 to expose endpoints map[pod1:[80]]
-Sep 21 16:45:16.976: INFO: successfully validated that service endpoint-test2 in namespace services-5457 exposes endpoints map[pod1:[80]] (2.444328925s elapsed)
-STEP: Creating pod pod2 in namespace services-5457
-STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-5457 to expose endpoints map[pod1:[80] pod2:[80]]
-Sep 21 16:45:20.631: INFO: successfully validated that service endpoint-test2 in namespace services-5457 exposes endpoints map[pod1:[80] pod2:[80]] (3.578741961s elapsed)
-STEP: Deleting pod pod1 in namespace services-5457
-STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-5457 to expose endpoints map[pod2:[80]]
-Sep 21 16:45:21.012: INFO: successfully validated that service endpoint-test2 in namespace services-5457 exposes endpoints map[pod2:[80]] (328.503106ms elapsed)
-STEP: Deleting pod pod2 in namespace services-5457
-STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-5457 to expose endpoints map[]
-Sep 21 16:45:21.112: INFO: successfully validated that service endpoint-test2 in namespace services-5457 exposes endpoints map[] (36.849636ms elapsed)
-[AfterEach] [sig-network] Services
+STEP: Creating pod liveness-466b0320-4d23-417d-b31a-923514241fd6 in namespace container-probe-3512
+Oct 26 16:48:29.606: INFO: Started pod liveness-466b0320-4d23-417d-b31a-923514241fd6 in namespace container-probe-3512
+STEP: checking the pod's current state and verifying that restartCount is present
+Oct 26 16:48:29.622: INFO: Initial restart count of pod liveness-466b0320-4d23-417d-b31a-923514241fd6 is 0
+Oct 26 16:48:49.807: INFO: Restart count of pod container-probe-3512/liveness-466b0320-4d23-417d-b31a-923514241fd6 is now 1 (20.184137639s elapsed)
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:45:21.307: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-5457" for this suite.
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
+Oct 26 16:48:49.861: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-3512" for this suite.
 
-• [SLOW TEST:7.829 seconds]
-[sig-network] Services
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should serve a basic endpoint from pods  [Conformance]
+• [SLOW TEST:24.567 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Services should serve a basic endpoint from pods  [Conformance]","total":280,"completed":172,"skipped":2871,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [k8s.io] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":280,"completed":176,"skipped":2762,"failed":0}
+SSSSSSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
+[BeforeEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:45:21.401: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
+Oct 26 16:48:49.904: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:45:21.868: INFO: Waiting up to 5m0s for pod "downwardapi-volume-542d5369-26ca-4064-8eb4-f55044b74035" in namespace "downward-api-9900" to be "success or failure"
-Sep 21 16:45:21.893: INFO: Pod "downwardapi-volume-542d5369-26ca-4064-8eb4-f55044b74035": Phase="Pending", Reason="", readiness=false. Elapsed: 25.349203ms
-Sep 21 16:45:23.932: INFO: Pod "downwardapi-volume-542d5369-26ca-4064-8eb4-f55044b74035": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064245119s
-Sep 21 16:45:25.960: INFO: Pod "downwardapi-volume-542d5369-26ca-4064-8eb4-f55044b74035": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.092175658s
+STEP: Creating configMap with name projected-configmap-test-volume-map-bf097506-9721-437d-8bbb-3c6815daeba3
+STEP: Creating a pod to test consume configMaps
+Oct 26 16:48:50.133: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-003d9709-0649-4867-8232-117c75ed1a54" in namespace "projected-3133" to be "success or failure"
+Oct 26 16:48:50.149: INFO: Pod "pod-projected-configmaps-003d9709-0649-4867-8232-117c75ed1a54": Phase="Pending", Reason="", readiness=false. Elapsed: 15.889294ms
+Oct 26 16:48:52.168: INFO: Pod "pod-projected-configmaps-003d9709-0649-4867-8232-117c75ed1a54": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.034956912s
 STEP: Saw pod success
-Sep 21 16:45:25.960: INFO: Pod "downwardapi-volume-542d5369-26ca-4064-8eb4-f55044b74035" satisfied condition "success or failure"
-Sep 21 16:45:25.988: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-542d5369-26ca-4064-8eb4-f55044b74035 container client-container: 
+Oct 26 16:48:52.168: INFO: Pod "pod-projected-configmaps-003d9709-0649-4867-8232-117c75ed1a54" satisfied condition "success or failure"
+Oct 26 16:48:52.185: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-configmaps-003d9709-0649-4867-8232-117c75ed1a54 container projected-configmap-volume-test: 
 STEP: delete the pod
-Sep 21 16:45:26.207: INFO: Waiting for pod downwardapi-volume-542d5369-26ca-4064-8eb4-f55044b74035 to disappear
-Sep 21 16:45:26.235: INFO: Pod downwardapi-volume-542d5369-26ca-4064-8eb4-f55044b74035 no longer exists
-[AfterEach] [sig-storage] Downward API volume
+Oct 26 16:48:52.289: INFO: Waiting for pod pod-projected-configmaps-003d9709-0649-4867-8232-117c75ed1a54 to disappear
+Oct 26 16:48:52.314: INFO: Pod pod-projected-configmaps-003d9709-0649-4867-8232-117c75ed1a54 no longer exists
+[AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:45:26.235: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-9900" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":173,"skipped":2909,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:48:52.314: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-3133" for this suite.
+•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","total":280,"completed":177,"skipped":2770,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] ReplicationController 
-  should release no longer matching pods [Conformance]
+[sig-storage] EmptyDir volumes 
+  volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] ReplicationController
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:45:26.310: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename replication-controller
+Oct 26 16:48:52.355: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should release no longer matching pods [Conformance]
+[It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Given a ReplicationController is created
-STEP: When the matched label of one of its pods change
-Sep 21 16:45:26.645: INFO: Pod name pod-release: Found 0 pods out of 1
-Sep 21 16:45:31.685: INFO: Pod name pod-release: Found 1 pods out of 1
-STEP: Then the pod is released
-[AfterEach] [sig-apps] ReplicationController
+STEP: Creating a pod to test emptydir volume type on tmpfs
+Oct 26 16:48:52.586: INFO: Waiting up to 5m0s for pod "pod-16f69d00-2dea-470f-b318-4db641197632" in namespace "emptydir-9029" to be "success or failure"
+Oct 26 16:48:52.603: INFO: Pod "pod-16f69d00-2dea-470f-b318-4db641197632": Phase="Pending", Reason="", readiness=false. Elapsed: 16.693899ms
+Oct 26 16:48:54.620: INFO: Pod "pod-16f69d00-2dea-470f-b318-4db641197632": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033754976s
+Oct 26 16:48:56.644: INFO: Pod "pod-16f69d00-2dea-470f-b318-4db641197632": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.058078159s
+STEP: Saw pod success
+Oct 26 16:48:56.645: INFO: Pod "pod-16f69d00-2dea-470f-b318-4db641197632" satisfied condition "success or failure"
+Oct 26 16:48:56.662: INFO: Trying to get logs from node 10.72.119.74 pod pod-16f69d00-2dea-470f-b318-4db641197632 container test-container: 
+STEP: delete the pod
+Oct 26 16:48:56.746: INFO: Waiting for pod pod-16f69d00-2dea-470f-b318-4db641197632 to disappear
+Oct 26 16:48:56.769: INFO: Pod pod-16f69d00-2dea-470f-b318-4db641197632 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:45:31.848: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replication-controller-9614" for this suite.
-
-• [SLOW TEST:5.625 seconds]
-[sig-apps] ReplicationController
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should release no longer matching pods [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] ReplicationController should release no longer matching pods [Conformance]","total":280,"completed":174,"skipped":2946,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:48:56.769: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-9029" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":178,"skipped":2798,"failed":0}
+SSSSSS
 ------------------------------
-[sig-api-machinery] Namespaces [Serial] 
-  should ensure that all pods are removed when a namespace is deleted [Conformance]
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with secret pod [LinuxOnly] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Namespaces [Serial]
+[BeforeEach] [sig-storage] Subpath
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:45:31.936: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename namespaces
+Oct 26 16:48:56.813: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename subpath
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should ensure that all pods are removed when a namespace is deleted [Conformance]
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
+STEP: Setting up data
+[It] should support subpaths with secret pod [LinuxOnly] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a test namespace
-STEP: Waiting for a default service account to be provisioned in namespace
-STEP: Creating a pod in the namespace
-STEP: Waiting for the pod to have running status
-STEP: Deleting the namespace
-STEP: Waiting for the namespace to be removed.
-STEP: Recreating the namespace
-STEP: Verifying there are no pods in the namespace
-[AfterEach] [sig-api-machinery] Namespaces [Serial]
+STEP: Creating pod pod-subpath-test-secret-7mdj
+STEP: Creating a pod to test atomic-volume-subpath
+Oct 26 16:48:57.088: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-7mdj" in namespace "subpath-6594" to be "success or failure"
+Oct 26 16:48:57.109: INFO: Pod "pod-subpath-test-secret-7mdj": Phase="Pending", Reason="", readiness=false. Elapsed: 20.685114ms
+Oct 26 16:48:59.126: INFO: Pod "pod-subpath-test-secret-7mdj": Phase="Running", Reason="", readiness=true. Elapsed: 2.037497085s
+Oct 26 16:49:01.142: INFO: Pod "pod-subpath-test-secret-7mdj": Phase="Running", Reason="", readiness=true. Elapsed: 4.054082167s
+Oct 26 16:49:03.160: INFO: Pod "pod-subpath-test-secret-7mdj": Phase="Running", Reason="", readiness=true. Elapsed: 6.07170517s
+Oct 26 16:49:05.176: INFO: Pod "pod-subpath-test-secret-7mdj": Phase="Running", Reason="", readiness=true. Elapsed: 8.088018813s
+Oct 26 16:49:07.194: INFO: Pod "pod-subpath-test-secret-7mdj": Phase="Running", Reason="", readiness=true. Elapsed: 10.105352316s
+Oct 26 16:49:09.210: INFO: Pod "pod-subpath-test-secret-7mdj": Phase="Running", Reason="", readiness=true. Elapsed: 12.121990906s
+Oct 26 16:49:11.228: INFO: Pod "pod-subpath-test-secret-7mdj": Phase="Running", Reason="", readiness=true. Elapsed: 14.139243581s
+Oct 26 16:49:13.249: INFO: Pod "pod-subpath-test-secret-7mdj": Phase="Running", Reason="", readiness=true. Elapsed: 16.160243329s
+Oct 26 16:49:15.266: INFO: Pod "pod-subpath-test-secret-7mdj": Phase="Running", Reason="", readiness=true. Elapsed: 18.177788389s
+Oct 26 16:49:17.284: INFO: Pod "pod-subpath-test-secret-7mdj": Phase="Running", Reason="", readiness=true. Elapsed: 20.19513369s
+Oct 26 16:49:19.299: INFO: Pod "pod-subpath-test-secret-7mdj": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.211085422s
+STEP: Saw pod success
+Oct 26 16:49:19.300: INFO: Pod "pod-subpath-test-secret-7mdj" satisfied condition "success or failure"
+Oct 26 16:49:19.317: INFO: Trying to get logs from node 10.72.119.74 pod pod-subpath-test-secret-7mdj container test-container-subpath-secret-7mdj: 
+STEP: delete the pod
+Oct 26 16:49:19.430: INFO: Waiting for pod pod-subpath-test-secret-7mdj to disappear
+Oct 26 16:49:19.447: INFO: Pod pod-subpath-test-secret-7mdj no longer exists
+STEP: Deleting pod pod-subpath-test-secret-7mdj
+Oct 26 16:49:19.447: INFO: Deleting pod "pod-subpath-test-secret-7mdj" in namespace "subpath-6594"
+[AfterEach] [sig-storage] Subpath
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:45:52.121: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "namespaces-5728" for this suite.
-STEP: Destroying namespace "nsdeletetest-5924" for this suite.
-Sep 21 16:45:52.284: INFO: Namespace nsdeletetest-5924 was already deleted
-STEP: Destroying namespace "nsdeletetest-6360" for this suite.
+Oct 26 16:49:19.464: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-6594" for this suite.
 
-• [SLOW TEST:20.394 seconds]
-[sig-api-machinery] Namespaces [Serial]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should ensure that all pods are removed when a namespace is deleted [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:22.686 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
+  Atomic writer volumes
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
+    should support subpaths with secret pod [LinuxOnly] [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance]","total":280,"completed":175,"skipped":2978,"failed":0}
-[k8s.io] InitContainer [NodeConformance] 
-  should invoke init containers on a RestartAlways pod [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:45:52.330: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename init-container
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
-[It] should invoke init containers on a RestartAlways pod [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-Sep 21 16:45:52.583: INFO: PodSpec: initContainers in spec.initContainers
-[AfterEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:45:56.471: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "init-container-4577" for this suite.
-•{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance]","total":280,"completed":176,"skipped":2978,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance]","total":280,"completed":179,"skipped":2804,"failed":0}
+SSSS
 ------------------------------
-[sig-node] Downward API 
-  should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
+[sig-cli] Kubectl client Kubectl describe 
+  should check if kubectl describe prints relevant information for rc and pods  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] Downward API
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:45:56.778: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
+Oct 26 16:49:19.500: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubectl
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
+[It] should check if kubectl describe prints relevant information for rc and pods  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward api env vars
-Sep 21 16:45:57.221: INFO: Waiting up to 5m0s for pod "downward-api-83280f8e-2b47-4b97-8375-03b815526641" in namespace "downward-api-5154" to be "success or failure"
-Sep 21 16:45:57.249: INFO: Pod "downward-api-83280f8e-2b47-4b97-8375-03b815526641": Phase="Pending", Reason="", readiness=false. Elapsed: 27.981991ms
-Sep 21 16:45:59.276: INFO: Pod "downward-api-83280f8e-2b47-4b97-8375-03b815526641": Phase="Pending", Reason="", readiness=false. Elapsed: 2.055305707s
-Sep 21 16:46:01.302: INFO: Pod "downward-api-83280f8e-2b47-4b97-8375-03b815526641": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.080788757s
-STEP: Saw pod success
-Sep 21 16:46:01.302: INFO: Pod "downward-api-83280f8e-2b47-4b97-8375-03b815526641" satisfied condition "success or failure"
-Sep 21 16:46:01.327: INFO: Trying to get logs from node 10.241.51.147 pod downward-api-83280f8e-2b47-4b97-8375-03b815526641 container dapi-container: 
-STEP: delete the pod
-Sep 21 16:46:01.477: INFO: Waiting for pod downward-api-83280f8e-2b47-4b97-8375-03b815526641 to disappear
-Sep 21 16:46:01.505: INFO: Pod downward-api-83280f8e-2b47-4b97-8375-03b815526641 no longer exists
-[AfterEach] [sig-node] Downward API
+Oct 26 16:49:19.653: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-1538'
+Oct 26 16:49:20.192: INFO: stderr: ""
+Oct 26 16:49:20.192: INFO: stdout: "replicationcontroller/agnhost-master created\n"
+Oct 26 16:49:20.192: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-1538'
+Oct 26 16:49:20.804: INFO: stderr: ""
+Oct 26 16:49:20.805: INFO: stdout: "service/agnhost-master created\n"
+STEP: Waiting for Agnhost master to start.
+Oct 26 16:49:21.826: INFO: Selector matched 1 pods for map[app:agnhost]
+Oct 26 16:49:21.826: INFO: Found 0 / 1
+Oct 26 16:49:22.828: INFO: Selector matched 1 pods for map[app:agnhost]
+Oct 26 16:49:22.828: INFO: Found 1 / 1
+Oct 26 16:49:22.828: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+Oct 26 16:49:22.848: INFO: Selector matched 1 pods for map[app:agnhost]
+Oct 26 16:49:22.848: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Oct 26 16:49:22.849: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 describe pod agnhost-master-f7msb --namespace=kubectl-1538'
+Oct 26 16:49:23.037: INFO: stderr: ""
+Oct 26 16:49:23.037: INFO: stdout: "Name:         agnhost-master-f7msb\nNamespace:    kubectl-1538\nPriority:     0\nNode:         10.72.119.74/10.72.119.74\nStart Time:   Mon, 26 Oct 2020 16:49:20 +0000\nLabels:       app=agnhost\n              role=master\nAnnotations:  cni.projectcalico.org/podIP: 172.30.194.126/32\n              cni.projectcalico.org/podIPs: 172.30.194.126/32\n              k8s.v1.cni.cncf.io/networks-status:\n                [{\n                    \"name\": \"k8s-pod-network\",\n                    \"ips\": [\n                        \"172.30.194.126\"\n                    ],\n                    \"dns\": {}\n                }]\n              openshift.io/scc: privileged\nStatus:       Running\nIP:           172.30.194.126\nIPs:\n  IP:           172.30.194.126\nControlled By:  ReplicationController/agnhost-master\nContainers:\n  agnhost-master:\n    Container ID:   cri-o://38d2f75c6b5b28844da28ac168b808e2240ed2b64566303adf025ab54bbcaf54\n    Image:          gcr.io/kubernetes-e2e-test-images/agnhost:2.8\n    Image ID:       gcr.io/kubernetes-e2e-test-images/agnhost@sha256:02b9e10ea1f7439a03d9bc3ee16d984172ca7eac1818792ad4d721c5c8f72ff4\n    Port:           6379/TCP\n    Host Port:      0/TCP\n    State:          Running\n      Started:      Mon, 26 Oct 2020 16:49:21 +0000\n    Ready:          True\n    Restart Count:  0\n    Environment:    \n    Mounts:\n      /var/run/secrets/kubernetes.io/serviceaccount from default-token-hfnz4 (ro)\nConditions:\n  Type              Status\n  Initialized       True \n  Ready             True \n  ContainersReady   True \n  PodScheduled      True \nVolumes:\n  default-token-hfnz4:\n    Type:        Secret (a volume populated by a Secret)\n    SecretName:  default-token-hfnz4\n    Optional:    false\nQoS Class:       BestEffort\nNode-Selectors:  \nTolerations:     node.kubernetes.io/not-ready:NoExecute for 300s\n                 node.kubernetes.io/unreachable:NoExecute for 300s\nEvents:\n  Type    Reason     Age   From                   Message\n  ----    ------     ----  ----                   -------\n  Normal  Scheduled  3s    default-scheduler      Successfully assigned kubectl-1538/agnhost-master-f7msb to 10.72.119.74\n  Normal  Pulled     2s    kubelet, 10.72.119.74  Container image \"gcr.io/kubernetes-e2e-test-images/agnhost:2.8\" already present on machine\n  Normal  Created    2s    kubelet, 10.72.119.74  Created container agnhost-master\n  Normal  Started    2s    kubelet, 10.72.119.74  Started container agnhost-master\n"
+Oct 26 16:49:23.037: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 describe rc agnhost-master --namespace=kubectl-1538'
+Oct 26 16:49:23.229: INFO: stderr: ""
+Oct 26 16:49:23.229: INFO: stdout: "Name:         agnhost-master\nNamespace:    kubectl-1538\nSelector:     app=agnhost,role=master\nLabels:       app=agnhost\n              role=master\nAnnotations:  \nReplicas:     1 current / 1 desired\nPods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n  Labels:  app=agnhost\n           role=master\n  Containers:\n   agnhost-master:\n    Image:        gcr.io/kubernetes-e2e-test-images/agnhost:2.8\n    Port:         6379/TCP\n    Host Port:    0/TCP\n    Environment:  \n    Mounts:       \n  Volumes:        \nEvents:\n  Type    Reason            Age   From                    Message\n  ----    ------            ----  ----                    -------\n  Normal  SuccessfulCreate  3s    replication-controller  Created pod: agnhost-master-f7msb\n"
+Oct 26 16:49:23.229: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 describe service agnhost-master --namespace=kubectl-1538'
+Oct 26 16:49:23.417: INFO: stderr: ""
+Oct 26 16:49:23.417: INFO: stdout: "Name:              agnhost-master\nNamespace:         kubectl-1538\nLabels:            app=agnhost\n                   role=master\nAnnotations:       \nSelector:          app=agnhost,role=master\nType:              ClusterIP\nIP:                172.21.235.29\nPort:                6379/TCP\nTargetPort:        agnhost-server/TCP\nEndpoints:         172.30.194.126:6379\nSession Affinity:  None\nEvents:            \n"
+Oct 26 16:49:23.436: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 describe node 10.72.119.74'
+Oct 26 16:49:23.792: INFO: stderr: ""
+Oct 26 16:49:23.792: INFO: stdout: "Name:               10.72.119.74\nRoles:              master,worker\nLabels:             arch=amd64\n                    beta.kubernetes.io/arch=amd64\n                    beta.kubernetes.io/instance-type=b3c.4x16.encrypted\n                    beta.kubernetes.io/os=linux\n                    failure-domain.beta.kubernetes.io/region=eu-gb\n                    failure-domain.beta.kubernetes.io/zone=lon06\n                    ibm-cloud.kubernetes.io/encrypted-docker-data=true\n                    ibm-cloud.kubernetes.io/external-ip=158.176.65.134\n                    ibm-cloud.kubernetes.io/iaas-provider=softlayer\n                    ibm-cloud.kubernetes.io/internal-ip=10.72.119.74\n                    ibm-cloud.kubernetes.io/machine-type=b3c.4x16.encrypted\n                    ibm-cloud.kubernetes.io/os=REDHAT_7_64\n                    ibm-cloud.kubernetes.io/region=eu-gb\n                    ibm-cloud.kubernetes.io/sgx-enabled=false\n                    ibm-cloud.kubernetes.io/worker-id=kube-bubdl5cl00f2uvkviqd0-kubee2epvgt-default-000003c8\n                    ibm-cloud.kubernetes.io/worker-pool-id=bubdl5cl00f2uvkviqd0-d359765\n                    ibm-cloud.kubernetes.io/worker-pool-name=default\n                    ibm-cloud.kubernetes.io/worker-version=4.4.26_1521_openshift\n                    ibm-cloud.kubernetes.io/zone=lon06\n                    kubernetes.io/arch=amd64\n                    kubernetes.io/hostname=10.72.119.74\n                    kubernetes.io/os=linux\n                    node-role.kubernetes.io/master=\n                    node-role.kubernetes.io/worker=\n                    node.kubernetes.io/instance-type=b3c.4x16.encrypted\n                    node.openshift.io/os_id=rhel\n                    privateVLAN=2722974\n                    publicVLAN=2722972\n                    topology.kubernetes.io/region=eu-gb\n                    topology.kubernetes.io/zone=lon06\nAnnotations:        projectcalico.org/IPv4Address: 10.72.119.74/26\n                    projectcalico.org/IPv4IPIPTunnelAddr: 172.30.194.64\nCreationTimestamp:  Mon, 26 Oct 2020 14:34:28 +0000\nTaints:             \nUnschedulable:      false\nLease:\n  HolderIdentity:  10.72.119.74\n  AcquireTime:     \n  RenewTime:       Mon, 26 Oct 2020 16:49:15 +0000\nConditions:\n  Type                 Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message\n  ----                 ------  -----------------                 ------------------                ------                       -------\n  NetworkUnavailable   False   Mon, 26 Oct 2020 14:35:53 +0000   Mon, 26 Oct 2020 14:35:53 +0000   CalicoIsUp                   Calico is running on this node\n  MemoryPressure       False   Mon, 26 Oct 2020 16:45:38 +0000   Mon, 26 Oct 2020 14:34:28 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available\n  DiskPressure         False   Mon, 26 Oct 2020 16:45:38 +0000   Mon, 26 Oct 2020 14:34:28 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure\n  PIDPressure          False   Mon, 26 Oct 2020 16:45:38 +0000   Mon, 26 Oct 2020 14:34:28 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available\n  Ready                True    Mon, 26 Oct 2020 16:45:38 +0000   Mon, 26 Oct 2020 14:35:58 +0000   KubeletReady                 kubelet is posting ready status\nAddresses:\n  InternalIP:  10.72.119.74\n  ExternalIP:  158.176.65.134\n  Hostname:    10.72.119.74\nCapacity:\n  cpu:                4\n  ephemeral-storage:  103078840Ki\n  hugepages-1Gi:      0\n  hugepages-2Mi:      0\n  memory:             16260872Ki\n  pods:               110\nAllocatable:\n  cpu:                3910m\n  ephemeral-storage:  100275095474\n  hugepages-1Gi:      0\n  hugepages-2Mi:      0\n  memory:             13484808Ki\n  pods:               110\nSystem Info:\n  Machine ID:                             5c8aa525c790457ea7eaaa541483dae0\n  System UUID:                            5801B4A4-F30B-FC2E-5BEF-331472FC905C\n  Boot ID:                                9898a7d1-4d38-4449-809d-b065010b6421\n  Kernel Version:                         3.10.0-1160.2.1.el7.x86_64\n  OS Image:                               Red Hat\n  Operating System:                       linux\n  Architecture:                           amd64\n  Container Runtime Version:              cri-o://1.17.5-8.rhaos4.4.git365b42a.el7\n  Kubelet Version:                        v1.17.1+c5fd4e8\n  Kube-Proxy Version:                     v1.17.1+c5fd4e8\nProviderID:                               ibm://fee034388aa6435883a1f720010ab3a2///bubdl5cl00f2uvkviqd0/kube-bubdl5cl00f2uvkviqd0-kubee2epvgt-default-000003c8\nNon-terminated Pods:                      (28 in total)\n  Namespace                               Name                                                       CPU Requests  CPU Limits  Memory Requests  Memory Limits  AGE\n  ---------                               ----                                                       ------------  ----------  ---------------  -------------  ---\n  calico-system                           calico-node-xlzf4                                          0 (0%)        0 (0%)      0 (0%)           0 (0%)         133m\n  calico-system                           calico-typha-7fd49c779d-46t4v                              0 (0%)        0 (0%)      0 (0%)           0 (0%)         131m\n  kube-system                             ibm-keepalived-watcher-49jbs                               5m (0%)       0 (0%)      10Mi (0%)        0 (0%)         134m\n  kube-system                             ibm-master-proxy-static-10.72.119.74                       25m (0%)      300m (7%)   32M (0%)         512M (3%)      134m\n  kube-system                             ibmcloud-block-storage-driver-ws85j                        50m (1%)      300m (7%)   100Mi (0%)       300Mi (2%)     134m\n  kube-system                             vpn-5b455f6f88-64q7j                                       5m (0%)       0 (0%)      5Mi (0%)         0 (0%)         39m\n  kubectl-1538                            agnhost-master-f7msb                                       0 (0%)        0 (0%)      0 (0%)           0 (0%)         3s\n  openshift-cluster-node-tuning-operator  tuned-p4cpc                                                10m (0%)      0 (0%)      50Mi (0%)        0 (0%)         132m\n  openshift-cluster-samples-operator      cluster-samples-operator-848994bd88-rw4rq                  20m (0%)      0 (0%)      0 (0%)           0 (0%)         39m\n  openshift-console                       console-5dd98d6f6f-zxw5x                                   10m (0%)      0 (0%)      100Mi (0%)       0 (0%)         39m\n  openshift-dns                           dns-default-s9kvh                                          110m (2%)     0 (0%)      70Mi (0%)        512Mi (3%)     131m\n  openshift-image-registry                node-ca-4ppdl                                              10m (0%)      0 (0%)      10Mi (0%)        0 (0%)         131m\n  openshift-kube-proxy                    openshift-kube-proxy-7d26g                                 100m (2%)     0 (0%)      200Mi (1%)       0 (0%)         134m\n  openshift-marketplace                   redhat-operators-5579777cd5-q5qm5                          10m (0%)      0 (0%)      100Mi (0%)       0 (0%)         9m55s\n  openshift-monitoring                    alertmanager-main-0                                        6m (0%)       0 (0%)      220Mi (1%)       0 (0%)         39m\n  openshift-monitoring                    grafana-f999495bf-t4tcv                                    5m (0%)       0 (0%)      120Mi (0%)       0 (0%)         39m\n  openshift-monitoring                    node-exporter-ls6pf                                        9m (0%)       0 (0%)      210Mi (1%)       0 (0%)         132m\n  openshift-monitoring                    prometheus-adapter-674456c5d6-p6bj5                        1m (0%)       0 (0%)      20Mi (0%)        0 (0%)         39m\n  openshift-monitoring                    telemeter-client-d5d9896c4-jpd8z                           3m (0%)       0 (0%)      20Mi (0%)        0 (0%)         39m\n  openshift-monitoring                    thanos-querier-84d7d58bc-c27ss                             8m (0%)       0 (0%)      72Mi (0%)        0 (0%)         39m\n  openshift-multus                        multus-admission-controller-dnt9r                          20m (0%)      0 (0%)      20Mi (0%)        0 (0%)         39m\n  openshift-multus                        multus-cc44t                                               10m (0%)      0 (0%)      150Mi (1%)       0 (0%)         134m\n  openshift-operator-lifecycle-manager    packageserver-5f76c64df7-jg7sj                             10m (0%)      0 (0%)      50Mi (0%)        0 (0%)         39m\n  openshift-service-ca                    service-ca-7d9db9c5b8-4w4cj                                10m (0%)      0 (0%)      120Mi (0%)       0 (0%)         39m\n  sonobuoy                                sonobuoy                                                   0 (0%)        0 (0%)      0 (0%)           0 (0%)         49m\n  sonobuoy                                sonobuoy-e2e-job-c6d919d638514ce4                          0 (0%)        0 (0%)      0 (0%)           0 (0%)         49m\n  sonobuoy                                sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-nbr6w    0 (0%)        0 (0%)      0 (0%)           0 (0%)         49m\n  tigera-operator                         tigera-operator-5c45c894b4-z9bnx                           100m (2%)     0 (0%)      40Mi (0%)        0 (0%)         39m\nAllocated resources:\n  (Total limits may be over 100 percent, i.e., overcommitted.)\n  Resource           Requests         Limits\n  --------           --------         ------\n  cpu                537m (13%)       600m (15%)\n  memory             1758738Ki (13%)  1363443712 (9%)\n  ephemeral-storage  0 (0%)           0 (0%)\nEvents:\n  Type    Reason                   Age                  From                      Message\n  ----    ------                   ----                 ----                      -------\n  Normal  Starting                 134m                 kubelet, 10.72.119.74     Starting kubelet.\n  Normal  NodeAllocatableEnforced  134m                 kubelet, 10.72.119.74     Updated Node Allocatable limit across pods\n  Normal  NodeHasSufficientMemory  134m (x8 over 134m)  kubelet, 10.72.119.74     Node 10.72.119.74 status is now: NodeHasSufficientMemory\n  Normal  NodeHasNoDiskPressure    134m (x8 over 134m)  kubelet, 10.72.119.74     Node 10.72.119.74 status is now: NodeHasNoDiskPressure\n  Normal  NodeHasSufficientPID     134m (x7 over 134m)  kubelet, 10.72.119.74     Node 10.72.119.74 status is now: NodeHasSufficientPID\n  Normal  Starting                 134m                 kube-proxy, 10.72.119.74  Starting kube-proxy.\n"
+Oct 26 16:49:23.793: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 describe namespace kubectl-1538'
+Oct 26 16:49:23.983: INFO: stderr: ""
+Oct 26 16:49:23.983: INFO: stdout: "Name:         kubectl-1538\nLabels:       e2e-framework=kubectl\n              e2e-run=2d1c55d1-9357-4e53-970b-33b80690768d\nAnnotations:  openshift.io/sa.scc.mcs: s0:c59,c44\n              openshift.io/sa.scc.supplemental-groups: 1003510000/10000\n              openshift.io/sa.scc.uid-range: 1003510000/10000\nStatus:       Active\n\nNo resource quota.\n\nNo LimitRange resource.\n"
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:46:01.505: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-5154" for this suite.
-•{"msg":"PASSED [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]","total":280,"completed":177,"skipped":3013,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:49:23.983: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-1538" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods  [Conformance]","total":280,"completed":180,"skipped":2808,"failed":0}
+SSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-network] Service endpoints latency 
-  should not be very high  [Conformance]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  listing mutating webhooks should work [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Service endpoints latency
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:46:01.589: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename svc-latency
+Oct 26 16:49:24.024: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not be very high  [Conformance]
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Oct 26 16:49:24.907: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 16:49:26.955: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327764, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327764, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327764, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739327764, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Oct 26 16:49:30.005: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] listing mutating webhooks should work [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:46:01.856: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: creating replication controller svc-latency-rc in namespace svc-latency-8508
-I0921 16:46:01.894954      24 runners.go:189] Created replication controller with name: svc-latency-rc, namespace: svc-latency-8508, replica count: 1
-I0921 16:46:02.945475      24 runners.go:189] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-I0921 16:46:03.945701      24 runners.go:189] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-Sep 21 16:46:04.128: INFO: Created: latency-svc-p4rf2
-Sep 21 16:46:04.157: INFO: Got endpoints: latency-svc-p4rf2 [111.966908ms]
-Sep 21 16:46:04.467: INFO: Created: latency-svc-9lq7w
-Sep 21 16:46:04.494: INFO: Got endpoints: latency-svc-9lq7w [335.061866ms]
-Sep 21 16:46:04.783: INFO: Created: latency-svc-ntn2h
-Sep 21 16:46:04.810: INFO: Got endpoints: latency-svc-ntn2h [651.613797ms]
-Sep 21 16:46:04.880: INFO: Created: latency-svc-86zl6
-Sep 21 16:46:04.906: INFO: Got endpoints: latency-svc-86zl6 [748.432657ms]
-Sep 21 16:46:04.930: INFO: Created: latency-svc-5x8md
-Sep 21 16:46:04.956: INFO: Got endpoints: latency-svc-5x8md [797.408335ms]
-Sep 21 16:46:04.981: INFO: Created: latency-svc-hmq2v
-Sep 21 16:46:05.015: INFO: Got endpoints: latency-svc-hmq2v [856.437571ms]
-Sep 21 16:46:05.034: INFO: Created: latency-svc-tpr52
-Sep 21 16:46:05.059: INFO: Got endpoints: latency-svc-tpr52 [901.02785ms]
-Sep 21 16:46:05.092: INFO: Created: latency-svc-j8jxd
-Sep 21 16:46:05.117: INFO: Got endpoints: latency-svc-j8jxd [958.623059ms]
-Sep 21 16:46:05.139: INFO: Created: latency-svc-b6rdv
-Sep 21 16:46:05.166: INFO: Got endpoints: latency-svc-b6rdv [1.007108781s]
-Sep 21 16:46:05.304: INFO: Created: latency-svc-qcvk4
-Sep 21 16:46:05.330: INFO: Got endpoints: latency-svc-qcvk4 [1.170451325s]
-Sep 21 16:46:05.364: INFO: Created: latency-svc-4tfqj
-Sep 21 16:46:05.387: INFO: Got endpoints: latency-svc-4tfqj [1.227219299s]
-Sep 21 16:46:05.414: INFO: Created: latency-svc-565w9
-Sep 21 16:46:05.444: INFO: Got endpoints: latency-svc-565w9 [1.283828248s]
-Sep 21 16:46:05.461: INFO: Created: latency-svc-p8zst
-Sep 21 16:46:05.493: INFO: Got endpoints: latency-svc-p8zst [1.333159032s]
-Sep 21 16:46:05.585: INFO: Created: latency-svc-bqxbq
-Sep 21 16:46:05.615: INFO: Got endpoints: latency-svc-bqxbq [1.456317027s]
-Sep 21 16:46:05.669: INFO: Created: latency-svc-7tb7s
-Sep 21 16:46:05.703: INFO: Got endpoints: latency-svc-7tb7s [1.543643784s]
-Sep 21 16:46:05.716: INFO: Created: latency-svc-f4mmd
-Sep 21 16:46:05.787: INFO: Got endpoints: latency-svc-f4mmd [1.627007301s]
-Sep 21 16:46:05.803: INFO: Created: latency-svc-4kzg5
-Sep 21 16:46:05.826: INFO: Got endpoints: latency-svc-4kzg5 [1.331866704s]
-Sep 21 16:46:05.859: INFO: Created: latency-svc-xzfbs
-Sep 21 16:46:05.893: INFO: Got endpoints: latency-svc-xzfbs [1.083048294s]
-Sep 21 16:46:06.384: INFO: Created: latency-svc-w4tg5
-Sep 21 16:46:06.411: INFO: Got endpoints: latency-svc-w4tg5 [1.505009972s]
-Sep 21 16:46:06.465: INFO: Created: latency-svc-wh2rx
-Sep 21 16:46:06.492: INFO: Got endpoints: latency-svc-wh2rx [1.536043268s]
-Sep 21 16:46:06.524: INFO: Created: latency-svc-mntt8
-Sep 21 16:46:06.549: INFO: Got endpoints: latency-svc-mntt8 [1.534025217s]
-Sep 21 16:46:06.573: INFO: Created: latency-svc-xl87k
-Sep 21 16:46:06.604: INFO: Got endpoints: latency-svc-xl87k [1.544320771s]
-Sep 21 16:46:06.612: INFO: Created: latency-svc-lm24t
-Sep 21 16:46:06.660: INFO: Created: latency-svc-l9zhq
-Sep 21 16:46:06.676: INFO: Got endpoints: latency-svc-lm24t [1.559473109s]
-Sep 21 16:46:06.699: INFO: Got endpoints: latency-svc-l9zhq [1.533061048s]
-Sep 21 16:46:06.705: INFO: Created: latency-svc-cwnxs
-Sep 21 16:46:06.737: INFO: Got endpoints: latency-svc-cwnxs [1.407711031s]
-Sep 21 16:46:06.751: INFO: Created: latency-svc-6z4r8
-Sep 21 16:46:06.786: INFO: Got endpoints: latency-svc-6z4r8 [1.399554223s]
-Sep 21 16:46:06.799: INFO: Created: latency-svc-gbbx9
-Sep 21 16:46:06.842: INFO: Got endpoints: latency-svc-gbbx9 [1.398289404s]
-Sep 21 16:46:06.856: INFO: Created: latency-svc-g6pnz
-Sep 21 16:46:06.892: INFO: Got endpoints: latency-svc-g6pnz [1.399155711s]
-Sep 21 16:46:06.911: INFO: Created: latency-svc-skw2g
-Sep 21 16:46:06.935: INFO: Got endpoints: latency-svc-skw2g [1.320315111s]
-Sep 21 16:46:06.946: INFO: Created: latency-svc-gv2pf
-Sep 21 16:46:06.979: INFO: Got endpoints: latency-svc-gv2pf [1.275687334s]
-Sep 21 16:46:07.016: INFO: Created: latency-svc-dcctx
-Sep 21 16:46:07.044: INFO: Got endpoints: latency-svc-dcctx [1.256655716s]
-Sep 21 16:46:07.058: INFO: Created: latency-svc-fm2wx
-Sep 21 16:46:07.081: INFO: Got endpoints: latency-svc-fm2wx [1.255554062s]
-Sep 21 16:46:07.102: INFO: Created: latency-svc-jbkbm
-Sep 21 16:46:07.133: INFO: Got endpoints: latency-svc-jbkbm [1.239779933s]
-Sep 21 16:46:07.152: INFO: Created: latency-svc-knbfw
-Sep 21 16:46:07.183: INFO: Got endpoints: latency-svc-knbfw [771.352797ms]
-Sep 21 16:46:07.212: INFO: Created: latency-svc-vs96w
-Sep 21 16:46:07.233: INFO: Got endpoints: latency-svc-vs96w [741.399492ms]
-Sep 21 16:46:07.255: INFO: Created: latency-svc-kr65p
-Sep 21 16:46:07.278: INFO: Got endpoints: latency-svc-kr65p [728.443958ms]
-Sep 21 16:46:07.307: INFO: Created: latency-svc-m6jrf
-Sep 21 16:46:07.334: INFO: Got endpoints: latency-svc-m6jrf [729.821211ms]
-Sep 21 16:46:07.345: INFO: Created: latency-svc-nncgn
-Sep 21 16:46:07.372: INFO: Got endpoints: latency-svc-nncgn [696.025286ms]
-Sep 21 16:46:07.396: INFO: Created: latency-svc-pgnb7
-Sep 21 16:46:07.423: INFO: Got endpoints: latency-svc-pgnb7 [723.700509ms]
-Sep 21 16:46:07.458: INFO: Created: latency-svc-k96fm
-Sep 21 16:46:07.485: INFO: Got endpoints: latency-svc-k96fm [747.355388ms]
-Sep 21 16:46:07.524: INFO: Created: latency-svc-6c4pt
-Sep 21 16:46:07.563: INFO: Got endpoints: latency-svc-6c4pt [776.192902ms]
-Sep 21 16:46:07.570: INFO: Created: latency-svc-p5m4h
-Sep 21 16:46:07.633: INFO: Created: latency-svc-r8722
-Sep 21 16:46:07.704: INFO: Created: latency-svc-pv75j
-Sep 21 16:46:07.746: INFO: Got endpoints: latency-svc-pv75j [810.67466ms]
-Sep 21 16:46:07.746: INFO: Got endpoints: latency-svc-p5m4h [904.190763ms]
-Sep 21 16:46:07.748: INFO: Got endpoints: latency-svc-r8722 [856.33859ms]
-Sep 21 16:46:07.814: INFO: Created: latency-svc-9xmxh
-Sep 21 16:46:07.838: INFO: Got endpoints: latency-svc-9xmxh [859.263705ms]
-Sep 21 16:46:07.946: INFO: Created: latency-svc-89bq6
-Sep 21 16:46:07.971: INFO: Got endpoints: latency-svc-89bq6 [927.089535ms]
-Sep 21 16:46:07.999: INFO: Created: latency-svc-msd8h
-Sep 21 16:46:08.029: INFO: Got endpoints: latency-svc-msd8h [948.048794ms]
-Sep 21 16:46:08.043: INFO: Created: latency-svc-xllgr
-Sep 21 16:46:08.069: INFO: Got endpoints: latency-svc-xllgr [935.759854ms]
-Sep 21 16:46:08.098: INFO: Created: latency-svc-h27ls
-Sep 21 16:46:08.120: INFO: Got endpoints: latency-svc-h27ls [937.830341ms]
-Sep 21 16:46:08.136: INFO: Created: latency-svc-qmzdk
-Sep 21 16:46:08.159: INFO: Got endpoints: latency-svc-qmzdk [925.687487ms]
-Sep 21 16:46:08.181: INFO: Created: latency-svc-n4qvd
-Sep 21 16:46:08.253: INFO: Got endpoints: latency-svc-n4qvd [975.337789ms]
-Sep 21 16:46:08.266: INFO: Created: latency-svc-69xf2
-Sep 21 16:46:08.291: INFO: Got endpoints: latency-svc-69xf2 [957.428033ms]
-Sep 21 16:46:08.309: INFO: Created: latency-svc-jg77j
-Sep 21 16:46:08.335: INFO: Got endpoints: latency-svc-jg77j [963.023841ms]
-Sep 21 16:46:08.357: INFO: Created: latency-svc-7wghr
-Sep 21 16:46:08.383: INFO: Got endpoints: latency-svc-7wghr [959.865227ms]
-Sep 21 16:46:08.397: INFO: Created: latency-svc-86skj
-Sep 21 16:46:08.418: INFO: Got endpoints: latency-svc-86skj [932.950638ms]
-Sep 21 16:46:08.435: INFO: Created: latency-svc-g4mx8
-Sep 21 16:46:08.471: INFO: Got endpoints: latency-svc-g4mx8 [908.576771ms]
-Sep 21 16:46:08.474: INFO: Created: latency-svc-djvjt
-Sep 21 16:46:08.505: INFO: Got endpoints: latency-svc-djvjt [759.085402ms]
-Sep 21 16:46:08.523: INFO: Created: latency-svc-jntr2
-Sep 21 16:46:08.561: INFO: Got endpoints: latency-svc-jntr2 [814.634775ms]
-Sep 21 16:46:08.589: INFO: Created: latency-svc-z59n5
-Sep 21 16:46:08.616: INFO: Got endpoints: latency-svc-z59n5 [867.583776ms]
-Sep 21 16:46:08.624: INFO: Created: latency-svc-9bl77
-Sep 21 16:46:08.653: INFO: Got endpoints: latency-svc-9bl77 [814.460199ms]
-Sep 21 16:46:08.673: INFO: Created: latency-svc-9kpm5
-Sep 21 16:46:08.699: INFO: Got endpoints: latency-svc-9kpm5 [728.021163ms]
-Sep 21 16:46:08.710: INFO: Created: latency-svc-xf2lt
-Sep 21 16:46:08.735: INFO: Got endpoints: latency-svc-xf2lt [705.097986ms]
-Sep 21 16:46:08.756: INFO: Created: latency-svc-6hbk7
-Sep 21 16:46:08.780: INFO: Got endpoints: latency-svc-6hbk7 [709.786946ms]
-Sep 21 16:46:08.802: INFO: Created: latency-svc-cq5kg
-Sep 21 16:46:08.829: INFO: Got endpoints: latency-svc-cq5kg [708.300438ms]
-Sep 21 16:46:08.839: INFO: Created: latency-svc-lxkrk
-Sep 21 16:46:08.869: INFO: Got endpoints: latency-svc-lxkrk [709.57045ms]
-Sep 21 16:46:08.875: INFO: Created: latency-svc-fshz8
-Sep 21 16:46:08.901: INFO: Got endpoints: latency-svc-fshz8 [647.651168ms]
-Sep 21 16:46:08.915: INFO: Created: latency-svc-c9gpl
-Sep 21 16:46:08.941: INFO: Got endpoints: latency-svc-c9gpl [649.281145ms]
-Sep 21 16:46:08.956: INFO: Created: latency-svc-czmw8
-Sep 21 16:46:08.990: INFO: Got endpoints: latency-svc-czmw8 [654.469404ms]
-Sep 21 16:46:08.995: INFO: Created: latency-svc-6c6cd
-Sep 21 16:46:09.022: INFO: Got endpoints: latency-svc-6c6cd [638.518164ms]
-Sep 21 16:46:09.049: INFO: Created: latency-svc-czk6m
-Sep 21 16:46:09.076: INFO: Created: latency-svc-zwpd4
-Sep 21 16:46:09.076: INFO: Got endpoints: latency-svc-czk6m [658.349738ms]
-Sep 21 16:46:09.103: INFO: Got endpoints: latency-svc-zwpd4 [631.686432ms]
-Sep 21 16:46:09.144: INFO: Created: latency-svc-5kspv
-Sep 21 16:46:09.169: INFO: Got endpoints: latency-svc-5kspv [664.095139ms]
-Sep 21 16:46:09.190: INFO: Created: latency-svc-2ztns
-Sep 21 16:46:09.218: INFO: Got endpoints: latency-svc-2ztns [656.968018ms]
-Sep 21 16:46:09.237: INFO: Created: latency-svc-5n49f
-Sep 21 16:46:09.259: INFO: Got endpoints: latency-svc-5n49f [642.778539ms]
-Sep 21 16:46:09.267: INFO: Created: latency-svc-dp64n
-Sep 21 16:46:09.294: INFO: Got endpoints: latency-svc-dp64n [640.803058ms]
-Sep 21 16:46:09.306: INFO: Created: latency-svc-2zxj9
-Sep 21 16:46:09.330: INFO: Got endpoints: latency-svc-2zxj9 [630.550559ms]
-Sep 21 16:46:09.352: INFO: Created: latency-svc-hnnfv
-Sep 21 16:46:09.382: INFO: Got endpoints: latency-svc-hnnfv [647.471775ms]
-Sep 21 16:46:09.384: INFO: Created: latency-svc-mw4ns
-Sep 21 16:46:09.411: INFO: Got endpoints: latency-svc-mw4ns [631.143589ms]
-Sep 21 16:46:09.415: INFO: Created: latency-svc-kzzcm
-Sep 21 16:46:09.443: INFO: Got endpoints: latency-svc-kzzcm [613.81285ms]
-Sep 21 16:46:09.456: INFO: Created: latency-svc-bmlfh
-Sep 21 16:46:09.478: INFO: Got endpoints: latency-svc-bmlfh [608.864788ms]
-Sep 21 16:46:09.488: INFO: Created: latency-svc-gm8zb
-Sep 21 16:46:09.510: INFO: Got endpoints: latency-svc-gm8zb [608.452129ms]
-Sep 21 16:46:09.541: INFO: Created: latency-svc-xtjsx
-Sep 21 16:46:09.569: INFO: Got endpoints: latency-svc-xtjsx [628.553889ms]
-Sep 21 16:46:09.575: INFO: Created: latency-svc-cqldf
-Sep 21 16:46:09.606: INFO: Got endpoints: latency-svc-cqldf [616.175022ms]
-Sep 21 16:46:09.628: INFO: Created: latency-svc-2n5nx
-Sep 21 16:46:09.667: INFO: Got endpoints: latency-svc-2n5nx [645.001204ms]
-Sep 21 16:46:09.679: INFO: Created: latency-svc-s42wq
-Sep 21 16:46:09.705: INFO: Got endpoints: latency-svc-s42wq [628.124507ms]
-Sep 21 16:46:09.719: INFO: Created: latency-svc-q96kg
-Sep 21 16:46:09.746: INFO: Got endpoints: latency-svc-q96kg [642.544379ms]
-Sep 21 16:46:09.756: INFO: Created: latency-svc-hvplf
-Sep 21 16:46:09.785: INFO: Got endpoints: latency-svc-hvplf [615.275575ms]
-Sep 21 16:46:09.800: INFO: Created: latency-svc-4xpg2
-Sep 21 16:46:09.830: INFO: Got endpoints: latency-svc-4xpg2 [611.33026ms]
-Sep 21 16:46:09.841: INFO: Created: latency-svc-j7gx5
-Sep 21 16:46:09.868: INFO: Got endpoints: latency-svc-j7gx5 [609.197425ms]
-Sep 21 16:46:09.873: INFO: Created: latency-svc-wpd5l
-Sep 21 16:46:09.902: INFO: Got endpoints: latency-svc-wpd5l [607.478765ms]
-Sep 21 16:46:09.907: INFO: Created: latency-svc-r7dj9
-Sep 21 16:46:09.930: INFO: Got endpoints: latency-svc-r7dj9 [600.145451ms]
-Sep 21 16:46:09.945: INFO: Created: latency-svc-r8lqk
-Sep 21 16:46:09.973: INFO: Got endpoints: latency-svc-r8lqk [590.778249ms]
-Sep 21 16:46:09.984: INFO: Created: latency-svc-qk2m8
-Sep 21 16:46:10.007: INFO: Got endpoints: latency-svc-qk2m8 [595.811735ms]
-Sep 21 16:46:10.013: INFO: Created: latency-svc-mvccq
-Sep 21 16:46:10.036: INFO: Got endpoints: latency-svc-mvccq [592.823159ms]
-Sep 21 16:46:10.056: INFO: Created: latency-svc-npzkz
-Sep 21 16:46:10.079: INFO: Got endpoints: latency-svc-npzkz [601.492459ms]
-Sep 21 16:46:10.086: INFO: Created: latency-svc-74tqp
-Sep 21 16:46:10.120: INFO: Created: latency-svc-cpfwl
-Sep 21 16:46:10.156: INFO: Created: latency-svc-bbx99
-Sep 21 16:46:10.272: INFO: Created: latency-svc-8vgpp
-Sep 21 16:46:10.305: INFO: Got endpoints: latency-svc-bbx99 [698.355437ms]
-Sep 21 16:46:10.305: INFO: Got endpoints: latency-svc-74tqp [795.34796ms]
-Sep 21 16:46:10.306: INFO: Got endpoints: latency-svc-cpfwl [736.074375ms]
-Sep 21 16:46:10.307: INFO: Got endpoints: latency-svc-8vgpp [640.203071ms]
-Sep 21 16:46:10.321: INFO: Created: latency-svc-n88jv
-Sep 21 16:46:10.342: INFO: Got endpoints: latency-svc-n88jv [637.433502ms]
-Sep 21 16:46:10.361: INFO: Created: latency-svc-f9xdj
-Sep 21 16:46:10.381: INFO: Got endpoints: latency-svc-f9xdj [635.228949ms]
-Sep 21 16:46:10.397: INFO: Created: latency-svc-vjrvn
-Sep 21 16:46:10.426: INFO: Got endpoints: latency-svc-vjrvn [641.525534ms]
-Sep 21 16:46:10.429: INFO: Created: latency-svc-tqgwv
-Sep 21 16:46:10.455: INFO: Got endpoints: latency-svc-tqgwv [625.112418ms]
-Sep 21 16:46:10.466: INFO: Created: latency-svc-r8pzl
-Sep 21 16:46:10.489: INFO: Got endpoints: latency-svc-r8pzl [620.919241ms]
-Sep 21 16:46:10.500: INFO: Created: latency-svc-wj46g
-Sep 21 16:46:10.524: INFO: Got endpoints: latency-svc-wj46g [622.556722ms]
-Sep 21 16:46:10.543: INFO: Created: latency-svc-c7m4m
-Sep 21 16:46:10.572: INFO: Got endpoints: latency-svc-c7m4m [640.97362ms]
-Sep 21 16:46:10.598: INFO: Created: latency-svc-7g4l4
-Sep 21 16:46:10.621: INFO: Got endpoints: latency-svc-7g4l4 [647.739995ms]
-Sep 21 16:46:10.634: INFO: Created: latency-svc-fr229
-Sep 21 16:46:10.656: INFO: Got endpoints: latency-svc-fr229 [649.442923ms]
-Sep 21 16:46:10.665: INFO: Created: latency-svc-zjmbk
-Sep 21 16:46:10.692: INFO: Got endpoints: latency-svc-zjmbk [654.972562ms]
-Sep 21 16:46:10.707: INFO: Created: latency-svc-f74l9
-Sep 21 16:46:10.729: INFO: Got endpoints: latency-svc-f74l9 [648.9956ms]
-Sep 21 16:46:10.787: INFO: Created: latency-svc-szbk6
-Sep 21 16:46:10.817: INFO: Created: latency-svc-grqrw
-Sep 21 16:46:10.860: INFO: Created: latency-svc-rnkjm
-Sep 21 16:46:10.895: INFO: Got endpoints: latency-svc-rnkjm [589.32093ms]
-Sep 21 16:46:10.895: INFO: Got endpoints: latency-svc-grqrw [589.736526ms]
-Sep 21 16:46:10.896: INFO: Got endpoints: latency-svc-szbk6 [590.774238ms]
-Sep 21 16:46:10.900: INFO: Created: latency-svc-p9szl
-Sep 21 16:46:10.943: INFO: Got endpoints: latency-svc-p9szl [635.680274ms]
-Sep 21 16:46:10.955: INFO: Created: latency-svc-sbm6k
-Sep 21 16:46:10.983: INFO: Got endpoints: latency-svc-sbm6k [640.095189ms]
-Sep 21 16:46:10.995: INFO: Created: latency-svc-mb2lt
-Sep 21 16:46:11.023: INFO: Got endpoints: latency-svc-mb2lt [641.452743ms]
-Sep 21 16:46:11.033: INFO: Created: latency-svc-zpdzj
-Sep 21 16:46:11.064: INFO: Created: latency-svc-fvl9c
-Sep 21 16:46:11.064: INFO: Got endpoints: latency-svc-zpdzj [637.546478ms]
-Sep 21 16:46:11.092: INFO: Got endpoints: latency-svc-fvl9c [637.024748ms]
-Sep 21 16:46:11.095: INFO: Created: latency-svc-whvd8
-Sep 21 16:46:11.124: INFO: Got endpoints: latency-svc-whvd8 [635.036258ms]
-Sep 21 16:46:11.133: INFO: Created: latency-svc-tm7v2
-Sep 21 16:46:11.170: INFO: Created: latency-svc-fkt6c
-Sep 21 16:46:11.170: INFO: Got endpoints: latency-svc-tm7v2 [645.997691ms]
-Sep 21 16:46:11.196: INFO: Got endpoints: latency-svc-fkt6c [624.083853ms]
-Sep 21 16:46:11.205: INFO: Created: latency-svc-p5rd9
-Sep 21 16:46:11.230: INFO: Got endpoints: latency-svc-p5rd9 [609.31077ms]
-Sep 21 16:46:11.239: INFO: Created: latency-svc-k5w5x
-Sep 21 16:46:11.270: INFO: Got endpoints: latency-svc-k5w5x [613.970587ms]
-Sep 21 16:46:11.278: INFO: Created: latency-svc-zr47p
-Sep 21 16:46:11.303: INFO: Got endpoints: latency-svc-zr47p [611.52718ms]
-Sep 21 16:46:11.317: INFO: Created: latency-svc-8zhvx
-Sep 21 16:46:11.339: INFO: Got endpoints: latency-svc-8zhvx [609.438531ms]
-Sep 21 16:46:11.346: INFO: Created: latency-svc-fp27x
-Sep 21 16:46:11.370: INFO: Got endpoints: latency-svc-fp27x [474.70541ms]
-Sep 21 16:46:11.390: INFO: Created: latency-svc-pgb5b
-Sep 21 16:46:11.416: INFO: Got endpoints: latency-svc-pgb5b [519.957214ms]
-Sep 21 16:46:11.427: INFO: Created: latency-svc-mmqs9
-Sep 21 16:46:11.465: INFO: Created: latency-svc-zc4zm
-Sep 21 16:46:11.466: INFO: Got endpoints: latency-svc-mmqs9 [570.137342ms]
-Sep 21 16:46:11.482: INFO: Got endpoints: latency-svc-zc4zm [538.803942ms]
-Sep 21 16:46:11.499: INFO: Created: latency-svc-dl755
-Sep 21 16:46:11.524: INFO: Got endpoints: latency-svc-dl755 [540.553083ms]
-Sep 21 16:46:11.534: INFO: Created: latency-svc-plljr
-Sep 21 16:46:11.560: INFO: Got endpoints: latency-svc-plljr [537.262913ms]
-Sep 21 16:46:11.574: INFO: Created: latency-svc-g2bwk
-Sep 21 16:46:11.600: INFO: Got endpoints: latency-svc-g2bwk [535.824264ms]
-Sep 21 16:46:11.609: INFO: Created: latency-svc-c2xfh
-Sep 21 16:46:11.633: INFO: Got endpoints: latency-svc-c2xfh [540.725126ms]
-Sep 21 16:46:11.644: INFO: Created: latency-svc-2gkxb
-Sep 21 16:46:11.669: INFO: Got endpoints: latency-svc-2gkxb [544.765116ms]
-Sep 21 16:46:11.688: INFO: Created: latency-svc-jvjn7
-Sep 21 16:46:11.715: INFO: Got endpoints: latency-svc-jvjn7 [544.174243ms]
-Sep 21 16:46:11.729: INFO: Created: latency-svc-qqwm8
-Sep 21 16:46:11.753: INFO: Got endpoints: latency-svc-qqwm8 [556.853518ms]
-Sep 21 16:46:11.772: INFO: Created: latency-svc-pr4fv
-Sep 21 16:46:11.806: INFO: Got endpoints: latency-svc-pr4fv [575.590454ms]
-Sep 21 16:46:11.815: INFO: Created: latency-svc-dv7j2
-Sep 21 16:46:11.848: INFO: Got endpoints: latency-svc-dv7j2 [577.290444ms]
-Sep 21 16:46:11.852: INFO: Created: latency-svc-sxmdz
-Sep 21 16:46:11.884: INFO: Got endpoints: latency-svc-sxmdz [581.068833ms]
-Sep 21 16:46:11.898: INFO: Created: latency-svc-94h2k
-Sep 21 16:46:11.932: INFO: Got endpoints: latency-svc-94h2k [593.005299ms]
-Sep 21 16:46:11.951: INFO: Created: latency-svc-wm87k
-Sep 21 16:46:11.978: INFO: Got endpoints: latency-svc-wm87k [608.815011ms]
-Sep 21 16:46:11.996: INFO: Created: latency-svc-hspqg
-Sep 21 16:46:12.022: INFO: Got endpoints: latency-svc-hspqg [606.401887ms]
-Sep 21 16:46:12.038: INFO: Created: latency-svc-r7bkj
-Sep 21 16:46:12.070: INFO: Got endpoints: latency-svc-r7bkj [604.56545ms]
-Sep 21 16:46:12.093: INFO: Created: latency-svc-8x4dp
-Sep 21 16:46:12.117: INFO: Created: latency-svc-zxqtg
-Sep 21 16:46:12.123: INFO: Got endpoints: latency-svc-8x4dp [640.87244ms]
-Sep 21 16:46:12.151: INFO: Got endpoints: latency-svc-zxqtg [627.023388ms]
-Sep 21 16:46:12.170: INFO: Created: latency-svc-ctc89
-Sep 21 16:46:12.206: INFO: Got endpoints: latency-svc-ctc89 [645.195031ms]
-Sep 21 16:46:12.229: INFO: Created: latency-svc-mfcfs
-Sep 21 16:46:12.260: INFO: Got endpoints: latency-svc-mfcfs [660.047957ms]
-Sep 21 16:46:12.276: INFO: Created: latency-svc-shvss
-Sep 21 16:46:12.304: INFO: Got endpoints: latency-svc-shvss [671.002092ms]
-Sep 21 16:46:12.318: INFO: Created: latency-svc-z9x9q
-Sep 21 16:46:12.350: INFO: Created: latency-svc-vc8lx
-Sep 21 16:46:12.352: INFO: Got endpoints: latency-svc-z9x9q [682.502202ms]
-Sep 21 16:46:12.379: INFO: Got endpoints: latency-svc-vc8lx [663.949011ms]
-Sep 21 16:46:12.394: INFO: Created: latency-svc-kcfbp
-Sep 21 16:46:12.434: INFO: Got endpoints: latency-svc-kcfbp [681.709946ms]
-Sep 21 16:46:12.435: INFO: Created: latency-svc-wsrl2
-Sep 21 16:46:12.462: INFO: Got endpoints: latency-svc-wsrl2 [656.164086ms]
-Sep 21 16:46:12.471: INFO: Created: latency-svc-kvw42
-Sep 21 16:46:12.496: INFO: Got endpoints: latency-svc-kvw42 [648.705622ms]
-Sep 21 16:46:12.575: INFO: Created: latency-svc-cd5vk
-Sep 21 16:46:12.604: INFO: Got endpoints: latency-svc-cd5vk [719.125783ms]
-Sep 21 16:46:12.628: INFO: Created: latency-svc-7l5dz
-Sep 21 16:46:12.673: INFO: Got endpoints: latency-svc-7l5dz [741.704595ms]
-Sep 21 16:46:12.694: INFO: Created: latency-svc-l2ddz
-Sep 21 16:46:12.719: INFO: Got endpoints: latency-svc-l2ddz [740.519424ms]
-Sep 21 16:46:12.772: INFO: Created: latency-svc-wcl5m
-Sep 21 16:46:12.804: INFO: Got endpoints: latency-svc-wcl5m [781.193707ms]
-Sep 21 16:46:12.834: INFO: Created: latency-svc-z2t7j
-Sep 21 16:46:12.864: INFO: Got endpoints: latency-svc-z2t7j [793.561337ms]
-Sep 21 16:46:12.927: INFO: Created: latency-svc-7h75m
-Sep 21 16:46:12.946: INFO: Created: latency-svc-8kbv7
-Sep 21 16:46:12.992: INFO: Got endpoints: latency-svc-8kbv7 [840.48192ms]
-Sep 21 16:46:12.992: INFO: Got endpoints: latency-svc-7h75m [868.890878ms]
-Sep 21 16:46:13.010: INFO: Created: latency-svc-64mjg
-Sep 21 16:46:13.042: INFO: Got endpoints: latency-svc-64mjg [836.425245ms]
-Sep 21 16:46:13.165: INFO: Created: latency-svc-sttsw
-Sep 21 16:46:13.189: INFO: Got endpoints: latency-svc-sttsw [928.955562ms]
-Sep 21 16:46:13.241: INFO: Created: latency-svc-6scjf
-Sep 21 16:46:13.269: INFO: Got endpoints: latency-svc-6scjf [964.960949ms]
-Sep 21 16:46:13.300: INFO: Created: latency-svc-56w9x
-Sep 21 16:46:13.316: INFO: Got endpoints: latency-svc-56w9x [964.001643ms]
-Sep 21 16:46:13.335: INFO: Created: latency-svc-9n78n
-Sep 21 16:46:13.365: INFO: Got endpoints: latency-svc-9n78n [985.913929ms]
-Sep 21 16:46:13.396: INFO: Created: latency-svc-j7x2n
-Sep 21 16:46:13.423: INFO: Got endpoints: latency-svc-j7x2n [988.65082ms]
-Sep 21 16:46:13.428: INFO: Created: latency-svc-dg7b8
-Sep 21 16:46:13.456: INFO: Got endpoints: latency-svc-dg7b8 [993.993645ms]
-Sep 21 16:46:13.481: INFO: Created: latency-svc-vjcd9
-Sep 21 16:46:13.503: INFO: Got endpoints: latency-svc-vjcd9 [1.00673235s]
-Sep 21 16:46:13.522: INFO: Created: latency-svc-q5d49
-Sep 21 16:46:13.552: INFO: Got endpoints: latency-svc-q5d49 [947.967243ms]
-Sep 21 16:46:13.573: INFO: Created: latency-svc-mxpkm
-Sep 21 16:46:13.596: INFO: Got endpoints: latency-svc-mxpkm [922.051178ms]
-Sep 21 16:46:13.617: INFO: Created: latency-svc-trzrq
-Sep 21 16:46:13.642: INFO: Got endpoints: latency-svc-trzrq [922.6517ms]
-Sep 21 16:46:13.657: INFO: Created: latency-svc-q9rdc
-Sep 21 16:46:13.681: INFO: Got endpoints: latency-svc-q9rdc [877.250183ms]
-Sep 21 16:46:13.703: INFO: Created: latency-svc-pwd7s
-Sep 21 16:46:13.725: INFO: Got endpoints: latency-svc-pwd7s [860.880303ms]
-Sep 21 16:46:13.742: INFO: Created: latency-svc-zvv27
-Sep 21 16:46:13.773: INFO: Got endpoints: latency-svc-zvv27 [781.074292ms]
-Sep 21 16:46:13.903: INFO: Created: latency-svc-h8jkk
-Sep 21 16:46:13.931: INFO: Got endpoints: latency-svc-h8jkk [939.747057ms]
-Sep 21 16:46:13.942: INFO: Created: latency-svc-987rw
-Sep 21 16:46:13.968: INFO: Got endpoints: latency-svc-987rw [925.527658ms]
-Sep 21 16:46:13.988: INFO: Created: latency-svc-qt84k
-Sep 21 16:46:14.015: INFO: Got endpoints: latency-svc-qt84k [826.331554ms]
-Sep 21 16:46:14.042: INFO: Created: latency-svc-5lclq
-Sep 21 16:46:14.068: INFO: Got endpoints: latency-svc-5lclq [798.608015ms]
-Sep 21 16:46:14.079: INFO: Created: latency-svc-f7wjm
-Sep 21 16:46:14.104: INFO: Got endpoints: latency-svc-f7wjm [787.980492ms]
-Sep 21 16:46:14.195: INFO: Created: latency-svc-sbpl2
-Sep 21 16:46:14.226: INFO: Got endpoints: latency-svc-sbpl2 [860.968155ms]
-Sep 21 16:46:14.240: INFO: Created: latency-svc-59ntp
-Sep 21 16:46:14.269: INFO: Got endpoints: latency-svc-59ntp [845.090214ms]
-Sep 21 16:46:14.282: INFO: Created: latency-svc-tm2wp
-Sep 21 16:46:14.309: INFO: Got endpoints: latency-svc-tm2wp [852.849724ms]
-Sep 21 16:46:14.317: INFO: Created: latency-svc-xc7zj
-Sep 21 16:46:14.350: INFO: Created: latency-svc-mz5bb
-Sep 21 16:46:14.350: INFO: Got endpoints: latency-svc-xc7zj [847.196929ms]
-Sep 21 16:46:14.372: INFO: Got endpoints: latency-svc-mz5bb [819.912081ms]
-Sep 21 16:46:14.472: INFO: Created: latency-svc-7rfl4
-Sep 21 16:46:14.496: INFO: Got endpoints: latency-svc-7rfl4 [900.449675ms]
-Sep 21 16:46:14.505: INFO: Created: latency-svc-rnv65
-Sep 21 16:46:14.532: INFO: Got endpoints: latency-svc-rnv65 [889.725364ms]
-Sep 21 16:46:14.543: INFO: Created: latency-svc-lwlzn
-Sep 21 16:46:14.567: INFO: Got endpoints: latency-svc-lwlzn [885.314385ms]
-Sep 21 16:46:14.581: INFO: Created: latency-svc-j5nl2
-Sep 21 16:46:14.607: INFO: Got endpoints: latency-svc-j5nl2 [881.825537ms]
-Sep 21 16:46:14.615: INFO: Created: latency-svc-74djh
-Sep 21 16:46:14.646: INFO: Got endpoints: latency-svc-74djh [872.831292ms]
-Sep 21 16:46:14.652: INFO: Created: latency-svc-gsrht
-Sep 21 16:46:14.681: INFO: Got endpoints: latency-svc-gsrht [749.141206ms]
-Sep 21 16:46:14.688: INFO: Created: latency-svc-t8d7f
-Sep 21 16:46:14.712: INFO: Got endpoints: latency-svc-t8d7f [744.294698ms]
-Sep 21 16:46:14.722: INFO: Created: latency-svc-vj5qx
-Sep 21 16:46:14.746: INFO: Got endpoints: latency-svc-vj5qx [730.80791ms]
-Sep 21 16:46:14.759: INFO: Created: latency-svc-8xp8p
-Sep 21 16:46:14.793: INFO: Got endpoints: latency-svc-8xp8p [724.473306ms]
-Sep 21 16:46:14.795: INFO: Created: latency-svc-zl9qt
-Sep 21 16:46:14.820: INFO: Got endpoints: latency-svc-zl9qt [715.651778ms]
-Sep 21 16:46:14.833: INFO: Created: latency-svc-mr54j
-Sep 21 16:46:14.861: INFO: Got endpoints: latency-svc-mr54j [634.844161ms]
-Sep 21 16:46:15.078: INFO: Created: latency-svc-hqgm4
-Sep 21 16:46:15.102: INFO: Got endpoints: latency-svc-hqgm4 [833.617144ms]
-Sep 21 16:46:15.115: INFO: Created: latency-svc-v2gbm
-Sep 21 16:46:15.134: INFO: Got endpoints: latency-svc-v2gbm [824.759695ms]
-Sep 21 16:46:15.211: INFO: Created: latency-svc-m4fmt
-Sep 21 16:46:15.236: INFO: Got endpoints: latency-svc-m4fmt [885.764622ms]
-Sep 21 16:46:15.259: INFO: Created: latency-svc-f726j
-Sep 21 16:46:15.289: INFO: Got endpoints: latency-svc-f726j [917.26902ms]
-Sep 21 16:46:15.301: INFO: Created: latency-svc-vp5xn
-Sep 21 16:46:15.326: INFO: Got endpoints: latency-svc-vp5xn [829.635404ms]
-Sep 21 16:46:15.370: INFO: Created: latency-svc-qrcmk
-Sep 21 16:46:15.393: INFO: Got endpoints: latency-svc-qrcmk [860.942458ms]
-Sep 21 16:46:15.393: INFO: Latencies: [335.061866ms 474.70541ms 519.957214ms 535.824264ms 537.262913ms 538.803942ms 540.553083ms 540.725126ms 544.174243ms 544.765116ms 556.853518ms 570.137342ms 575.590454ms 577.290444ms 581.068833ms 589.32093ms 589.736526ms 590.774238ms 590.778249ms 592.823159ms 593.005299ms 595.811735ms 600.145451ms 601.492459ms 604.56545ms 606.401887ms 607.478765ms 608.452129ms 608.815011ms 608.864788ms 609.197425ms 609.31077ms 609.438531ms 611.33026ms 611.52718ms 613.81285ms 613.970587ms 615.275575ms 616.175022ms 620.919241ms 622.556722ms 624.083853ms 625.112418ms 627.023388ms 628.124507ms 628.553889ms 630.550559ms 631.143589ms 631.686432ms 634.844161ms 635.036258ms 635.228949ms 635.680274ms 637.024748ms 637.433502ms 637.546478ms 638.518164ms 640.095189ms 640.203071ms 640.803058ms 640.87244ms 640.97362ms 641.452743ms 641.525534ms 642.544379ms 642.778539ms 645.001204ms 645.195031ms 645.997691ms 647.471775ms 647.651168ms 647.739995ms 648.705622ms 648.9956ms 649.281145ms 649.442923ms 651.613797ms 654.469404ms 654.972562ms 656.164086ms 656.968018ms 658.349738ms 660.047957ms 663.949011ms 664.095139ms 671.002092ms 681.709946ms 682.502202ms 696.025286ms 698.355437ms 705.097986ms 708.300438ms 709.57045ms 709.786946ms 715.651778ms 719.125783ms 723.700509ms 724.473306ms 728.021163ms 728.443958ms 729.821211ms 730.80791ms 736.074375ms 740.519424ms 741.399492ms 741.704595ms 744.294698ms 747.355388ms 748.432657ms 749.141206ms 759.085402ms 771.352797ms 776.192902ms 781.074292ms 781.193707ms 787.980492ms 793.561337ms 795.34796ms 797.408335ms 798.608015ms 810.67466ms 814.460199ms 814.634775ms 819.912081ms 824.759695ms 826.331554ms 829.635404ms 833.617144ms 836.425245ms 840.48192ms 845.090214ms 847.196929ms 852.849724ms 856.33859ms 856.437571ms 859.263705ms 860.880303ms 860.942458ms 860.968155ms 867.583776ms 868.890878ms 872.831292ms 877.250183ms 881.825537ms 885.314385ms 885.764622ms 889.725364ms 900.449675ms 901.02785ms 904.190763ms 908.576771ms 917.26902ms 922.051178ms 922.6517ms 925.527658ms 925.687487ms 927.089535ms 928.955562ms 932.950638ms 935.759854ms 937.830341ms 939.747057ms 947.967243ms 948.048794ms 957.428033ms 958.623059ms 959.865227ms 963.023841ms 964.001643ms 964.960949ms 975.337789ms 985.913929ms 988.65082ms 993.993645ms 1.00673235s 1.007108781s 1.083048294s 1.170451325s 1.227219299s 1.239779933s 1.255554062s 1.256655716s 1.275687334s 1.283828248s 1.320315111s 1.331866704s 1.333159032s 1.398289404s 1.399155711s 1.399554223s 1.407711031s 1.456317027s 1.505009972s 1.533061048s 1.534025217s 1.536043268s 1.543643784s 1.544320771s 1.559473109s 1.627007301s]
-Sep 21 16:46:15.393: INFO: 50 %ile: 729.821211ms
-Sep 21 16:46:15.393: INFO: 90 %ile: 1.255554062s
-Sep 21 16:46:15.393: INFO: 99 %ile: 1.559473109s
-Sep 21 16:46:15.393: INFO: Total sample count: 200
-[AfterEach] [sig-network] Service endpoints latency
+STEP: Listing all of the created validation webhooks
+STEP: Creating a configMap that should be mutated
+STEP: Deleting the collection of validation webhooks
+STEP: Creating a configMap that should not be mutated
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:46:15.393: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "svc-latency-8508" for this suite.
+Oct 26 16:49:30.690: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-1444" for this suite.
+STEP: Destroying namespace "webhook-1444-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:13.877 seconds]
-[sig-network] Service endpoints latency
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should not be very high  [Conformance]
+• [SLOW TEST:6.932 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  listing mutating webhooks should work [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Service endpoints latency should not be very high  [Conformance]","total":280,"completed":178,"skipped":3070,"failed":0}
-SSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance]","total":280,"completed":181,"skipped":2825,"failed":0}
+SSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+[k8s.io] Variable Expansion 
+  should allow substituting values in a container's args [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [k8s.io] Variable Expansion
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:46:15.468: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename emptydir
+Oct 26 16:49:30.956: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename var-expansion
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+[It] should allow substituting values in a container's args [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0777 on tmpfs
-Sep 21 16:46:15.863: INFO: Waiting up to 5m0s for pod "pod-12f799a3-7ae5-4887-9a0a-3bda07a8a568" in namespace "emptydir-3660" to be "success or failure"
-Sep 21 16:46:15.902: INFO: Pod "pod-12f799a3-7ae5-4887-9a0a-3bda07a8a568": Phase="Pending", Reason="", readiness=false. Elapsed: 39.707272ms
-Sep 21 16:46:17.927: INFO: Pod "pod-12f799a3-7ae5-4887-9a0a-3bda07a8a568": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.064322196s
+STEP: Creating a pod to test substitution in container's args
+Oct 26 16:49:31.184: INFO: Waiting up to 5m0s for pod "var-expansion-b372f555-0f44-4c99-9310-d05cfb2fca96" in namespace "var-expansion-2965" to be "success or failure"
+Oct 26 16:49:31.202: INFO: Pod "var-expansion-b372f555-0f44-4c99-9310-d05cfb2fca96": Phase="Pending", Reason="", readiness=false. Elapsed: 18.81344ms
+Oct 26 16:49:33.218: INFO: Pod "var-expansion-b372f555-0f44-4c99-9310-d05cfb2fca96": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.034784754s
 STEP: Saw pod success
-Sep 21 16:46:17.927: INFO: Pod "pod-12f799a3-7ae5-4887-9a0a-3bda07a8a568" satisfied condition "success or failure"
-Sep 21 16:46:17.965: INFO: Trying to get logs from node 10.241.51.147 pod pod-12f799a3-7ae5-4887-9a0a-3bda07a8a568 container test-container: 
+Oct 26 16:49:33.219: INFO: Pod "var-expansion-b372f555-0f44-4c99-9310-d05cfb2fca96" satisfied condition "success or failure"
+Oct 26 16:49:33.235: INFO: Trying to get logs from node 10.72.119.74 pod var-expansion-b372f555-0f44-4c99-9310-d05cfb2fca96 container dapi-container: 
 STEP: delete the pod
-Sep 21 16:46:18.108: INFO: Waiting for pod pod-12f799a3-7ae5-4887-9a0a-3bda07a8a568 to disappear
-Sep 21 16:46:18.134: INFO: Pod pod-12f799a3-7ae5-4887-9a0a-3bda07a8a568 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+Oct 26 16:49:33.320: INFO: Waiting for pod var-expansion-b372f555-0f44-4c99-9310-d05cfb2fca96 to disappear
+Oct 26 16:49:33.335: INFO: Pod var-expansion-b372f555-0f44-4c99-9310-d05cfb2fca96 no longer exists
+[AfterEach] [k8s.io] Variable Expansion
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:46:18.134: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-3660" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":179,"skipped":3085,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:49:33.335: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "var-expansion-2965" for this suite.
+•{"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance]","total":280,"completed":182,"skipped":2829,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide container's cpu limit [NodeConformance] [Conformance]
+[sig-storage] EmptyDir volumes 
+  should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:46:18.199: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
+Oct 26 16:49:33.376: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should provide container's cpu limit [NodeConformance] [Conformance]
+[It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:46:18.779: INFO: Waiting up to 5m0s for pod "downwardapi-volume-328bfe61-13fd-4cfb-8156-0ca2b7b9b398" in namespace "downward-api-4799" to be "success or failure"
-Sep 21 16:46:18.809: INFO: Pod "downwardapi-volume-328bfe61-13fd-4cfb-8156-0ca2b7b9b398": Phase="Pending", Reason="", readiness=false. Elapsed: 29.716855ms
-Sep 21 16:46:20.830: INFO: Pod "downwardapi-volume-328bfe61-13fd-4cfb-8156-0ca2b7b9b398": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.050549498s
+STEP: Creating a pod to test emptydir 0644 on node default medium
+Oct 26 16:49:33.603: INFO: Waiting up to 5m0s for pod "pod-9b5da8c0-bc09-4959-b70f-9643ab0b23c8" in namespace "emptydir-8213" to be "success or failure"
+Oct 26 16:49:33.619: INFO: Pod "pod-9b5da8c0-bc09-4959-b70f-9643ab0b23c8": Phase="Pending", Reason="", readiness=false. Elapsed: 16.045906ms
+Oct 26 16:49:35.637: INFO: Pod "pod-9b5da8c0-bc09-4959-b70f-9643ab0b23c8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034488642s
+Oct 26 16:49:37.655: INFO: Pod "pod-9b5da8c0-bc09-4959-b70f-9643ab0b23c8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.051663089s
 STEP: Saw pod success
-Sep 21 16:46:20.830: INFO: Pod "downwardapi-volume-328bfe61-13fd-4cfb-8156-0ca2b7b9b398" satisfied condition "success or failure"
-Sep 21 16:46:20.863: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-328bfe61-13fd-4cfb-8156-0ca2b7b9b398 container client-container: 
+Oct 26 16:49:37.655: INFO: Pod "pod-9b5da8c0-bc09-4959-b70f-9643ab0b23c8" satisfied condition "success or failure"
+Oct 26 16:49:37.671: INFO: Trying to get logs from node 10.72.119.74 pod pod-9b5da8c0-bc09-4959-b70f-9643ab0b23c8 container test-container: 
 STEP: delete the pod
-Sep 21 16:46:21.014: INFO: Waiting for pod downwardapi-volume-328bfe61-13fd-4cfb-8156-0ca2b7b9b398 to disappear
-Sep 21 16:46:21.044: INFO: Pod downwardapi-volume-328bfe61-13fd-4cfb-8156-0ca2b7b9b398 no longer exists
-[AfterEach] [sig-storage] Downward API volume
+Oct 26 16:49:37.770: INFO: Waiting for pod pod-9b5da8c0-bc09-4959-b70f-9643ab0b23c8 to disappear
+Oct 26 16:49:37.785: INFO: Pod pod-9b5da8c0-bc09-4959-b70f-9643ab0b23c8 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:46:21.044: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-4799" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance]","total":280,"completed":180,"skipped":3119,"failed":0}
-SSSSSSSSSSSS
+Oct 26 16:49:37.786: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-8213" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":183,"skipped":2864,"failed":0}
+SS
 ------------------------------
-[k8s.io] InitContainer [NodeConformance] 
-  should invoke init containers on a RestartNever pod [Conformance]
+[sig-apps] Job 
+  should adopt matching orphans and release non-matching pods [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+[BeforeEach] [sig-apps] Job
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:46:21.121: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename init-container
+Oct 26 16:49:37.825: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename job
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
-[It] should invoke init containers on a RestartNever pod [Conformance]
+[It] should adopt matching orphans and release non-matching pods [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-Sep 21 16:46:21.413: INFO: PodSpec: initContainers in spec.initContainers
-[AfterEach] [k8s.io] InitContainer [NodeConformance]
+STEP: Creating a job
+STEP: Ensuring active pods == parallelism
+STEP: Orphaning one of the Job's Pods
+Oct 26 16:49:42.608: INFO: Successfully updated pod "adopt-release-cczrp"
+STEP: Checking that the Job readopts the Pod
+Oct 26 16:49:42.608: INFO: Waiting up to 15m0s for pod "adopt-release-cczrp" in namespace "job-1010" to be "adopted"
+Oct 26 16:49:42.625: INFO: Pod "adopt-release-cczrp": Phase="Running", Reason="", readiness=true. Elapsed: 16.606312ms
+Oct 26 16:49:44.642: INFO: Pod "adopt-release-cczrp": Phase="Running", Reason="", readiness=true. Elapsed: 2.033872272s
+Oct 26 16:49:44.642: INFO: Pod "adopt-release-cczrp" satisfied condition "adopted"
+STEP: Removing the labels from the Job's Pod
+Oct 26 16:49:45.201: INFO: Successfully updated pod "adopt-release-cczrp"
+STEP: Checking that the Job releases the Pod
+Oct 26 16:49:45.201: INFO: Waiting up to 15m0s for pod "adopt-release-cczrp" in namespace "job-1010" to be "released"
+Oct 26 16:49:45.221: INFO: Pod "adopt-release-cczrp": Phase="Running", Reason="", readiness=true. Elapsed: 19.043285ms
+Oct 26 16:49:45.221: INFO: Pod "adopt-release-cczrp" satisfied condition "released"
+[AfterEach] [sig-apps] Job
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:46:25.556: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "init-container-6023" for this suite.
-•{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance]","total":280,"completed":181,"skipped":3131,"failed":0}
-SSSSSSS
-------------------------------
-[k8s.io] Container Runtime blackbox test on terminated container 
-  should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:46:25.637: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-runtime
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
+Oct 26 16:49:45.221: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "job-1010" for this suite.
+
+• [SLOW TEST:7.451 seconds]
+[sig-apps] Job
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should adopt matching orphans and release non-matching pods [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the container
-STEP: wait for the container to reach Succeeded
-STEP: get the container status
-STEP: the container should be terminated
-STEP: the termination message should be set
-Sep 21 16:46:29.333: INFO: Expected: &{} to match Container's Termination Message:  --
-STEP: delete the container
-[AfterEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:46:29.446: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-runtime-8210" for this suite.
-•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":280,"completed":182,"skipped":3138,"failed":0}
-SSSSSSSSSSSSS
 ------------------------------
-[sig-apps] ReplicaSet 
-  should adopt matching pods on creation and release no longer matching pods [Conformance]
+{"msg":"PASSED [sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance]","total":280,"completed":184,"skipped":2866,"failed":0}
+SS
+------------------------------
+[k8s.io] Security Context when creating containers with AllowPrivilegeEscalation 
+  should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] ReplicaSet
+[BeforeEach] [k8s.io] Security Context
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:46:29.573: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename replicaset
+Oct 26 16:49:45.278: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename security-context-test
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should adopt matching pods on creation and release no longer matching pods [Conformance]
+[BeforeEach] [k8s.io] Security Context
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39
+[It] should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Given a Pod with a 'name' label pod-adoption-release is created
-STEP: When a replicaset with a matching selector is created
-STEP: Then the orphan pod is adopted
-STEP: When the matched label of one of its pods change
-Sep 21 16:46:33.196: INFO: Pod name pod-adoption-release: Found 1 pods out of 1
-STEP: Then the pod is released
-[AfterEach] [sig-apps] ReplicaSet
+Oct 26 16:49:45.549: INFO: Waiting up to 5m0s for pod "alpine-nnp-false-c77ef131-03dc-4503-acc4-fa19bb5706e4" in namespace "security-context-test-9392" to be "success or failure"
+Oct 26 16:49:45.567: INFO: Pod "alpine-nnp-false-c77ef131-03dc-4503-acc4-fa19bb5706e4": Phase="Pending", Reason="", readiness=false. Elapsed: 17.625067ms
+Oct 26 16:49:47.585: INFO: Pod "alpine-nnp-false-c77ef131-03dc-4503-acc4-fa19bb5706e4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035360106s
+Oct 26 16:49:49.603: INFO: Pod "alpine-nnp-false-c77ef131-03dc-4503-acc4-fa19bb5706e4": Phase="Pending", Reason="", readiness=false. Elapsed: 4.053579661s
+Oct 26 16:49:51.619: INFO: Pod "alpine-nnp-false-c77ef131-03dc-4503-acc4-fa19bb5706e4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.070066727s
+Oct 26 16:49:51.619: INFO: Pod "alpine-nnp-false-c77ef131-03dc-4503-acc4-fa19bb5706e4" satisfied condition "success or failure"
+[AfterEach] [k8s.io] Security Context
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:46:33.319: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replicaset-2870" for this suite.
-•{"msg":"PASSED [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance]","total":280,"completed":183,"skipped":3151,"failed":0}
-SSSSSSSSSSSSSSS
+Oct 26 16:49:51.651: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "security-context-test-9392" for this suite.
+
+• [SLOW TEST:6.408 seconds]
+[k8s.io] Security Context
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  when creating containers with AllowPrivilegeEscalation
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:289
+    should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+{"msg":"PASSED [k8s.io] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":185,"skipped":2868,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with projected pod [LinuxOnly] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [sig-storage] Subpath
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:46:33.401: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename emptydir
+Oct 26 16:49:51.687: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename subpath
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
+STEP: Setting up data
+[It] should support subpaths with projected pod [LinuxOnly] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0644 on tmpfs
-Sep 21 16:46:33.963: INFO: Waiting up to 5m0s for pod "pod-47959133-2b40-461d-925d-d3d361fd686a" in namespace "emptydir-4142" to be "success or failure"
-Sep 21 16:46:33.989: INFO: Pod "pod-47959133-2b40-461d-925d-d3d361fd686a": Phase="Pending", Reason="", readiness=false. Elapsed: 25.76146ms
-Sep 21 16:46:36.014: INFO: Pod "pod-47959133-2b40-461d-925d-d3d361fd686a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.050932051s
+STEP: Creating pod pod-subpath-test-projected-sdch
+STEP: Creating a pod to test atomic-volume-subpath
+Oct 26 16:49:51.917: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-sdch" in namespace "subpath-1059" to be "success or failure"
+Oct 26 16:49:51.939: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Pending", Reason="", readiness=false. Elapsed: 21.902407ms
+Oct 26 16:49:53.956: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Pending", Reason="", readiness=false. Elapsed: 2.039048096s
+Oct 26 16:49:55.972: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Running", Reason="", readiness=true. Elapsed: 4.055153281s
+Oct 26 16:49:57.994: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Running", Reason="", readiness=true. Elapsed: 6.077369911s
+Oct 26 16:50:00.011: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Running", Reason="", readiness=true. Elapsed: 8.09454684s
+Oct 26 16:50:02.029: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Running", Reason="", readiness=true. Elapsed: 10.112154679s
+Oct 26 16:50:04.046: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Running", Reason="", readiness=true. Elapsed: 12.129238914s
+Oct 26 16:50:06.064: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Running", Reason="", readiness=true. Elapsed: 14.146829542s
+Oct 26 16:50:08.080: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Running", Reason="", readiness=true. Elapsed: 16.162821294s
+Oct 26 16:50:10.101: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Running", Reason="", readiness=true. Elapsed: 18.184581258s
+Oct 26 16:50:12.118: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Running", Reason="", readiness=true. Elapsed: 20.201377236s
+Oct 26 16:50:14.136: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Running", Reason="", readiness=true. Elapsed: 22.218682724s
+Oct 26 16:50:16.153: INFO: Pod "pod-subpath-test-projected-sdch": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.235887982s
 STEP: Saw pod success
-Sep 21 16:46:36.015: INFO: Pod "pod-47959133-2b40-461d-925d-d3d361fd686a" satisfied condition "success or failure"
-Sep 21 16:46:36.037: INFO: Trying to get logs from node 10.241.51.147 pod pod-47959133-2b40-461d-925d-d3d361fd686a container test-container: 
+Oct 26 16:50:16.153: INFO: Pod "pod-subpath-test-projected-sdch" satisfied condition "success or failure"
+Oct 26 16:50:16.169: INFO: Trying to get logs from node 10.72.119.74 pod pod-subpath-test-projected-sdch container test-container-subpath-projected-sdch: 
 STEP: delete the pod
-Sep 21 16:46:36.452: INFO: Waiting for pod pod-47959133-2b40-461d-925d-d3d361fd686a to disappear
-Sep 21 16:46:36.485: INFO: Pod pod-47959133-2b40-461d-925d-d3d361fd686a no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+Oct 26 16:50:16.269: INFO: Waiting for pod pod-subpath-test-projected-sdch to disappear
+Oct 26 16:50:16.286: INFO: Pod pod-subpath-test-projected-sdch no longer exists
+STEP: Deleting pod pod-subpath-test-projected-sdch
+Oct 26 16:50:16.287: INFO: Deleting pod "pod-subpath-test-projected-sdch" in namespace "subpath-1059"
+[AfterEach] [sig-storage] Subpath
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:46:36.485: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-4142" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":184,"skipped":3166,"failed":0}
-SSSSSSSSSSSSSS
+Oct 26 16:50:16.303: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-1059" for this suite.
+
+• [SLOW TEST:24.683 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
+  Atomic writer volumes
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
+    should support subpaths with projected pod [LinuxOnly] [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-network] DNS 
-  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
+{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance]","total":280,"completed":186,"skipped":2891,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] Daemon set [Serial] 
+  should run and stop simple daemon [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] DNS
+[BeforeEach] [sig-apps] Daemon set [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:46:36.570: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename dns
+Oct 26 16:50:16.371: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename daemonsets
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
+[It] should run and stop simple daemon [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-6309.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-6309.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6309.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
-
-STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-6309.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-6309.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6309.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+STEP: Creating simple DaemonSet "daemon-set"
+STEP: Check that daemon pods launch on every node of the cluster.
+Oct 26 16:50:16.662: INFO: Number of nodes with available pods: 0
+Oct 26 16:50:16.662: INFO: Node 10.72.119.74 is running more than one daemon pod
+Oct 26 16:50:17.700: INFO: Number of nodes with available pods: 0
+Oct 26 16:50:17.700: INFO: Node 10.72.119.74 is running more than one daemon pod
+Oct 26 16:50:18.698: INFO: Number of nodes with available pods: 0
+Oct 26 16:50:18.698: INFO: Node 10.72.119.74 is running more than one daemon pod
+Oct 26 16:50:19.700: INFO: Number of nodes with available pods: 3
+Oct 26 16:50:19.700: INFO: Number of running nodes: 3, number of available pods: 3
+STEP: Stop a daemon pod, check that the daemon pod is revived.
+Oct 26 16:50:19.819: INFO: Number of nodes with available pods: 2
+Oct 26 16:50:19.819: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:50:20.855: INFO: Number of nodes with available pods: 2
+Oct 26 16:50:20.855: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:50:21.856: INFO: Number of nodes with available pods: 2
+Oct 26 16:50:21.856: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:50:22.859: INFO: Number of nodes with available pods: 2
+Oct 26 16:50:22.859: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:50:23.853: INFO: Number of nodes with available pods: 2
+Oct 26 16:50:23.853: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:50:24.857: INFO: Number of nodes with available pods: 2
+Oct 26 16:50:24.857: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 16:50:25.852: INFO: Number of nodes with available pods: 3
+Oct 26 16:50:25.852: INFO: Number of running nodes: 3, number of available pods: 3
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-5302, will wait for the garbage collector to delete the pods
+Oct 26 16:50:25.975: INFO: Deleting DaemonSet.extensions daemon-set took: 40.10559ms
+Oct 26 16:50:26.075: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.25057ms
+Oct 26 16:50:39.592: INFO: Number of nodes with available pods: 0
+Oct 26 16:50:39.592: INFO: Number of running nodes: 0, number of available pods: 0
+Oct 26 16:50:39.607: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-5302/daemonsets","resourceVersion":"76917"},"items":null}
 
-STEP: creating a pod to probe /etc/hosts
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Sep 21 16:46:49.431: INFO: DNS probes using dns-6309/dns-test-23af684d-2476-48c9-9021-2fd8b139aa39 succeeded
+Oct 26 16:50:39.622: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-5302/pods","resourceVersion":"76917"},"items":null}
 
-STEP: deleting the pod
-[AfterEach] [sig-network] DNS
+[AfterEach] [sig-apps] Daemon set [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:46:49.554: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-6309" for this suite.
+Oct 26 16:50:39.681: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "daemonsets-5302" for this suite.
 
-• [SLOW TEST:13.053 seconds]
-[sig-network] DNS
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
+• [SLOW TEST:23.346 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should run and stop simple daemon [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] DNS should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]","total":280,"completed":185,"skipped":3180,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-network] Services 
-  should be able to create a functioning NodePort service [Conformance]
+{"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance]","total":280,"completed":187,"skipped":2918,"failed":0}
+[sig-apps] ReplicaSet 
+  should serve a basic image on each replica with a public image  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Services
+[BeforeEach] [sig-apps] ReplicaSet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:46:49.626: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename services
+Oct 26 16:50:39.717: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename replicaset
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should be able to create a functioning NodePort service [Conformance]
+[It] should serve a basic image on each replica with a public image  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating service nodeport-test with type=NodePort in namespace services-4587
-STEP: creating replication controller nodeport-test in namespace services-4587
-I0921 16:46:50.061787      24 runners.go:189] Created replication controller with name: nodeport-test, namespace: services-4587, replica count: 2
-Sep 21 16:46:53.112: INFO: Creating new exec pod
-I0921 16:46:53.112190      24 runners.go:189] nodeport-test Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-Sep 21 16:46:58.294: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-4587 execpodjtcqf -- /bin/sh -x -c nc -zv -t -w 2 nodeport-test 80'
-Sep 21 16:46:58.824: INFO: stderr: "+ nc -zv -t -w 2 nodeport-test 80\nConnection to nodeport-test 80 port [tcp/http] succeeded!\n"
-Sep 21 16:46:58.824: INFO: stdout: ""
-Sep 21 16:46:58.824: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-4587 execpodjtcqf -- /bin/sh -x -c nc -zv -t -w 2 172.21.224.99 80'
-Sep 21 16:46:59.201: INFO: stderr: "+ nc -zv -t -w 2 172.21.224.99 80\nConnection to 172.21.224.99 80 port [tcp/http] succeeded!\n"
-Sep 21 16:46:59.201: INFO: stdout: ""
-Sep 21 16:46:59.201: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-4587 execpodjtcqf -- /bin/sh -x -c nc -zv -t -w 2 10.241.51.150 31788'
-Sep 21 16:46:59.591: INFO: stderr: "+ nc -zv -t -w 2 10.241.51.150 31788\nConnection to 10.241.51.150 31788 port [tcp/31788] succeeded!\n"
-Sep 21 16:46:59.591: INFO: stdout: ""
-Sep 21 16:46:59.591: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-4587 execpodjtcqf -- /bin/sh -x -c nc -zv -t -w 2 10.241.51.147 31788'
-Sep 21 16:46:59.967: INFO: stderr: "+ nc -zv -t -w 2 10.241.51.147 31788\nConnection to 10.241.51.147 31788 port [tcp/31788] succeeded!\n"
-Sep 21 16:46:59.967: INFO: stdout: ""
-Sep 21 16:46:59.967: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-4587 execpodjtcqf -- /bin/sh -x -c nc -zv -t -w 2 169.48.196.37 31788'
-Sep 21 16:47:00.384: INFO: stderr: "+ nc -zv -t -w 2 169.48.196.37 31788\nConnection to 169.48.196.37 31788 port [tcp/31788] succeeded!\n"
-Sep 21 16:47:00.384: INFO: stdout: ""
-Sep 21 16:47:00.384: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-4587 execpodjtcqf -- /bin/sh -x -c nc -zv -t -w 2 169.48.196.44 31788'
-Sep 21 16:47:00.756: INFO: stderr: "+ nc -zv -t -w 2 169.48.196.44 31788\nConnection to 169.48.196.44 31788 port [tcp/31788] succeeded!\n"
-Sep 21 16:47:00.756: INFO: stdout: ""
-[AfterEach] [sig-network] Services
+Oct 26 16:50:39.851: INFO: Creating ReplicaSet my-hostname-basic-9b08988a-eadb-4df9-aa04-df0966decde9
+Oct 26 16:50:39.900: INFO: Pod name my-hostname-basic-9b08988a-eadb-4df9-aa04-df0966decde9: Found 0 pods out of 1
+Oct 26 16:50:44.919: INFO: Pod name my-hostname-basic-9b08988a-eadb-4df9-aa04-df0966decde9: Found 1 pods out of 1
+Oct 26 16:50:44.919: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-9b08988a-eadb-4df9-aa04-df0966decde9" is running
+Oct 26 16:50:44.935: INFO: Pod "my-hostname-basic-9b08988a-eadb-4df9-aa04-df0966decde9-gjzhz" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 16:50:39 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 16:50:42 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 16:50:42 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-10-26 16:50:39 +0000 UTC Reason: Message:}])
+Oct 26 16:50:44.935: INFO: Trying to dial the pod
+Oct 26 16:50:50.003: INFO: Controller my-hostname-basic-9b08988a-eadb-4df9-aa04-df0966decde9: Got expected result from replica 1 [my-hostname-basic-9b08988a-eadb-4df9-aa04-df0966decde9-gjzhz]: "my-hostname-basic-9b08988a-eadb-4df9-aa04-df0966decde9-gjzhz", 1 of 1 required successes so far
+[AfterEach] [sig-apps] ReplicaSet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:00.756: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-4587" for this suite.
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
+Oct 26 16:50:50.003: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replicaset-2205" for this suite.
 
-• [SLOW TEST:11.195 seconds]
-[sig-network] Services
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should be able to create a functioning NodePort service [Conformance]
+• [SLOW TEST:10.324 seconds]
+[sig-apps] ReplicaSet
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should serve a basic image on each replica with a public image  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Services should be able to create a functioning NodePort service [Conformance]","total":280,"completed":186,"skipped":3204,"failed":0}
-S
-------------------------------
-[sig-storage] Projected secret 
-  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:47:00.822: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name projected-secret-test-072c5192-507e-4d68-999b-4fbed7b19f21
-STEP: Creating a pod to test consume secrets
-Sep 21 16:47:01.199: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-426aeb2a-cc92-42eb-a17d-8333b57c1c27" in namespace "projected-5733" to be "success or failure"
-Sep 21 16:47:01.246: INFO: Pod "pod-projected-secrets-426aeb2a-cc92-42eb-a17d-8333b57c1c27": Phase="Pending", Reason="", readiness=false. Elapsed: 46.943802ms
-Sep 21 16:47:03.264: INFO: Pod "pod-projected-secrets-426aeb2a-cc92-42eb-a17d-8333b57c1c27": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.064246471s
-STEP: Saw pod success
-Sep 21 16:47:03.264: INFO: Pod "pod-projected-secrets-426aeb2a-cc92-42eb-a17d-8333b57c1c27" satisfied condition "success or failure"
-Sep 21 16:47:03.284: INFO: Trying to get logs from node 10.241.51.147 pod pod-projected-secrets-426aeb2a-cc92-42eb-a17d-8333b57c1c27 container secret-volume-test: 
-STEP: delete the pod
-Sep 21 16:47:03.402: INFO: Waiting for pod pod-projected-secrets-426aeb2a-cc92-42eb-a17d-8333b57c1c27 to disappear
-Sep 21 16:47:03.433: INFO: Pod pod-projected-secrets-426aeb2a-cc92-42eb-a17d-8333b57c1c27 no longer exists
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:03.433: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-5733" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":280,"completed":187,"skipped":3205,"failed":0}
+{"msg":"PASSED [sig-apps] ReplicaSet should serve a basic image on each replica with a public image  [Conformance]","total":280,"completed":188,"skipped":2918,"failed":0}
 SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected configMap 
-  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
+[sig-api-machinery] Namespaces [Serial] 
+  should ensure that all services are removed when a namespace is deleted [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
+[BeforeEach] [sig-api-machinery] Namespaces [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:47:03.489: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 16:50:50.044: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename namespaces
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
+[It] should ensure that all services are removed when a namespace is deleted [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name projected-configmap-test-volume-map-8acc95ac-7aa6-4378-aa59-9c2551da6180
-STEP: Creating a pod to test consume configMaps
-Sep 21 16:47:03.841: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-6328df19-71c9-41e9-8ef8-067dcd314dc2" in namespace "projected-6945" to be "success or failure"
-Sep 21 16:47:03.867: INFO: Pod "pod-projected-configmaps-6328df19-71c9-41e9-8ef8-067dcd314dc2": Phase="Pending", Reason="", readiness=false. Elapsed: 26.626936ms
-Sep 21 16:47:05.888: INFO: Pod "pod-projected-configmaps-6328df19-71c9-41e9-8ef8-067dcd314dc2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.047386836s
-STEP: Saw pod success
-Sep 21 16:47:05.888: INFO: Pod "pod-projected-configmaps-6328df19-71c9-41e9-8ef8-067dcd314dc2" satisfied condition "success or failure"
-Sep 21 16:47:05.909: INFO: Trying to get logs from node 10.241.51.147 pod pod-projected-configmaps-6328df19-71c9-41e9-8ef8-067dcd314dc2 container projected-configmap-volume-test: 
-STEP: delete the pod
-Sep 21 16:47:06.041: INFO: Waiting for pod pod-projected-configmaps-6328df19-71c9-41e9-8ef8-067dcd314dc2 to disappear
-Sep 21 16:47:06.275: INFO: Pod pod-projected-configmaps-6328df19-71c9-41e9-8ef8-067dcd314dc2 no longer exists
-[AfterEach] [sig-storage] Projected configMap
+STEP: Creating a test namespace
+STEP: Waiting for a default service account to be provisioned in namespace
+STEP: Creating a service in the namespace
+STEP: Deleting the namespace
+STEP: Waiting for the namespace to be removed.
+STEP: Recreating the namespace
+STEP: Verifying there is no service in the namespace
+[AfterEach] [sig-api-machinery] Namespaces [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:06.275: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-6945" for this suite.
-•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","total":280,"completed":188,"skipped":3238,"failed":0}
+Oct 26 16:50:57.507: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "namespaces-3367" for this suite.
+STEP: Destroying namespace "nsdeletetest-5790" for this suite.
+Oct 26 16:50:57.561: INFO: Namespace nsdeletetest-5790 was already deleted
+STEP: Destroying namespace "nsdeletetest-6768" for this suite.
+
+• [SLOW TEST:7.534 seconds]
+[sig-api-machinery] Namespaces [Serial]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should ensure that all services are removed when a namespace is deleted [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance]","total":280,"completed":189,"skipped":2951,"failed":0}
 SSS
 ------------------------------
 [sig-cli] Kubectl client Kubectl patch 
@@ -9571,7 +8751,7 @@ SSS
 [BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:47:06.333: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 16:50:57.580: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename kubectl
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
@@ -9579,256 +8759,249 @@ STEP: Waiting for a default service account to be provisioned in namespace
 [It] should add annotations for pods in rc  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: creating Agnhost RC
-Sep 21 16:47:06.557: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-2445'
-Sep 21 16:47:06.860: INFO: stderr: ""
-Sep 21 16:47:06.860: INFO: stdout: "replicationcontroller/agnhost-master created\n"
+Oct 26 16:50:57.723: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-5372'
+Oct 26 16:50:58.326: INFO: stderr: ""
+Oct 26 16:50:58.326: INFO: stdout: "replicationcontroller/agnhost-master created\n"
 STEP: Waiting for Agnhost master to start.
-Sep 21 16:47:07.881: INFO: Selector matched 1 pods for map[app:agnhost]
-Sep 21 16:47:07.881: INFO: Found 0 / 1
-Sep 21 16:47:08.885: INFO: Selector matched 1 pods for map[app:agnhost]
-Sep 21 16:47:08.885: INFO: Found 1 / 1
-Sep 21 16:47:08.885: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+Oct 26 16:50:59.344: INFO: Selector matched 1 pods for map[app:agnhost]
+Oct 26 16:50:59.344: INFO: Found 0 / 1
+Oct 26 16:51:00.343: INFO: Selector matched 1 pods for map[app:agnhost]
+Oct 26 16:51:00.343: INFO: Found 0 / 1
+Oct 26 16:51:01.343: INFO: Selector matched 1 pods for map[app:agnhost]
+Oct 26 16:51:01.343: INFO: Found 1 / 1
+Oct 26 16:51:01.343: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
 STEP: patching all pods
-Sep 21 16:47:08.910: INFO: Selector matched 1 pods for map[app:agnhost]
-Sep 21 16:47:08.910: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
-Sep 21 16:47:08.910: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 patch pod agnhost-master-9kv6c --namespace=kubectl-2445 -p {"metadata":{"annotations":{"x":"y"}}}'
-Sep 21 16:47:09.139: INFO: stderr: ""
-Sep 21 16:47:09.139: INFO: stdout: "pod/agnhost-master-9kv6c patched\n"
+Oct 26 16:51:01.360: INFO: Selector matched 1 pods for map[app:agnhost]
+Oct 26 16:51:01.360: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Oct 26 16:51:01.360: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 patch pod agnhost-master-6c6pz --namespace=kubectl-5372 -p {"metadata":{"annotations":{"x":"y"}}}'
+Oct 26 16:51:01.547: INFO: stderr: ""
+Oct 26 16:51:01.547: INFO: stdout: "pod/agnhost-master-6c6pz patched\n"
 STEP: checking annotations
-Sep 21 16:47:09.168: INFO: Selector matched 1 pods for map[app:agnhost]
-Sep 21 16:47:09.168: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Oct 26 16:51:01.565: INFO: Selector matched 1 pods for map[app:agnhost]
+Oct 26 16:51:01.565: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
 [AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:09.168: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-2445" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc  [Conformance]","total":280,"completed":189,"skipped":3241,"failed":0}
-SSSS
+Oct 26 16:51:01.565: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-5372" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc  [Conformance]","total":280,"completed":190,"skipped":2954,"failed":0}
+SSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
+[k8s.io] Probing container 
+  should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 16:51:01.598: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename container-probe
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating pod test-webserver-a4b3987b-6112-4b69-9bb4-68bbbfb378a1 in namespace container-probe-4300
+Oct 26 16:51:03.855: INFO: Started pod test-webserver-a4b3987b-6112-4b69-9bb4-68bbbfb378a1 in namespace container-probe-4300
+STEP: checking the pod's current state and verifying that restartCount is present
+Oct 26 16:51:03.871: INFO: Initial restart count of pod test-webserver-a4b3987b-6112-4b69-9bb4-68bbbfb378a1 is 0
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 16:55:04.059: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-4300" for this suite.
+
+• [SLOW TEST:242.497 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":280,"completed":191,"skipped":2967,"failed":0}
+SSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Downward API volume 
   should provide container's cpu request [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
+[BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:47:09.237: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 16:55:04.098: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename downward-api
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
 [It] should provide container's cpu request [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:47:09.708: INFO: Waiting up to 5m0s for pod "downwardapi-volume-66eec4ca-6a28-4a5e-babf-ffb38fd91d24" in namespace "projected-6858" to be "success or failure"
-Sep 21 16:47:09.728: INFO: Pod "downwardapi-volume-66eec4ca-6a28-4a5e-babf-ffb38fd91d24": Phase="Pending", Reason="", readiness=false. Elapsed: 19.559489ms
-Sep 21 16:47:11.753: INFO: Pod "downwardapi-volume-66eec4ca-6a28-4a5e-babf-ffb38fd91d24": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.04464801s
+Oct 26 16:55:04.301: INFO: Waiting up to 5m0s for pod "downwardapi-volume-4538776b-7a3d-4ee1-815e-1413029f692c" in namespace "downward-api-3316" to be "success or failure"
+Oct 26 16:55:04.319: INFO: Pod "downwardapi-volume-4538776b-7a3d-4ee1-815e-1413029f692c": Phase="Pending", Reason="", readiness=false. Elapsed: 17.562689ms
+Oct 26 16:55:06.335: INFO: Pod "downwardapi-volume-4538776b-7a3d-4ee1-815e-1413029f692c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.033623839s
 STEP: Saw pod success
-Sep 21 16:47:11.753: INFO: Pod "downwardapi-volume-66eec4ca-6a28-4a5e-babf-ffb38fd91d24" satisfied condition "success or failure"
-Sep 21 16:47:11.784: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-66eec4ca-6a28-4a5e-babf-ffb38fd91d24 container client-container: 
+Oct 26 16:55:06.335: INFO: Pod "downwardapi-volume-4538776b-7a3d-4ee1-815e-1413029f692c" satisfied condition "success or failure"
+Oct 26 16:55:06.350: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-4538776b-7a3d-4ee1-815e-1413029f692c container client-container: 
 STEP: delete the pod
-Sep 21 16:47:11.914: INFO: Waiting for pod downwardapi-volume-66eec4ca-6a28-4a5e-babf-ffb38fd91d24 to disappear
-Sep 21 16:47:11.947: INFO: Pod downwardapi-volume-66eec4ca-6a28-4a5e-babf-ffb38fd91d24 no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
+Oct 26 16:55:06.468: INFO: Waiting for pod downwardapi-volume-4538776b-7a3d-4ee1-815e-1413029f692c to disappear
+Oct 26 16:55:06.482: INFO: Pod downwardapi-volume-4538776b-7a3d-4ee1-815e-1413029f692c no longer exists
+[AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:11.947: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-6858" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance]","total":280,"completed":190,"skipped":3245,"failed":0}
-SSSSSSSSSSS
+Oct 26 16:55:06.483: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-3316" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance]","total":280,"completed":192,"skipped":2986,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume [NodeConformance] [Conformance]
+[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch 
+  watch on custom resource definition objects [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:47:12.033: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 16:55:06.529: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename crd-watch
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+[It] watch on custom resource definition objects [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating projection with secret that has name projected-secret-test-5c358b57-1629-4150-b608-246396295638
-STEP: Creating a pod to test consume secrets
-Sep 21 16:47:12.488: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-d54dcf08-4204-42f4-9977-5fafa6c44d94" in namespace "projected-7129" to be "success or failure"
-Sep 21 16:47:12.515: INFO: Pod "pod-projected-secrets-d54dcf08-4204-42f4-9977-5fafa6c44d94": Phase="Pending", Reason="", readiness=false. Elapsed: 27.567065ms
-Sep 21 16:47:14.543: INFO: Pod "pod-projected-secrets-d54dcf08-4204-42f4-9977-5fafa6c44d94": Phase="Pending", Reason="", readiness=false. Elapsed: 2.055549921s
-Sep 21 16:47:16.569: INFO: Pod "pod-projected-secrets-d54dcf08-4204-42f4-9977-5fafa6c44d94": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.081813579s
-STEP: Saw pod success
-Sep 21 16:47:16.570: INFO: Pod "pod-projected-secrets-d54dcf08-4204-42f4-9977-5fafa6c44d94" satisfied condition "success or failure"
-Sep 21 16:47:16.594: INFO: Trying to get logs from node 10.241.51.147 pod pod-projected-secrets-d54dcf08-4204-42f4-9977-5fafa6c44d94 container projected-secret-volume-test: 
-STEP: delete the pod
-Sep 21 16:47:16.777: INFO: Waiting for pod pod-projected-secrets-d54dcf08-4204-42f4-9977-5fafa6c44d94 to disappear
-Sep 21 16:47:16.807: INFO: Pod pod-projected-secrets-d54dcf08-4204-42f4-9977-5fafa6c44d94 no longer exists
-[AfterEach] [sig-storage] Projected secret
+Oct 26 16:55:06.678: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Creating first CR 
+Oct 26 16:55:07.412: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-10-26T16:55:07Z generation:1 name:name1 resourceVersion:78536 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:529872bb-0c67-464d-9c6e-00fbf0cd9670] num:map[num1:9223372036854775807 num2:1000000]]}
+STEP: Creating second CR
+Oct 26 16:55:17.430: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-10-26T16:55:17Z generation:1 name:name2 resourceVersion:78620 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:2534caad-a2e6-4719-ae17-2616aec95a80] num:map[num1:9223372036854775807 num2:1000000]]}
+STEP: Modifying first CR
+Oct 26 16:55:27.451: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-10-26T16:55:07Z generation:2 name:name1 resourceVersion:78660 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:529872bb-0c67-464d-9c6e-00fbf0cd9670] num:map[num1:9223372036854775807 num2:1000000]]}
+STEP: Modifying second CR
+Oct 26 16:55:37.468: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-10-26T16:55:17Z generation:2 name:name2 resourceVersion:78701 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:2534caad-a2e6-4719-ae17-2616aec95a80] num:map[num1:9223372036854775807 num2:1000000]]}
+STEP: Deleting first CR
+Oct 26 16:55:47.503: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-10-26T16:55:07Z generation:2 name:name1 resourceVersion:78743 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:529872bb-0c67-464d-9c6e-00fbf0cd9670] num:map[num1:9223372036854775807 num2:1000000]]}
+STEP: Deleting second CR
+Oct 26 16:55:57.534: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-10-26T16:55:17Z generation:2 name:name2 resourceVersion:78787 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:2534caad-a2e6-4719-ae17-2616aec95a80] num:map[num1:9223372036854775807 num2:1000000]]}
+[AfterEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:16.807: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-7129" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":191,"skipped":3256,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 16:56:08.082: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-watch-3924" for this suite.
+
+• [SLOW TEST:61.603 seconds]
+[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  CustomResourceDefinition Watch
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_watch.go:41
+    watch on custom resource definition objects [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  pod should support shared volumes between containers [Conformance]
+{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance]","total":280,"completed":193,"skipped":3034,"failed":0}
+SSSSSSSSSS
+------------------------------
+[sig-storage] Downward API volume 
+  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:47:16.875: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename emptydir
+Oct 26 16:56:08.134: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename downward-api
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] pod should support shared volumes between containers [Conformance]
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating Pod
-STEP: Waiting for the pod running
-STEP: Geting the pod
-STEP: Reading file content from the nginx-container
-Sep 21 16:47:21.322: INFO: ExecWithOptions {Command:[/bin/sh -c cat /usr/share/volumeshare/shareddata.txt] Namespace:emptydir-2208 PodName:pod-sharedvolume-b5183867-db16-41f5-aa2d-747eb6fe3450 ContainerName:busybox-main-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 16:47:21.322: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:47:21.553: INFO: Exec stderr: ""
-[AfterEach] [sig-storage] EmptyDir volumes
+STEP: Creating a pod to test downward API volume plugin
+Oct 26 16:56:08.324: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f5363f6c-4f8e-403d-a0ed-a8879f819803" in namespace "downward-api-1283" to be "success or failure"
+Oct 26 16:56:08.341: INFO: Pod "downwardapi-volume-f5363f6c-4f8e-403d-a0ed-a8879f819803": Phase="Pending", Reason="", readiness=false. Elapsed: 16.456457ms
+Oct 26 16:56:10.358: INFO: Pod "downwardapi-volume-f5363f6c-4f8e-403d-a0ed-a8879f819803": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03374537s
+Oct 26 16:56:12.374: INFO: Pod "downwardapi-volume-f5363f6c-4f8e-403d-a0ed-a8879f819803": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.049990105s
+STEP: Saw pod success
+Oct 26 16:56:12.374: INFO: Pod "downwardapi-volume-f5363f6c-4f8e-403d-a0ed-a8879f819803" satisfied condition "success or failure"
+Oct 26 16:56:12.390: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-f5363f6c-4f8e-403d-a0ed-a8879f819803 container client-container: 
+STEP: delete the pod
+Oct 26 16:56:12.475: INFO: Waiting for pod downwardapi-volume-f5363f6c-4f8e-403d-a0ed-a8879f819803 to disappear
+Oct 26 16:56:12.491: INFO: Pod downwardapi-volume-f5363f6c-4f8e-403d-a0ed-a8879f819803 no longer exists
+[AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:21.553: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-2208" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance]","total":280,"completed":192,"skipped":3281,"failed":0}
-SSSSSSSSSSSS
+Oct 26 16:56:12.491: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-1283" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":194,"skipped":3044,"failed":0}
+SSSSS
 ------------------------------
-[k8s.io] Pods 
-  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+[k8s.io] Probing container 
+  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Pods
+[BeforeEach] [k8s.io] Probing container
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:47:21.628: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename pods
+Oct 26 16:56:12.535: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename container-probe
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
-[It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-STEP: submitting the pod to kubernetes
-STEP: verifying the pod is in kubernetes
-STEP: updating the pod
-Sep 21 16:47:24.804: INFO: Successfully updated pod "pod-update-activedeadlineseconds-d6bd7300-62a6-4e10-bda2-b5ba20f28696"
-Sep 21 16:47:24.804: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-d6bd7300-62a6-4e10-bda2-b5ba20f28696" in namespace "pods-8862" to be "terminated due to deadline exceeded"
-Sep 21 16:47:24.828: INFO: Pod "pod-update-activedeadlineseconds-d6bd7300-62a6-4e10-bda2-b5ba20f28696": Phase="Running", Reason="", readiness=true. Elapsed: 23.872242ms
-Sep 21 16:47:26.852: INFO: Pod "pod-update-activedeadlineseconds-d6bd7300-62a6-4e10-bda2-b5ba20f28696": Phase="Running", Reason="", readiness=true. Elapsed: 2.04778903s
-Sep 21 16:47:28.878: INFO: Pod "pod-update-activedeadlineseconds-d6bd7300-62a6-4e10-bda2-b5ba20f28696": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 4.074474129s
-Sep 21 16:47:28.879: INFO: Pod "pod-update-activedeadlineseconds-d6bd7300-62a6-4e10-bda2-b5ba20f28696" satisfied condition "terminated due to deadline exceeded"
-[AfterEach] [k8s.io] Pods
+STEP: Creating pod busybox-6567d098-4393-4a4b-811e-f8bcd19ea647 in namespace container-probe-734
+Oct 26 16:56:14.771: INFO: Started pod busybox-6567d098-4393-4a4b-811e-f8bcd19ea647 in namespace container-probe-734
+STEP: checking the pod's current state and verifying that restartCount is present
+Oct 26 16:56:14.786: INFO: Initial restart count of pod busybox-6567d098-4393-4a4b-811e-f8bcd19ea647 is 0
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:28.879: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-8862" for this suite.
+Oct 26 17:00:14.933: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-734" for this suite.
 
-• [SLOW TEST:7.318 seconds]
-[k8s.io] Pods
+• [SLOW TEST:242.452 seconds]
+[k8s.io] Probing container
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]","total":280,"completed":193,"skipped":3293,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":280,"completed":195,"skipped":3049,"failed":0}
+SSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl logs 
-  should be able to retrieve and filter logs  [Conformance]
+[sig-api-machinery] Servers with support for Table transformation 
+  should return a 406 for a backend which does not implement metadata [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-api-machinery] Servers with support for Table transformation
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:47:28.947: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubectl
+Oct 26 17:00:14.989: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename tables
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[BeforeEach] Kubectl logs
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1464
-STEP: creating an pod
-Sep 21 16:47:29.192: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 run logs-generator --generator=run-pod/v1 --image=gcr.io/kubernetes-e2e-test-images/agnhost:2.8 --namespace=kubectl-4454 -- logs-generator --log-lines-total 100 --run-duration 20s'
-Sep 21 16:47:29.437: INFO: stderr: ""
-Sep 21 16:47:29.437: INFO: stdout: "pod/logs-generator created\n"
-[It] should be able to retrieve and filter logs  [Conformance]
+[BeforeEach] [sig-api-machinery] Servers with support for Table transformation
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/table_conversion.go:46
+[It] should return a 406 for a backend which does not implement metadata [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Waiting for log generator to start.
-Sep 21 16:47:29.437: INFO: Waiting up to 5m0s for 1 pods to be running and ready, or succeeded: [logs-generator]
-Sep 21 16:47:29.437: INFO: Waiting up to 5m0s for pod "logs-generator" in namespace "kubectl-4454" to be "running and ready, or succeeded"
-Sep 21 16:47:29.462: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 25.013345ms
-Sep 21 16:47:31.498: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 2.061055291s
-Sep 21 16:47:33.526: INFO: Pod "logs-generator": Phase="Running", Reason="", readiness=true. Elapsed: 4.088554227s
-Sep 21 16:47:33.526: INFO: Pod "logs-generator" satisfied condition "running and ready, or succeeded"
-Sep 21 16:47:33.526: INFO: Wanted all 1 pods to be running and ready, or succeeded. Result: true. Pods: [logs-generator]
-STEP: checking for a matching strings
-Sep 21 16:47:33.526: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 logs logs-generator logs-generator --namespace=kubectl-4454'
-Sep 21 16:47:33.768: INFO: stderr: ""
-Sep 21 16:47:33.768: INFO: stdout: "I0921 16:47:30.809008       1 logs_generator.go:76] 0 GET /api/v1/namespaces/kube-system/pods/qmc 508\nI0921 16:47:31.009184       1 logs_generator.go:76] 1 POST /api/v1/namespaces/kube-system/pods/hhh 552\nI0921 16:47:31.209291       1 logs_generator.go:76] 2 GET /api/v1/namespaces/ns/pods/h59d 368\nI0921 16:47:31.409240       1 logs_generator.go:76] 3 PUT /api/v1/namespaces/default/pods/mg6 388\nI0921 16:47:31.609192       1 logs_generator.go:76] 4 POST /api/v1/namespaces/ns/pods/5s49 525\nI0921 16:47:31.809160       1 logs_generator.go:76] 5 GET /api/v1/namespaces/ns/pods/9h7 456\nI0921 16:47:32.009206       1 logs_generator.go:76] 6 GET /api/v1/namespaces/kube-system/pods/sg5l 292\nI0921 16:47:32.209243       1 logs_generator.go:76] 7 POST /api/v1/namespaces/default/pods/knv 478\nI0921 16:47:32.409227       1 logs_generator.go:76] 8 GET /api/v1/namespaces/ns/pods/mcbw 589\nI0921 16:47:32.609210       1 logs_generator.go:76] 9 POST /api/v1/namespaces/default/pods/bz7j 496\nI0921 16:47:32.809182       1 logs_generator.go:76] 10 GET /api/v1/namespaces/kube-system/pods/cqc4 386\nI0921 16:47:33.009186       1 logs_generator.go:76] 11 PUT /api/v1/namespaces/default/pods/sp9p 487\nI0921 16:47:33.209163       1 logs_generator.go:76] 12 PUT /api/v1/namespaces/kube-system/pods/dcr2 435\nI0921 16:47:33.409203       1 logs_generator.go:76] 13 POST /api/v1/namespaces/default/pods/js8 427\nI0921 16:47:33.610316       1 logs_generator.go:76] 14 PUT /api/v1/namespaces/kube-system/pods/kjh 348\n"
-STEP: limiting log lines
-Sep 21 16:47:33.768: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 logs logs-generator logs-generator --namespace=kubectl-4454 --tail=1'
-Sep 21 16:47:33.968: INFO: stderr: ""
-Sep 21 16:47:33.968: INFO: stdout: "I0921 16:47:33.812297       1 logs_generator.go:76] 15 POST /api/v1/namespaces/kube-system/pods/rhdf 504\n"
-Sep 21 16:47:33.968: INFO: got output "I0921 16:47:33.812297       1 logs_generator.go:76] 15 POST /api/v1/namespaces/kube-system/pods/rhdf 504\n"
-STEP: limiting log bytes
-Sep 21 16:47:33.968: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 logs logs-generator logs-generator --namespace=kubectl-4454 --limit-bytes=1'
-Sep 21 16:47:34.148: INFO: stderr: ""
-Sep 21 16:47:34.148: INFO: stdout: "I"
-Sep 21 16:47:34.148: INFO: got output "I"
-STEP: exposing timestamps
-Sep 21 16:47:34.148: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 logs logs-generator logs-generator --namespace=kubectl-4454 --tail=1 --timestamps'
-Sep 21 16:47:34.341: INFO: stderr: ""
-Sep 21 16:47:34.342: INFO: stdout: "2020-09-21T11:47:34.209544507-05:00 I0921 16:47:34.209436       1 logs_generator.go:76] 17 PUT /api/v1/namespaces/ns/pods/sg2 286\n"
-Sep 21 16:47:34.342: INFO: got output "2020-09-21T11:47:34.209544507-05:00 I0921 16:47:34.209436       1 logs_generator.go:76] 17 PUT /api/v1/namespaces/ns/pods/sg2 286\n"
-STEP: restricting to a time range
-Sep 21 16:47:36.842: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 logs logs-generator logs-generator --namespace=kubectl-4454 --since=1s'
-Sep 21 16:47:37.010: INFO: stderr: ""
-Sep 21 16:47:37.010: INFO: stdout: "I0921 16:47:36.010176       1 logs_generator.go:76] 26 GET /api/v1/namespaces/kube-system/pods/88n 531\nI0921 16:47:36.209224       1 logs_generator.go:76] 27 GET /api/v1/namespaces/kube-system/pods/xglf 222\nI0921 16:47:36.409355       1 logs_generator.go:76] 28 PUT /api/v1/namespaces/kube-system/pods/sj8n 242\nI0921 16:47:36.609224       1 logs_generator.go:76] 29 GET /api/v1/namespaces/default/pods/qjgs 279\nI0921 16:47:36.809263       1 logs_generator.go:76] 30 POST /api/v1/namespaces/kube-system/pods/n6m 593\n"
-Sep 21 16:47:37.010: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 logs logs-generator logs-generator --namespace=kubectl-4454 --since=24h'
-Sep 21 16:47:37.148: INFO: stderr: ""
-Sep 21 16:47:37.148: INFO: stdout: "I0921 16:47:30.809008       1 logs_generator.go:76] 0 GET /api/v1/namespaces/kube-system/pods/qmc 508\nI0921 16:47:31.009184       1 logs_generator.go:76] 1 POST /api/v1/namespaces/kube-system/pods/hhh 552\nI0921 16:47:31.209291       1 logs_generator.go:76] 2 GET /api/v1/namespaces/ns/pods/h59d 368\nI0921 16:47:31.409240       1 logs_generator.go:76] 3 PUT /api/v1/namespaces/default/pods/mg6 388\nI0921 16:47:31.609192       1 logs_generator.go:76] 4 POST /api/v1/namespaces/ns/pods/5s49 525\nI0921 16:47:31.809160       1 logs_generator.go:76] 5 GET /api/v1/namespaces/ns/pods/9h7 456\nI0921 16:47:32.009206       1 logs_generator.go:76] 6 GET /api/v1/namespaces/kube-system/pods/sg5l 292\nI0921 16:47:32.209243       1 logs_generator.go:76] 7 POST /api/v1/namespaces/default/pods/knv 478\nI0921 16:47:32.409227       1 logs_generator.go:76] 8 GET /api/v1/namespaces/ns/pods/mcbw 589\nI0921 16:47:32.609210       1 logs_generator.go:76] 9 POST /api/v1/namespaces/default/pods/bz7j 496\nI0921 16:47:32.809182       1 logs_generator.go:76] 10 GET /api/v1/namespaces/kube-system/pods/cqc4 386\nI0921 16:47:33.009186       1 logs_generator.go:76] 11 PUT /api/v1/namespaces/default/pods/sp9p 487\nI0921 16:47:33.209163       1 logs_generator.go:76] 12 PUT /api/v1/namespaces/kube-system/pods/dcr2 435\nI0921 16:47:33.409203       1 logs_generator.go:76] 13 POST /api/v1/namespaces/default/pods/js8 427\nI0921 16:47:33.610316       1 logs_generator.go:76] 14 PUT /api/v1/namespaces/kube-system/pods/kjh 348\nI0921 16:47:33.812297       1 logs_generator.go:76] 15 POST /api/v1/namespaces/kube-system/pods/rhdf 504\nI0921 16:47:34.010625       1 logs_generator.go:76] 16 POST /api/v1/namespaces/ns/pods/wnw 261\nI0921 16:47:34.209436       1 logs_generator.go:76] 17 PUT /api/v1/namespaces/ns/pods/sg2 286\nI0921 16:47:34.409298       1 logs_generator.go:76] 18 GET /api/v1/namespaces/default/pods/qwk 244\nI0921 16:47:34.609321       1 logs_generator.go:76] 19 POST /api/v1/namespaces/default/pods/bcbp 480\nI0921 16:47:34.809237       1 logs_generator.go:76] 20 POST /api/v1/namespaces/kube-system/pods/7c7w 345\nI0921 16:47:35.009236       1 logs_generator.go:76] 21 PUT /api/v1/namespaces/default/pods/zbjk 547\nI0921 16:47:35.209361       1 logs_generator.go:76] 22 POST /api/v1/namespaces/default/pods/7j2 242\nI0921 16:47:35.409319       1 logs_generator.go:76] 23 PUT /api/v1/namespaces/ns/pods/8ft 237\nI0921 16:47:35.609245       1 logs_generator.go:76] 24 POST /api/v1/namespaces/ns/pods/zf4x 237\nI0921 16:47:35.809288       1 logs_generator.go:76] 25 PUT /api/v1/namespaces/kube-system/pods/qgc7 220\nI0921 16:47:36.010176       1 logs_generator.go:76] 26 GET /api/v1/namespaces/kube-system/pods/88n 531\nI0921 16:47:36.209224       1 logs_generator.go:76] 27 GET /api/v1/namespaces/kube-system/pods/xglf 222\nI0921 16:47:36.409355       1 logs_generator.go:76] 28 PUT /api/v1/namespaces/kube-system/pods/sj8n 242\nI0921 16:47:36.609224       1 logs_generator.go:76] 29 GET /api/v1/namespaces/default/pods/qjgs 279\nI0921 16:47:36.809263       1 logs_generator.go:76] 30 POST /api/v1/namespaces/kube-system/pods/n6m 593\nI0921 16:47:37.009215       1 logs_generator.go:76] 31 POST /api/v1/namespaces/ns/pods/wzx 354\n"
-[AfterEach] Kubectl logs
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1470
-Sep 21 16:47:37.149: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete pod logs-generator --namespace=kubectl-4454'
-Sep 21 16:47:39.887: INFO: stderr: ""
-Sep 21 16:47:39.887: INFO: stdout: "pod \"logs-generator\" deleted\n"
-[AfterEach] [sig-cli] Kubectl client
+[AfterEach] [sig-api-machinery] Servers with support for Table transformation
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:39.887: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-4454" for this suite.
-
-• [SLOW TEST:11.003 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Kubectl logs
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1460
-    should be able to retrieve and filter logs  [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs  [Conformance]","total":280,"completed":194,"skipped":3324,"failed":0}
-SSSSSSSSSS
+Oct 26 17:00:15.192: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "tables-8723" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance]","total":280,"completed":196,"skipped":3056,"failed":0}
+SSSSSSSSSSSSS
 ------------------------------
-[k8s.io] InitContainer [NodeConformance] 
-  should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]
+[sig-api-machinery] ResourceQuota 
+  should be able to update and delete ResourceQuota. [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+[BeforeEach] [sig-api-machinery] ResourceQuota
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:47:39.950: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename init-container
+Oct 26 17:00:15.251: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename resourcequota
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
-[It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]
+[It] should be able to update and delete ResourceQuota. [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-Sep 21 16:47:40.207: INFO: PodSpec: initContainers in spec.initContainers
-[AfterEach] [k8s.io] InitContainer [NodeConformance]
+STEP: Creating a ResourceQuota
+STEP: Getting a ResourceQuota
+STEP: Updating a ResourceQuota
+STEP: Verifying a ResourceQuota was modified
+STEP: Deleting a ResourceQuota
+STEP: Verifying the deleted ResourceQuota
+[AfterEach] [sig-api-machinery] ResourceQuota
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:43.876: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "init-container-5791" for this suite.
-•{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]","total":280,"completed":195,"skipped":3334,"failed":0}
-SSSSSSSSSS
+Oct 26 17:00:15.492: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-7747" for this suite.
+•{"msg":"PASSED [sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance]","total":280,"completed":197,"skipped":3069,"failed":0}
+
 ------------------------------
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  patching/updating a mutating webhook should work [Conformance]
+  should include webhook resources in discovery documents [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:47:43.941: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:00:15.537: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
@@ -9837,688 +9010,209 @@ STEP: Setting up server cert
 STEP: Create role binding to let webhook read extension-apiserver-authentication
 STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Sep 21 16:47:45.077: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Sep 21 16:47:47.135: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303665, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303665, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303665, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303665, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:00:16.794: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 17:00:18.862: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328416, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328416, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328416, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328416, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Sep 21 16:47:50.234: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] patching/updating a mutating webhook should work [Conformance]
+Oct 26 17:00:21.919: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should include webhook resources in discovery documents [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a mutating webhook configuration
-STEP: Updating a mutating webhook configuration's rules to not include the create operation
-STEP: Creating a configMap that should not be mutated
-STEP: Patching a mutating webhook configuration's rules to include the create operation
-STEP: Creating a configMap that should be mutated
+STEP: fetching the /apis discovery document
+STEP: finding the admissionregistration.k8s.io API group in the /apis discovery document
+STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis discovery document
+STEP: fetching the /apis/admissionregistration.k8s.io discovery document
+STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis/admissionregistration.k8s.io discovery document
+STEP: fetching the /apis/admissionregistration.k8s.io/v1 discovery document
+STEP: finding mutatingwebhookconfigurations and validatingwebhookconfigurations resources in the /apis/admissionregistration.k8s.io/v1 discovery document
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:50.669: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-6442" for this suite.
-STEP: Destroying namespace "webhook-6442-markers" for this suite.
+Oct 26 17:00:21.957: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-9148" for this suite.
+STEP: Destroying namespace "webhook-9148-markers" for this suite.
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:7.196 seconds]
+• [SLOW TEST:6.700 seconds]
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  patching/updating a mutating webhook should work [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance]","total":280,"completed":196,"skipped":3344,"failed":0}
-SSSSSSSS
-------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:47:51.137: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating projection with secret that has name projected-secret-test-ddb11504-443c-4547-935c-4887caa42bfe
-STEP: Creating a pod to test consume secrets
-Sep 21 16:47:51.566: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-91258b71-0c13-4e98-be9c-3c8193f2486f" in namespace "projected-3651" to be "success or failure"
-Sep 21 16:47:51.774: INFO: Pod "pod-projected-secrets-91258b71-0c13-4e98-be9c-3c8193f2486f": Phase="Pending", Reason="", readiness=false. Elapsed: 208.047716ms
-Sep 21 16:47:53.796: INFO: Pod "pod-projected-secrets-91258b71-0c13-4e98-be9c-3c8193f2486f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.230114975s
-Sep 21 16:47:55.829: INFO: Pod "pod-projected-secrets-91258b71-0c13-4e98-be9c-3c8193f2486f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.263422852s
-STEP: Saw pod success
-Sep 21 16:47:55.829: INFO: Pod "pod-projected-secrets-91258b71-0c13-4e98-be9c-3c8193f2486f" satisfied condition "success or failure"
-Sep 21 16:47:55.853: INFO: Trying to get logs from node 10.241.51.147 pod pod-projected-secrets-91258b71-0c13-4e98-be9c-3c8193f2486f container projected-secret-volume-test: 
-STEP: delete the pod
-Sep 21 16:47:56.003: INFO: Waiting for pod pod-projected-secrets-91258b71-0c13-4e98-be9c-3c8193f2486f to disappear
-Sep 21 16:47:56.028: INFO: Pod pod-projected-secrets-91258b71-0c13-4e98-be9c-3c8193f2486f no longer exists
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:56.029: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-3651" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":197,"skipped":3352,"failed":0}
-SS
-------------------------------
-[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
-  should be possible to delete [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:47:56.098: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubelet-test
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
-[BeforeEach] when scheduling a busybox command that always fails in a pod
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
-[It] should be possible to delete [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:56.500: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubelet-test-6445" for this suite.
-•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance]","total":280,"completed":198,"skipped":3354,"failed":0}
-SSSSSSSSSSSSSSSS
-------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:47:56.604: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename secrets
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-map-e69fb4f6-0341-496c-9493-d6fcc747cb3c
-STEP: Creating a pod to test consume secrets
-Sep 21 16:47:57.023: INFO: Waiting up to 5m0s for pod "pod-secrets-7c9dd106-d8b4-4080-9fc3-ae01ca9a75cf" in namespace "secrets-848" to be "success or failure"
-Sep 21 16:47:57.045: INFO: Pod "pod-secrets-7c9dd106-d8b4-4080-9fc3-ae01ca9a75cf": Phase="Pending", Reason="", readiness=false. Elapsed: 22.096372ms
-Sep 21 16:47:59.069: INFO: Pod "pod-secrets-7c9dd106-d8b4-4080-9fc3-ae01ca9a75cf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.045887178s
-STEP: Saw pod success
-Sep 21 16:47:59.069: INFO: Pod "pod-secrets-7c9dd106-d8b4-4080-9fc3-ae01ca9a75cf" satisfied condition "success or failure"
-Sep 21 16:47:59.091: INFO: Trying to get logs from node 10.241.51.147 pod pod-secrets-7c9dd106-d8b4-4080-9fc3-ae01ca9a75cf container secret-volume-test: 
-STEP: delete the pod
-Sep 21 16:47:59.233: INFO: Waiting for pod pod-secrets-7c9dd106-d8b4-4080-9fc3-ae01ca9a75cf to disappear
-Sep 21 16:47:59.256: INFO: Pod pod-secrets-7c9dd106-d8b4-4080-9fc3-ae01ca9a75cf no longer exists
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:47:59.256: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-848" for this suite.
-•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":199,"skipped":3370,"failed":0}
-SSSSSSSSS
+  should include webhook resources in discovery documents [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance]","total":280,"completed":198,"skipped":3069,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Pods 
+  should contain environment variables for services [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
+[BeforeEach] [k8s.io] Pods
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:47:59.360: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename emptydir
+Oct 26 17:00:22.238: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename pods
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
+[It] should contain environment variables for services [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0666 on node default medium
-Sep 21 16:47:59.735: INFO: Waiting up to 5m0s for pod "pod-6a3bf9fc-63ee-40fe-b77d-d8b66c6951c1" in namespace "emptydir-7057" to be "success or failure"
-Sep 21 16:47:59.763: INFO: Pod "pod-6a3bf9fc-63ee-40fe-b77d-d8b66c6951c1": Phase="Pending", Reason="", readiness=false. Elapsed: 27.307042ms
-Sep 21 16:48:01.790: INFO: Pod "pod-6a3bf9fc-63ee-40fe-b77d-d8b66c6951c1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.055084103s
-Sep 21 16:48:03.813: INFO: Pod "pod-6a3bf9fc-63ee-40fe-b77d-d8b66c6951c1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.077570136s
+Oct 26 17:00:24.653: INFO: Waiting up to 5m0s for pod "client-envvars-02bea53b-a9be-4b30-8ca2-9d4571b7c097" in namespace "pods-386" to be "success or failure"
+Oct 26 17:00:24.669: INFO: Pod "client-envvars-02bea53b-a9be-4b30-8ca2-9d4571b7c097": Phase="Pending", Reason="", readiness=false. Elapsed: 16.193433ms
+Oct 26 17:00:26.685: INFO: Pod "client-envvars-02bea53b-a9be-4b30-8ca2-9d4571b7c097": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.032588978s
 STEP: Saw pod success
-Sep 21 16:48:03.813: INFO: Pod "pod-6a3bf9fc-63ee-40fe-b77d-d8b66c6951c1" satisfied condition "success or failure"
-Sep 21 16:48:03.836: INFO: Trying to get logs from node 10.241.51.147 pod pod-6a3bf9fc-63ee-40fe-b77d-d8b66c6951c1 container test-container: 
+Oct 26 17:00:26.685: INFO: Pod "client-envvars-02bea53b-a9be-4b30-8ca2-9d4571b7c097" satisfied condition "success or failure"
+Oct 26 17:00:26.707: INFO: Trying to get logs from node 10.72.119.74 pod client-envvars-02bea53b-a9be-4b30-8ca2-9d4571b7c097 container env3cont: 
 STEP: delete the pod
-Sep 21 16:48:03.995: INFO: Waiting for pod pod-6a3bf9fc-63ee-40fe-b77d-d8b66c6951c1 to disappear
-Sep 21 16:48:04.021: INFO: Pod pod-6a3bf9fc-63ee-40fe-b77d-d8b66c6951c1 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
+Oct 26 17:00:26.852: INFO: Waiting for pod client-envvars-02bea53b-a9be-4b30-8ca2-9d4571b7c097 to disappear
+Oct 26 17:00:26.868: INFO: Pod client-envvars-02bea53b-a9be-4b30-8ca2-9d4571b7c097 no longer exists
+[AfterEach] [k8s.io] Pods
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:48:04.022: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-7057" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":200,"skipped":3379,"failed":0}
-SSSSSSSSSSSS
+Oct 26 17:00:26.868: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-386" for this suite.
+•{"msg":"PASSED [k8s.io] Pods should contain environment variables for services [NodeConformance] [Conformance]","total":280,"completed":199,"skipped":3091,"failed":0}
+SSSSSSSSSSSSSSS
 ------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates resource limits of pods that are allowed to run  [Conformance]
+[k8s.io] [sig-node] PreStop 
+  should call prestop when killing a pod  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+[BeforeEach] [k8s.io] [sig-node] PreStop
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:48:04.099: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename sched-pred
+Oct 26 17:00:26.908: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename prestop
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
-Sep 21 16:48:04.459: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-Sep 21 16:48:04.560: INFO: Waiting for terminating namespaces to be deleted...
-Sep 21 16:48:04.589: INFO: 
-Logging pods the kubelet thinks is on node 10.241.51.147 before test
-Sep 21 16:48:04.725: INFO: openshift-kube-proxy-tvbvw from openshift-kube-proxy started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.725: INFO: 	Container kube-proxy ready: true, restart count 0
-Sep 21 16:48:04.725: INFO: multus-admission-controller-7rkpv from openshift-multus started at 2020-09-21 16:14:37 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.725: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:04.725: INFO: 	Container multus-admission-controller ready: true, restart count 0
-Sep 21 16:48:04.725: INFO: multus-7xt8r from openshift-multus started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.725: INFO: 	Container kube-multus ready: true, restart count 0
-Sep 21 16:48:04.725: INFO: ibm-master-proxy-static-10.241.51.147 from kube-system started at 2020-09-21 14:11:47 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.725: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
-Sep 21 16:48:04.725: INFO: 	Container pause ready: true, restart count 0
-Sep 21 16:48:04.725: INFO: sonobuoy-e2e-job-36e1517c33ed4f81 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.725: INFO: 	Container e2e ready: true, restart count 0
-Sep 21 16:48:04.725: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Sep 21 16:48:04.725: INFO: sonobuoy from sonobuoy started at 2020-09-21 15:39:44 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.725: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-Sep 21 16:48:04.725: INFO: node-exporter-46mwx from openshift-monitoring started at 2020-09-21 14:11:49 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.726: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:04.726: INFO: 	Container node-exporter ready: true, restart count 0
-Sep 21 16:48:04.726: INFO: node-ca-6r9nd from openshift-image-registry started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.726: INFO: 	Container node-ca ready: true, restart count 0
-Sep 21 16:48:04.726: INFO: dns-default-f6vtl from openshift-dns started at 2020-09-21 14:11:49 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.726: INFO: 	Container dns ready: true, restart count 0
-Sep 21 16:48:04.726: INFO: 	Container dns-node-resolver ready: true, restart count 0
-Sep 21 16:48:04.726: INFO: tuned-6qvw2 from openshift-cluster-node-tuning-operator started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.726: INFO: 	Container tuned ready: true, restart count 0
-Sep 21 16:48:04.726: INFO: calico-typha-6fc9ff8b5f-9pmb4 from calico-system started at 2020-09-21 14:13:14 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.726: INFO: 	Container calico-typha ready: true, restart count 0
-Sep 21 16:48:04.726: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-shgq6 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.726: INFO: 	Container sonobuoy-worker ready: true, restart count 1
-Sep 21 16:48:04.726: INFO: 	Container systemd-logs ready: true, restart count 0
-Sep 21 16:48:04.726: INFO: packageserver-75876789fb-w6q6s from openshift-operator-lifecycle-manager started at 2020-09-21 16:30:37 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.726: INFO: 	Container packageserver ready: true, restart count 0
-Sep 21 16:48:04.726: INFO: ibm-keepalived-watcher-h4dbn from kube-system started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.726: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Sep 21 16:48:04.726: INFO: calico-node-t2n87 from calico-system started at 2020-09-21 14:11:49 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.726: INFO: 	Container calico-node ready: true, restart count 0
-Sep 21 16:48:04.727: INFO: ibmcloud-block-storage-driver-8r6lc from kube-system started at 2020-09-21 14:12:04 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.727: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
-Sep 21 16:48:04.727: INFO: 
-Logging pods the kubelet thinks is on node 10.241.51.149 before test
-Sep 21 16:48:04.965: INFO: openshift-kube-proxy-9svk2 from openshift-kube-proxy started at 2020-09-21 14:06:38 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.965: INFO: 	Container kube-proxy ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: service-ca-operator-9f84dfdb5-wdhvg from openshift-service-ca-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container operator ready: true, restart count 1
-Sep 21 16:48:04.966: INFO: ingress-operator-57546dfcd9-2sg5v from openshift-ingress-operator started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container ingress-operator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: catalog-operator-69d4f5c59f-lqjgz from openshift-operator-lifecycle-manager started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container catalog-operator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-lgbbw from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container sonobuoy-worker ready: true, restart count 1
-Sep 21 16:48:04.966: INFO: 	Container systemd-logs ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: router-default-84dfcdc9b5-tk98j from openshift-ingress started at 2020-09-21 14:09:06 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container router ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: downloads-7989988989-9rhh4 from openshift-console started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container download-server ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: tuned-gpfhj from openshift-cluster-node-tuning-operator started at 2020-09-21 14:08:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container tuned ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: node-exporter-jnrzm from openshift-monitoring started at 2020-09-21 14:08:24 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container node-exporter ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: dns-default-958xc from openshift-dns started at 2020-09-21 14:09:05 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container dns ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container dns-node-resolver ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: alertmanager-main-0 from openshift-monitoring started at 2020-09-21 14:14:12 +0000 UTC (3 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container alertmanager ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container alertmanager-proxy ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container config-reloader ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: ibm-keepalived-watcher-r4tn7 from kube-system started at 2020-09-21 14:06:04 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: marketplace-operator-56c6694c8f-x9rcq from openshift-marketplace started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container marketplace-operator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: downloads-7989988989-cjvjw from openshift-console started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container download-server ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-44pnn from ibm-system started at 2020-09-21 14:12:56 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container ibm-cloud-provider-ip-169-48-211-85 ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: tigera-operator-679798d94d-hdlvj from tigera-operator started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container tigera-operator ready: true, restart count 2
-Sep 21 16:48:04.966: INFO: cluster-image-registry-operator-695bf78ffc-95l5m from openshift-image-registry started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container cluster-image-registry-operator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container cluster-image-registry-operator-watch ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: console-operator-58fd84b95c-dc6vq from openshift-console-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container console-operator ready: true, restart count 1
-Sep 21 16:48:04.966: INFO: migrator-6c94c8c775-vxsnp from openshift-kube-storage-version-migrator started at 2020-09-21 14:07:55 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container migrator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: cluster-storage-operator-5bcf8d9f8f-dls4n from openshift-cluster-storage-operator started at 2020-09-21 14:07:36 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container cluster-storage-operator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: image-registry-775dd777d7-5qtdd from openshift-image-registry started at 2020-09-21 14:11:01 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container registry ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: ibm-master-proxy-static-10.241.51.149 from kube-system started at 2020-09-21 14:06:02 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container pause ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: cluster-node-tuning-operator-bdbf98554-22kv8 from openshift-cluster-node-tuning-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container cluster-node-tuning-operator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: redhat-operators-54dcdd9697-pcwgz from openshift-marketplace started at 2020-09-21 14:09:34 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container redhat-operators ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: registry-pvc-permissions-rlfxm from openshift-image-registry started at 2020-09-21 14:11:01 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container pvc-permissions ready: false, restart count 0
-Sep 21 16:48:04.966: INFO: calico-node-9gqht from calico-system started at 2020-09-21 14:07:18 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container calico-node ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: openshift-service-catalog-controller-manager-operator-b9ccczlcs from openshift-service-catalog-controller-manager-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container operator ready: true, restart count 1
-Sep 21 16:48:04.966: INFO: olm-operator-b5f47bf98-8d8cw from openshift-operator-lifecycle-manager started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container olm-operator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: console-8549dd6c4f-8vk99 from openshift-console started at 2020-09-21 16:14:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container console ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: node-ca-bxmsf from openshift-image-registry started at 2020-09-21 14:09:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container node-ca ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: prometheus-operator-8d96bcc68-ct2jw from openshift-monitoring started at 2020-09-21 16:14:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container prometheus-operator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: ibm-file-plugin-66fdcfc4c9-tm6qk from kube-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: cluster-samples-operator-5cfb985bf-n7nxd from openshift-cluster-samples-operator started at 2020-09-21 14:09:54 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container cluster-samples-operator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container cluster-samples-operator-watch ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: openshift-service-catalog-apiserver-operator-6988647bc8-t9z7w from openshift-service-catalog-apiserver-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container operator ready: true, restart count 1
-Sep 21 16:48:04.966: INFO: ibmcloud-block-storage-plugin-68d5c65db9-7q8bv from kube-system started at 2020-09-21 14:07:36 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container ibmcloud-block-storage-plugin-container ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: openshift-state-metrics-678b6c786f-698gn from openshift-monitoring started at 2020-09-21 14:08:22 +0000 UTC (3 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container openshift-state-metrics ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: redhat-marketplace-7b9bcdf96b-p4khw from openshift-marketplace started at 2020-09-21 14:09:34 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container redhat-marketplace ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: multus-hprgb from openshift-multus started at 2020-09-21 14:06:32 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container kube-multus ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: calico-typha-6fc9ff8b5f-gzhgw from calico-system started at 2020-09-21 14:07:18 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container calico-typha ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: kube-state-metrics-7455b6c8-455w7 from openshift-monitoring started at 2020-09-21 14:08:22 +0000 UTC (3 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container kube-state-metrics ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: prometheus-adapter-5f78ddc679-vb486 from openshift-monitoring started at 2020-09-21 14:08:28 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container prometheus-adapter ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: network-operator-f456cfdbb-4m6pj from openshift-network-operator started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container network-operator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: cluster-monitoring-operator-75f4ddd85f-579q6 from openshift-monitoring started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container cluster-monitoring-operator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: kube-storage-version-migrator-operator-86d8fc6986-vl4fl from openshift-kube-storage-version-migrator-operator started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container kube-storage-version-migrator-operator ready: true, restart count 1
-Sep 21 16:48:04.966: INFO: calico-kube-controllers-79d75767dd-xkvb6 from calico-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container calico-kube-controllers ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: multus-admission-controller-gkchr from openshift-multus started at 2020-09-21 14:07:34 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container multus-admission-controller ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: prometheus-adapter-5f78ddc679-9hj7w from openshift-monitoring started at 2020-09-21 14:08:28 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container prometheus-adapter ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: ibm-storage-watcher-7956f954f-26xx6 from kube-system started at 2020-09-21 14:07:33 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: service-ca-bd8866b9b-qzggv from openshift-service-ca started at 2020-09-21 14:07:50 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container service-ca-controller ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: certified-operators-5cfd68758d-q6c6n from openshift-marketplace started at 2020-09-21 16:14:05 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container certified-operators ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: community-operators-6566687cd-smnd2 from openshift-marketplace started at 2020-09-21 14:09:35 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container community-operators ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: ibmcloud-block-storage-driver-pldbm from kube-system started at 2020-09-21 14:06:08 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: dns-operator-7886f4f4f4-rk7dv from openshift-dns-operator started at 2020-09-21 14:07:33 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container dns-operator ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: prometheus-k8s-1 from openshift-monitoring started at 2020-09-21 16:14:17 +0000 UTC (7 container statuses recorded)
-Sep 21 16:48:04.966: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container prometheus ready: true, restart count 1
-Sep 21 16:48:04.966: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container prometheus-proxy ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 	Container thanos-sidecar ready: true, restart count 0
-Sep 21 16:48:04.966: INFO: 
-Logging pods the kubelet thinks is on node 10.241.51.150 before test
-Sep 21 16:48:05.200: INFO: console-8549dd6c4f-b5jnn from openshift-console started at 2020-09-21 14:18:13 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container console ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: node-exporter-flb5l from openshift-monitoring started at 2020-09-21 14:12:21 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container node-exporter ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: calico-node-dn4jv from calico-system started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container calico-node ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: grafana-c745ff78c-9pgfx from openshift-monitoring started at 2020-09-21 14:13:42 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container grafana ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container grafana-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-ghw5q from ibm-system started at 2020-09-21 16:07:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container ibm-cloud-provider-ip-169-48-211-85 ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: dns-default-z8v9j from openshift-dns started at 2020-09-21 14:12:21 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container dns ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container dns-node-resolver ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: calico-typha-6fc9ff8b5f-cbhcz from calico-system started at 2020-09-21 14:13:14 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container calico-typha ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: multus-admission-controller-snglx from openshift-multus started at 2020-09-21 14:13:31 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container multus-admission-controller ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: alertmanager-main-1 from openshift-monitoring started at 2020-09-21 14:14:02 +0000 UTC (3 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container alertmanager ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container alertmanager-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container config-reloader ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: router-default-84dfcdc9b5-mt4b7 from openshift-ingress started at 2020-09-21 16:07:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container router ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: multus-n5qc8 from openshift-multus started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container kube-multus ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: openshift-kube-proxy-9bv2v from openshift-kube-proxy started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container kube-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: prometheus-k8s-0 from openshift-monitoring started at 2020-09-21 14:15:10 +0000 UTC (7 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container prometheus ready: true, restart count 1
-Sep 21 16:48:05.200: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container prometheus-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container thanos-sidecar ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: ibm-master-proxy-static-10.241.51.150 from kube-system started at 2020-09-21 14:12:19 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container pause ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: telemeter-client-68bdb7f795-nnhfl from openshift-monitoring started at 2020-09-21 14:13:34 +0000 UTC (3 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container reload ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container telemeter-client ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: vpn-64d8d4987c-lxt95 from kube-system started at 2020-09-21 16:07:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container vpn ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: thanos-querier-8f4c5c746-xv2gj from openshift-monitoring started at 2020-09-21 16:14:05 +0000 UTC (4 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container oauth-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container thanos-querier ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: ibmcloud-block-storage-driver-jf9s4 from kube-system started at 2020-09-21 14:12:26 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: thanos-querier-8f4c5c746-v982g from openshift-monitoring started at 2020-09-21 14:14:42 +0000 UTC (4 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container oauth-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container prom-label-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container thanos-querier ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: packageserver-75876789fb-h2vxx from openshift-operator-lifecycle-manager started at 2020-09-21 16:30:42 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container packageserver ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: node-ca-ql97q from openshift-image-registry started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container node-ca ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: tuned-5n2xm from openshift-cluster-node-tuning-operator started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container tuned ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-xzz27 from sonobuoy started at 2020-09-21 15:39:52 +0000 UTC (2 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container sonobuoy-worker ready: true, restart count 1
-Sep 21 16:48:05.200: INFO: 	Container systemd-logs ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: ibm-keepalived-watcher-bntgc from kube-system started at 2020-09-21 14:12:21 +0000 UTC (1 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container keepalived-watcher ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: alertmanager-main-2 from openshift-monitoring started at 2020-09-21 16:14:17 +0000 UTC (3 container statuses recorded)
-Sep 21 16:48:05.200: INFO: 	Container alertmanager ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container alertmanager-proxy ready: true, restart count 0
-Sep 21 16:48:05.200: INFO: 	Container config-reloader ready: true, restart count 0
-[It] validates resource limits of pods that are allowed to run  [Conformance]
+[BeforeEach] [k8s.io] [sig-node] PreStop
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pre_stop.go:172
+[It] should call prestop when killing a pod  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: verifying the node has the label node 10.241.51.147
-STEP: verifying the node has the label node 10.241.51.149
-STEP: verifying the node has the label node 10.241.51.150
-Sep 21 16:48:05.872: INFO: Pod calico-kube-controllers-79d75767dd-xkvb6 requesting resource cpu=0m on Node 10.241.51.149
-Sep 21 16:48:05.872: INFO: Pod calico-node-9gqht requesting resource cpu=0m on Node 10.241.51.149
-Sep 21 16:48:05.872: INFO: Pod calico-node-dn4jv requesting resource cpu=0m on Node 10.241.51.150
-Sep 21 16:48:05.872: INFO: Pod calico-node-t2n87 requesting resource cpu=0m on Node 10.241.51.147
-Sep 21 16:48:05.872: INFO: Pod calico-typha-6fc9ff8b5f-9pmb4 requesting resource cpu=0m on Node 10.241.51.147
-Sep 21 16:48:05.872: INFO: Pod calico-typha-6fc9ff8b5f-cbhcz requesting resource cpu=0m on Node 10.241.51.150
-Sep 21 16:48:05.872: INFO: Pod calico-typha-6fc9ff8b5f-gzhgw requesting resource cpu=0m on Node 10.241.51.149
-Sep 21 16:48:05.872: INFO: Pod ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-44pnn requesting resource cpu=5m on Node 10.241.51.149
-Sep 21 16:48:05.872: INFO: Pod ibm-cloud-provider-ip-169-48-211-85-7f7cb67554-ghw5q requesting resource cpu=5m on Node 10.241.51.150
-Sep 21 16:48:05.872: INFO: Pod ibm-file-plugin-66fdcfc4c9-tm6qk requesting resource cpu=50m on Node 10.241.51.149
-Sep 21 16:48:05.872: INFO: Pod ibm-keepalived-watcher-bntgc requesting resource cpu=5m on Node 10.241.51.150
-Sep 21 16:48:05.872: INFO: Pod ibm-keepalived-watcher-h4dbn requesting resource cpu=5m on Node 10.241.51.147
-Sep 21 16:48:05.872: INFO: Pod ibm-keepalived-watcher-r4tn7 requesting resource cpu=5m on Node 10.241.51.149
-Sep 21 16:48:05.872: INFO: Pod ibm-master-proxy-static-10.241.51.147 requesting resource cpu=25m on Node 10.241.51.147
-Sep 21 16:48:05.873: INFO: Pod ibm-master-proxy-static-10.241.51.149 requesting resource cpu=25m on Node 10.241.51.149
-Sep 21 16:48:05.873: INFO: Pod ibm-master-proxy-static-10.241.51.150 requesting resource cpu=25m on Node 10.241.51.150
-Sep 21 16:48:05.873: INFO: Pod ibm-storage-watcher-7956f954f-26xx6 requesting resource cpu=50m on Node 10.241.51.149
-Sep 21 16:48:05.873: INFO: Pod ibmcloud-block-storage-driver-8r6lc requesting resource cpu=50m on Node 10.241.51.147
-Sep 21 16:48:05.873: INFO: Pod ibmcloud-block-storage-driver-jf9s4 requesting resource cpu=50m on Node 10.241.51.150
-Sep 21 16:48:05.873: INFO: Pod ibmcloud-block-storage-driver-pldbm requesting resource cpu=50m on Node 10.241.51.149
-Sep 21 16:48:05.873: INFO: Pod ibmcloud-block-storage-plugin-68d5c65db9-7q8bv requesting resource cpu=50m on Node 10.241.51.149
-Sep 21 16:48:05.873: INFO: Pod vpn-64d8d4987c-lxt95 requesting resource cpu=5m on Node 10.241.51.150
-Sep 21 16:48:05.873: INFO: Pod cluster-node-tuning-operator-bdbf98554-22kv8 requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.873: INFO: Pod tuned-5n2xm requesting resource cpu=10m on Node 10.241.51.150
-Sep 21 16:48:05.873: INFO: Pod tuned-6qvw2 requesting resource cpu=10m on Node 10.241.51.147
-Sep 21 16:48:05.873: INFO: Pod tuned-gpfhj requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.873: INFO: Pod cluster-samples-operator-5cfb985bf-n7nxd requesting resource cpu=20m on Node 10.241.51.149
-Sep 21 16:48:05.873: INFO: Pod cluster-storage-operator-5bcf8d9f8f-dls4n requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.873: INFO: Pod console-operator-58fd84b95c-dc6vq requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.873: INFO: Pod console-8549dd6c4f-8vk99 requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.873: INFO: Pod console-8549dd6c4f-b5jnn requesting resource cpu=10m on Node 10.241.51.150
-Sep 21 16:48:05.873: INFO: Pod downloads-7989988989-9rhh4 requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.873: INFO: Pod downloads-7989988989-cjvjw requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.874: INFO: Pod dns-operator-7886f4f4f4-rk7dv requesting resource cpu=20m on Node 10.241.51.149
-Sep 21 16:48:05.874: INFO: Pod dns-default-958xc requesting resource cpu=110m on Node 10.241.51.149
-Sep 21 16:48:05.874: INFO: Pod dns-default-f6vtl requesting resource cpu=110m on Node 10.241.51.147
-Sep 21 16:48:05.874: INFO: Pod dns-default-z8v9j requesting resource cpu=110m on Node 10.241.51.150
-Sep 21 16:48:05.874: INFO: Pod cluster-image-registry-operator-695bf78ffc-95l5m requesting resource cpu=20m on Node 10.241.51.149
-Sep 21 16:48:05.874: INFO: Pod image-registry-775dd777d7-5qtdd requesting resource cpu=100m on Node 10.241.51.149
-Sep 21 16:48:05.874: INFO: Pod node-ca-6r9nd requesting resource cpu=10m on Node 10.241.51.147
-Sep 21 16:48:05.874: INFO: Pod node-ca-bxmsf requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.874: INFO: Pod node-ca-ql97q requesting resource cpu=10m on Node 10.241.51.150
-Sep 21 16:48:05.874: INFO: Pod ingress-operator-57546dfcd9-2sg5v requesting resource cpu=20m on Node 10.241.51.149
-Sep 21 16:48:05.874: INFO: Pod router-default-84dfcdc9b5-mt4b7 requesting resource cpu=100m on Node 10.241.51.150
-Sep 21 16:48:05.874: INFO: Pod router-default-84dfcdc9b5-tk98j requesting resource cpu=100m on Node 10.241.51.149
-Sep 21 16:48:05.874: INFO: Pod openshift-kube-proxy-9bv2v requesting resource cpu=100m on Node 10.241.51.150
-Sep 21 16:48:05.874: INFO: Pod openshift-kube-proxy-9svk2 requesting resource cpu=100m on Node 10.241.51.149
-Sep 21 16:48:05.874: INFO: Pod openshift-kube-proxy-tvbvw requesting resource cpu=100m on Node 10.241.51.147
-Sep 21 16:48:05.874: INFO: Pod kube-storage-version-migrator-operator-86d8fc6986-vl4fl requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod migrator-6c94c8c775-vxsnp requesting resource cpu=100m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod certified-operators-5cfd68758d-q6c6n requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod community-operators-6566687cd-smnd2 requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod marketplace-operator-56c6694c8f-x9rcq requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod redhat-marketplace-7b9bcdf96b-p4khw requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod redhat-operators-54dcdd9697-pcwgz requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod alertmanager-main-0 requesting resource cpu=6m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod alertmanager-main-1 requesting resource cpu=6m on Node 10.241.51.150
-Sep 21 16:48:05.875: INFO: Pod alertmanager-main-2 requesting resource cpu=6m on Node 10.241.51.150
-Sep 21 16:48:05.875: INFO: Pod cluster-monitoring-operator-75f4ddd85f-579q6 requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod grafana-c745ff78c-9pgfx requesting resource cpu=5m on Node 10.241.51.150
-Sep 21 16:48:05.875: INFO: Pod kube-state-metrics-7455b6c8-455w7 requesting resource cpu=4m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod node-exporter-46mwx requesting resource cpu=9m on Node 10.241.51.147
-Sep 21 16:48:05.875: INFO: Pod node-exporter-flb5l requesting resource cpu=9m on Node 10.241.51.150
-Sep 21 16:48:05.875: INFO: Pod node-exporter-jnrzm requesting resource cpu=9m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod openshift-state-metrics-678b6c786f-698gn requesting resource cpu=3m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod prometheus-adapter-5f78ddc679-9hj7w requesting resource cpu=1m on Node 10.241.51.149
-Sep 21 16:48:05.875: INFO: Pod prometheus-adapter-5f78ddc679-vb486 requesting resource cpu=1m on Node 10.241.51.149
-Sep 21 16:48:05.876: INFO: Pod prometheus-k8s-0 requesting resource cpu=76m on Node 10.241.51.150
-Sep 21 16:48:05.876: INFO: Pod prometheus-k8s-1 requesting resource cpu=76m on Node 10.241.51.149
-Sep 21 16:48:05.876: INFO: Pod prometheus-operator-8d96bcc68-ct2jw requesting resource cpu=5m on Node 10.241.51.149
-Sep 21 16:48:05.876: INFO: Pod telemeter-client-68bdb7f795-nnhfl requesting resource cpu=3m on Node 10.241.51.150
-Sep 21 16:48:05.876: INFO: Pod thanos-querier-8f4c5c746-v982g requesting resource cpu=8m on Node 10.241.51.150
-Sep 21 16:48:05.876: INFO: Pod thanos-querier-8f4c5c746-xv2gj requesting resource cpu=8m on Node 10.241.51.150
-Sep 21 16:48:05.876: INFO: Pod multus-7xt8r requesting resource cpu=10m on Node 10.241.51.147
-Sep 21 16:48:05.876: INFO: Pod multus-admission-controller-7rkpv requesting resource cpu=20m on Node 10.241.51.147
-Sep 21 16:48:05.876: INFO: Pod multus-admission-controller-gkchr requesting resource cpu=20m on Node 10.241.51.149
-Sep 21 16:48:05.876: INFO: Pod multus-admission-controller-snglx requesting resource cpu=20m on Node 10.241.51.150
-Sep 21 16:48:05.876: INFO: Pod multus-hprgb requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.876: INFO: Pod multus-n5qc8 requesting resource cpu=10m on Node 10.241.51.150
-Sep 21 16:48:05.876: INFO: Pod network-operator-f456cfdbb-4m6pj requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.876: INFO: Pod catalog-operator-69d4f5c59f-lqjgz requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.876: INFO: Pod olm-operator-b5f47bf98-8d8cw requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.876: INFO: Pod packageserver-75876789fb-h2vxx requesting resource cpu=10m on Node 10.241.51.150
-Sep 21 16:48:05.876: INFO: Pod packageserver-75876789fb-w6q6s requesting resource cpu=10m on Node 10.241.51.147
-Sep 21 16:48:05.876: INFO: Pod service-ca-operator-9f84dfdb5-wdhvg requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.877: INFO: Pod service-ca-bd8866b9b-qzggv requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.877: INFO: Pod openshift-service-catalog-apiserver-operator-6988647bc8-t9z7w requesting resource cpu=0m on Node 10.241.51.149
-Sep 21 16:48:05.877: INFO: Pod openshift-service-catalog-controller-manager-operator-b9ccczlcs requesting resource cpu=10m on Node 10.241.51.149
-Sep 21 16:48:05.877: INFO: Pod sonobuoy requesting resource cpu=0m on Node 10.241.51.147
-Sep 21 16:48:05.877: INFO: Pod sonobuoy-e2e-job-36e1517c33ed4f81 requesting resource cpu=0m on Node 10.241.51.147
-Sep 21 16:48:05.877: INFO: Pod sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-lgbbw requesting resource cpu=0m on Node 10.241.51.149
-Sep 21 16:48:05.877: INFO: Pod sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-shgq6 requesting resource cpu=0m on Node 10.241.51.147
-Sep 21 16:48:05.877: INFO: Pod sonobuoy-systemd-logs-daemon-set-0933bd4da07a4275-xzz27 requesting resource cpu=0m on Node 10.241.51.150
-Sep 21 16:48:05.877: INFO: Pod tigera-operator-679798d94d-hdlvj requesting resource cpu=100m on Node 10.241.51.149
-STEP: Starting Pods to consume most of the cluster CPU.
-Sep 21 16:48:05.877: INFO: Creating a pod which consumes cpu=1848m on Node 10.241.51.149
-Sep 21 16:48:06.007: INFO: Creating a pod which consumes cpu=2323m on Node 10.241.51.150
-Sep 21 16:48:06.134: INFO: Creating a pod which consumes cpu=2485m on Node 10.241.51.147
-STEP: Creating another pod that requires unavailable amount of CPU.
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-00adc854-0dc9-4c52-87f1-fa17f9b3605f.1636da6e9aa2a3c6], Reason = [Scheduled], Message = [Successfully assigned sched-pred-7631/filler-pod-00adc854-0dc9-4c52-87f1-fa17f9b3605f to 10.241.51.150]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-00adc854-0dc9-4c52-87f1-fa17f9b3605f.1636da6edfcbb548], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-00adc854-0dc9-4c52-87f1-fa17f9b3605f.1636da6eedae89af], Reason = [Created], Message = [Created container filler-pod-00adc854-0dc9-4c52-87f1-fa17f9b3605f]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-00adc854-0dc9-4c52-87f1-fa17f9b3605f.1636da6ef0839ad3], Reason = [Started], Message = [Started container filler-pod-00adc854-0dc9-4c52-87f1-fa17f9b3605f]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-469232f0-d9b4-4df7-a3db-83f161bed072.1636da6ea0fcdac3], Reason = [Scheduled], Message = [Successfully assigned sched-pred-7631/filler-pod-469232f0-d9b4-4df7-a3db-83f161bed072 to 10.241.51.147]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-469232f0-d9b4-4df7-a3db-83f161bed072.1636da6eebd3b446], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-469232f0-d9b4-4df7-a3db-83f161bed072.1636da6ef7b3500e], Reason = [Created], Message = [Created container filler-pod-469232f0-d9b4-4df7-a3db-83f161bed072]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-469232f0-d9b4-4df7-a3db-83f161bed072.1636da6ef9b2eea2], Reason = [Started], Message = [Started container filler-pod-469232f0-d9b4-4df7-a3db-83f161bed072]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-822800e6-cd50-4aec-acb9-257065979cfa.1636da6e92bd76f2], Reason = [Scheduled], Message = [Successfully assigned sched-pred-7631/filler-pod-822800e6-cd50-4aec-acb9-257065979cfa to 10.241.51.149]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-822800e6-cd50-4aec-acb9-257065979cfa.1636da6ed9bf98bc], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-822800e6-cd50-4aec-acb9-257065979cfa.1636da6ee5ce3735], Reason = [Created], Message = [Created container filler-pod-822800e6-cd50-4aec-acb9-257065979cfa]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-822800e6-cd50-4aec-acb9-257065979cfa.1636da6ee8383a12], Reason = [Started], Message = [Started container filler-pod-822800e6-cd50-4aec-acb9-257065979cfa]
-STEP: Considering event: 
-Type = [Warning], Name = [additional-pod.1636da6f2691a5b1], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 Insufficient cpu.]
-STEP: removing the label node off the node 10.241.51.147
-STEP: verifying the node doesn't have the label node
-STEP: removing the label node off the node 10.241.51.149
-STEP: verifying the node doesn't have the label node
-STEP: removing the label node off the node 10.241.51.150
-STEP: verifying the node doesn't have the label node
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+STEP: Creating server pod server in namespace prestop-205
+STEP: Waiting for pods to come up.
+STEP: Creating tester pod tester in namespace prestop-205
+STEP: Deleting pre-stop pod
+Oct 26 17:00:40.314: INFO: Saw: {
+	"Hostname": "server",
+	"Sent": null,
+	"Received": {
+		"prestop": 1
+	},
+	"Errors": null,
+	"Log": [
+		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.",
+		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.",
+		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up."
+	],
+	"StillContactingPeers": true
+}
+STEP: Deleting the server pod
+[AfterEach] [k8s.io] [sig-node] PreStop
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:48:09.893: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "sched-pred-7631" for this suite.
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
+Oct 26 17:00:40.349: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "prestop-205" for this suite.
 
-• [SLOW TEST:5.856 seconds]
-[sig-scheduling] SchedulerPredicates [Serial]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
-  validates resource limits of pods that are allowed to run  [Conformance]
+• [SLOW TEST:13.483 seconds]
+[k8s.io] [sig-node] PreStop
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should call prestop when killing a pod  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run  [Conformance]","total":280,"completed":201,"skipped":3391,"failed":0}
-SSSSSSSSS
+{"msg":"PASSED [k8s.io] [sig-node] PreStop should call prestop when killing a pod  [Conformance]","total":280,"completed":200,"skipped":3106,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected configMap 
-  updates should be reflected in volume [NodeConformance] [Conformance]
+[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition 
+  getting/updating/patching custom resource definition status sub-resource works  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:48:09.957: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 17:00:40.398: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename custom-resource-definition
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] updates should be reflected in volume [NodeConformance] [Conformance]
+[It] getting/updating/patching custom resource definition status sub-resource works  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:48:10.266: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
-STEP: Creating projection with configMap that has name projected-configmap-test-upd-4269238a-f64a-4efc-82eb-020237ecfed0
-STEP: Creating the pod
-STEP: Updating configmap projected-configmap-test-upd-4269238a-f64a-4efc-82eb-020237ecfed0
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] Projected configMap
+Oct 26 17:00:40.527: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:48:17.058: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-9282" for this suite.
-
-• [SLOW TEST:7.163 seconds]
-[sig-storage] Projected configMap
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
-  updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":202,"skipped":3400,"failed":0}
-SSSSSS
+Oct 26 17:00:41.176: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "custom-resource-definition-6871" for this suite.
+•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works  [Conformance]","total":280,"completed":201,"skipped":3139,"failed":0}
+SS
 ------------------------------
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  works for CRD preserving unknown fields in an embedded object [Conformance]
+[sig-network] Networking Granular Checks: Pods 
+  should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+[BeforeEach] [sig-network] Networking
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:48:17.122: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename crd-publish-openapi
+Oct 26 17:00:41.215: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename pod-network-test
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] works for CRD preserving unknown fields in an embedded object [Conformance]
+[It] should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:48:17.352: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: client-side validation (kubectl create and apply) allows request with any unknown properties
-Sep 21 16:48:25.847: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-5100 create -f -'
-Sep 21 16:48:26.701: INFO: stderr: ""
-Sep 21 16:48:26.701: INFO: stdout: "e2e-test-crd-publish-openapi-9356-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n"
-Sep 21 16:48:26.701: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-5100 delete e2e-test-crd-publish-openapi-9356-crds test-cr'
-Sep 21 16:48:26.932: INFO: stderr: ""
-Sep 21 16:48:26.932: INFO: stdout: "e2e-test-crd-publish-openapi-9356-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n"
-Sep 21 16:48:26.932: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-5100 apply -f -'
-Sep 21 16:48:27.319: INFO: stderr: ""
-Sep 21 16:48:27.319: INFO: stdout: "e2e-test-crd-publish-openapi-9356-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n"
-Sep 21 16:48:27.319: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-5100 delete e2e-test-crd-publish-openapi-9356-crds test-cr'
-Sep 21 16:48:27.783: INFO: stderr: ""
-Sep 21 16:48:27.783: INFO: stdout: "e2e-test-crd-publish-openapi-9356-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n"
-STEP: kubectl explain works to explain CR
-Sep 21 16:48:27.783: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 explain e2e-test-crd-publish-openapi-9356-crds'
-Sep 21 16:48:28.256: INFO: stderr: ""
-Sep 21 16:48:28.256: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-9356-crd\nVERSION:  crd-publish-openapi-test-unknown-in-nested.example.com/v1\n\nDESCRIPTION:\n     preserve-unknown-properties in nested field for Testing\n\nFIELDS:\n   apiVersion\t\n     APIVersion defines the versioned schema of this representation of an\n     object. Servers should convert recognized schemas to the latest internal\n     value, and may reject unrecognized values. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n   kind\t\n     Kind is a string value representing the REST resource this object\n     represents. Servers may infer this from the endpoint the client submits\n     requests to. Cannot be updated. In CamelCase. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n   metadata\t\n     Standard object's metadata. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   spec\t\n     Specification of Waldo\n\n   status\t\n     Status of Waldo\n\n"
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+STEP: Performing setup for networking test in namespace pod-network-test-6766
+STEP: creating a selector
+STEP: Creating the service pods in kubernetes
+Oct 26 17:00:41.353: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+STEP: Creating test pods
+Oct 26 17:01:01.829: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.194.79:8080/dial?request=hostname&protocol=http&host=172.30.194.65&port=8080&tries=1'] Namespace:pod-network-test-6766 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 17:01:01.829: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 17:01:02.121: INFO: Waiting for responses: map[]
+Oct 26 17:01:02.136: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.194.79:8080/dial?request=hostname&protocol=http&host=172.30.103.234&port=8080&tries=1'] Namespace:pod-network-test-6766 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 17:01:02.136: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 17:01:02.379: INFO: Waiting for responses: map[]
+Oct 26 17:01:02.394: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.194.79:8080/dial?request=hostname&protocol=http&host=172.30.54.159&port=8080&tries=1'] Namespace:pod-network-test-6766 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 17:01:02.394: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 17:01:02.644: INFO: Waiting for responses: map[]
+[AfterEach] [sig-network] Networking
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:48:36.731: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-5100" for this suite.
+Oct 26 17:01:02.644: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pod-network-test-6766" for this suite.
 
-• [SLOW TEST:19.683 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for CRD preserving unknown fields in an embedded object [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:21.479 seconds]
+[sig-network] Networking
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
+  Granular Checks: Pods
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
+    should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance]","total":280,"completed":203,"skipped":3406,"failed":0}
-SSSSSSSSS
+{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":202,"skipped":3141,"failed":0}
+SSSSSSSSSSS
 ------------------------------
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with downward pod [LinuxOnly] [Conformance]
+[sig-node] ConfigMap 
+  should be consumable via environment variable [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Subpath
+[BeforeEach] [sig-node] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:48:36.806: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename subpath
+Oct 26 17:01:02.695: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename configmap
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
-STEP: Setting up data
-[It] should support subpaths with downward pod [LinuxOnly] [Conformance]
+[It] should be consumable via environment variable [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod pod-subpath-test-downwardapi-2flp
-STEP: Creating a pod to test atomic-volume-subpath
-Sep 21 16:48:37.205: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-2flp" in namespace "subpath-7374" to be "success or failure"
-Sep 21 16:48:37.227: INFO: Pod "pod-subpath-test-downwardapi-2flp": Phase="Pending", Reason="", readiness=false. Elapsed: 21.791507ms
-Sep 21 16:48:39.248: INFO: Pod "pod-subpath-test-downwardapi-2flp": Phase="Running", Reason="", readiness=true. Elapsed: 2.042348614s
-Sep 21 16:48:41.271: INFO: Pod "pod-subpath-test-downwardapi-2flp": Phase="Running", Reason="", readiness=true. Elapsed: 4.065990546s
-Sep 21 16:48:43.292: INFO: Pod "pod-subpath-test-downwardapi-2flp": Phase="Running", Reason="", readiness=true. Elapsed: 6.086327432s
-Sep 21 16:48:45.309: INFO: Pod "pod-subpath-test-downwardapi-2flp": Phase="Running", Reason="", readiness=true. Elapsed: 8.103587843s
-Sep 21 16:48:47.344: INFO: Pod "pod-subpath-test-downwardapi-2flp": Phase="Running", Reason="", readiness=true. Elapsed: 10.13835353s
-Sep 21 16:48:49.372: INFO: Pod "pod-subpath-test-downwardapi-2flp": Phase="Running", Reason="", readiness=true. Elapsed: 12.166361896s
-Sep 21 16:48:51.409: INFO: Pod "pod-subpath-test-downwardapi-2flp": Phase="Running", Reason="", readiness=true. Elapsed: 14.203473367s
-Sep 21 16:48:53.433: INFO: Pod "pod-subpath-test-downwardapi-2flp": Phase="Running", Reason="", readiness=true. Elapsed: 16.227524537s
-Sep 21 16:48:55.458: INFO: Pod "pod-subpath-test-downwardapi-2flp": Phase="Running", Reason="", readiness=true. Elapsed: 18.25279179s
-Sep 21 16:48:57.480: INFO: Pod "pod-subpath-test-downwardapi-2flp": Phase="Running", Reason="", readiness=true. Elapsed: 20.274514419s
-Sep 21 16:48:59.503: INFO: Pod "pod-subpath-test-downwardapi-2flp": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.29721672s
+STEP: Creating configMap configmap-189/configmap-test-b9ca8789-ba91-44d7-a7d3-d7046bc2137f
+STEP: Creating a pod to test consume configMaps
+Oct 26 17:01:02.911: INFO: Waiting up to 5m0s for pod "pod-configmaps-ceb8013b-b7b0-4672-b32b-253a364b8786" in namespace "configmap-189" to be "success or failure"
+Oct 26 17:01:02.928: INFO: Pod "pod-configmaps-ceb8013b-b7b0-4672-b32b-253a364b8786": Phase="Pending", Reason="", readiness=false. Elapsed: 17.692905ms
+Oct 26 17:01:04.944: INFO: Pod "pod-configmaps-ceb8013b-b7b0-4672-b32b-253a364b8786": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033463481s
+Oct 26 17:01:06.959: INFO: Pod "pod-configmaps-ceb8013b-b7b0-4672-b32b-253a364b8786": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.048109681s
 STEP: Saw pod success
-Sep 21 16:48:59.503: INFO: Pod "pod-subpath-test-downwardapi-2flp" satisfied condition "success or failure"
-Sep 21 16:48:59.520: INFO: Trying to get logs from node 10.241.51.147 pod pod-subpath-test-downwardapi-2flp container test-container-subpath-downwardapi-2flp: 
+Oct 26 17:01:06.959: INFO: Pod "pod-configmaps-ceb8013b-b7b0-4672-b32b-253a364b8786" satisfied condition "success or failure"
+Oct 26 17:01:06.973: INFO: Trying to get logs from node 10.72.119.74 pod pod-configmaps-ceb8013b-b7b0-4672-b32b-253a364b8786 container env-test: 
 STEP: delete the pod
-Sep 21 16:48:59.647: INFO: Waiting for pod pod-subpath-test-downwardapi-2flp to disappear
-Sep 21 16:48:59.669: INFO: Pod pod-subpath-test-downwardapi-2flp no longer exists
-STEP: Deleting pod pod-subpath-test-downwardapi-2flp
-Sep 21 16:48:59.669: INFO: Deleting pod "pod-subpath-test-downwardapi-2flp" in namespace "subpath-7374"
-[AfterEach] [sig-storage] Subpath
+Oct 26 17:01:07.060: INFO: Waiting for pod pod-configmaps-ceb8013b-b7b0-4672-b32b-253a364b8786 to disappear
+Oct 26 17:01:07.074: INFO: Pod pod-configmaps-ceb8013b-b7b0-4672-b32b-253a364b8786 no longer exists
+[AfterEach] [sig-node] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:48:59.687: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "subpath-7374" for this suite.
-
-• [SLOW TEST:22.939 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
-  Atomic writer volumes
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
-    should support subpaths with downward pod [LinuxOnly] [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Oct 26 17:01:07.074: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-189" for this suite.
+•{"msg":"PASSED [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance]","total":280,"completed":203,"skipped":3152,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [LinuxOnly] [Conformance]","total":280,"completed":204,"skipped":3415,"failed":0}
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should mutate pod and apply defaults after mutation [Conformance]
+  should honor timeout [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:48:59.745: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:01:07.110: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
@@ -10527,394 +9221,722 @@ STEP: Setting up server cert
 STEP: Create role binding to let webhook read extension-apiserver-authentication
 STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Sep 21 16:49:00.696: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 17:01:07.965: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 17:01:10.019: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328467, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328467, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328468, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328467, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Sep 21 16:49:03.868: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should mutate pod and apply defaults after mutation [Conformance]
+Oct 26 17:01:13.071: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should honor timeout [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering the mutating pod webhook via the AdmissionRegistration API
-STEP: create a pod that should be updated by the webhook
+STEP: Setting timeout (1s) shorter than webhook latency (5s)
+STEP: Registering slow webhook via the AdmissionRegistration API
+STEP: Request fails when timeout (1s) is shorter than slow webhook latency (5s)
+STEP: Having no error when timeout is shorter than webhook latency and failure policy is ignore
+STEP: Registering slow webhook via the AdmissionRegistration API
+STEP: Having no error when timeout is longer than webhook latency
+STEP: Registering slow webhook via the AdmissionRegistration API
+STEP: Having no error when timeout is empty (defaulted to 10s in v1)
+STEP: Registering slow webhook via the AdmissionRegistration API
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:49:04.219: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-8678" for this suite.
-STEP: Destroying namespace "webhook-8678-markers" for this suite.
+Oct 26 17:01:25.679: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-8278" for this suite.
+STEP: Destroying namespace "webhook-8278-markers" for this suite.
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
-•{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance]","total":280,"completed":205,"skipped":3415,"failed":0}
-SSSSSSSSSSSSSSSSSSS
+
+• [SLOW TEST:18.824 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should honor timeout [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-api-machinery] Watchers 
-  should be able to start watching from a specific resource version [Conformance]
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance]","total":280,"completed":204,"skipped":3189,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Garbage collector 
+  should not be blocked by dependency circle [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Watchers
+[BeforeEach] [sig-api-machinery] Garbage collector
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:49:04.549: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename watch
+Oct 26 17:01:25.934: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename gc
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to start watching from a specific resource version [Conformance]
+[It] should not be blocked by dependency circle [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Oct 26 17:01:26.281: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"dc6b7a2e-a813-446e-80d3-59c82871f5dd", Controller:(*bool)(0xc004596532), BlockOwnerDeletion:(*bool)(0xc004596533)}}
+Oct 26 17:01:26.317: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"d472ca9d-1311-4277-985f-1f1e824072e2", Controller:(*bool)(0xc0044f1606), BlockOwnerDeletion:(*bool)(0xc0044f1607)}}
+Oct 26 17:01:26.340: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"690a63db-ab1a-466a-a68d-d8649d97eac4", Controller:(*bool)(0xc004596776), BlockOwnerDeletion:(*bool)(0xc004596777)}}
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:01:31.389: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-8380" for this suite.
+
+• [SLOW TEST:5.490 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should not be blocked by dependency circle [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]","total":280,"completed":205,"skipped":3215,"failed":0}
+[sig-cli] Kubectl client Kubectl expose 
+  should create services for rc  [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:01:31.425: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubectl
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
+[It] should create services for rc  [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating Agnhost RC
+Oct 26 17:01:31.632: INFO: namespace kubectl-4200
+Oct 26 17:01:31.632: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-4200'
+Oct 26 17:01:32.429: INFO: stderr: ""
+Oct 26 17:01:32.429: INFO: stdout: "replicationcontroller/agnhost-master created\n"
+STEP: Waiting for Agnhost master to start.
+Oct 26 17:01:33.449: INFO: Selector matched 1 pods for map[app:agnhost]
+Oct 26 17:01:33.449: INFO: Found 0 / 1
+Oct 26 17:01:34.447: INFO: Selector matched 1 pods for map[app:agnhost]
+Oct 26 17:01:34.447: INFO: Found 0 / 1
+Oct 26 17:01:35.449: INFO: Selector matched 1 pods for map[app:agnhost]
+Oct 26 17:01:35.449: INFO: Found 1 / 1
+Oct 26 17:01:35.449: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+Oct 26 17:01:35.465: INFO: Selector matched 1 pods for map[app:agnhost]
+Oct 26 17:01:35.465: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Oct 26 17:01:35.465: INFO: wait on agnhost-master startup in kubectl-4200 
+Oct 26 17:01:35.465: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 logs agnhost-master-qlcwk agnhost-master --namespace=kubectl-4200'
+Oct 26 17:01:35.645: INFO: stderr: ""
+Oct 26 17:01:35.645: INFO: stdout: "Paused\n"
+STEP: exposing RC
+Oct 26 17:01:35.645: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 expose rc agnhost-master --name=rm2 --port=1234 --target-port=6379 --namespace=kubectl-4200'
+Oct 26 17:01:35.852: INFO: stderr: ""
+Oct 26 17:01:35.852: INFO: stdout: "service/rm2 exposed\n"
+Oct 26 17:01:35.864: INFO: Service rm2 in namespace kubectl-4200 found.
+STEP: exposing service
+Oct 26 17:01:37.897: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=kubectl-4200'
+Oct 26 17:01:38.091: INFO: stderr: ""
+Oct 26 17:01:38.091: INFO: stdout: "service/rm3 exposed\n"
+Oct 26 17:01:38.102: INFO: Service rm3 in namespace kubectl-4200 found.
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:01:40.134: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-4200" for this suite.
+
+• [SLOW TEST:8.745 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Kubectl expose
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1295
+    should create services for rc  [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-cli] Kubectl client Kubectl expose should create services for rc  [Conformance]","total":280,"completed":206,"skipped":3215,"failed":0}
+[k8s.io] InitContainer [NodeConformance] 
+  should invoke init containers on a RestartNever pod [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:01:40.170: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename init-container
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
+[It] should invoke init containers on a RestartNever pod [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating the pod
+Oct 26 17:01:40.323: INFO: PodSpec: initContainers in spec.initContainers
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:01:45.695: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "init-container-2935" for this suite.
+
+• [SLOW TEST:5.566 seconds]
+[k8s.io] InitContainer [NodeConformance]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should invoke init containers on a RestartNever pod [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance]","total":280,"completed":207,"skipped":3215,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected configMap 
+  updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:01:45.736: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a new configmap
-STEP: modifying the configmap once
-STEP: modifying the configmap a second time
-STEP: deleting the configmap
-STEP: creating a watch on configmaps from the resource version returned by the first update
-STEP: Expecting to observe notifications for all changes to the configmap after the first update
-Sep 21 16:49:04.909: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version  watch-1646 /api/v1/namespaces/watch-1646/configmaps/e2e-watch-test-resource-version 0a38a5c6-038b-40d1-9e80-804b0b1289ac 84139 0 2020-09-21 16:49:04 +0000 UTC   map[watch-this-configmap:from-resource-version] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Sep 21 16:49:04.909: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version  watch-1646 /api/v1/namespaces/watch-1646/configmaps/e2e-watch-test-resource-version 0a38a5c6-038b-40d1-9e80-804b0b1289ac 84141 0 2020-09-21 16:49:04 +0000 UTC   map[watch-this-configmap:from-resource-version] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-[AfterEach] [sig-api-machinery] Watchers
+Oct 26 17:01:45.898: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
+STEP: Creating projection with configMap that has name projected-configmap-test-upd-30859602-9762-4148-8faf-85d30937af95
+STEP: Creating the pod
+STEP: Updating configmap projected-configmap-test-upd-30859602-9762-4148-8faf-85d30937af95
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:49:04.909: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "watch-1646" for this suite.
-•{"msg":"PASSED [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance]","total":280,"completed":206,"skipped":3434,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 17:01:50.142: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9918" for this suite.
+•{"msg":"PASSED [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":208,"skipped":3241,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Update Demo 
-  should scale a replication controller  [Conformance]
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-storage] Subpath
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:49:04.985: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubectl
+Oct 26 17:01:50.191: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename subpath
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[BeforeEach] Update Demo
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:330
-[It] should scale a replication controller  [Conformance]
+[BeforeEach] Atomic writer volumes
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
+STEP: Setting up data
+[It] should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a replication controller
-Sep 21 16:49:05.194: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-812'
-Sep 21 16:49:05.763: INFO: stderr: ""
-Sep 21 16:49:05.763: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Sep 21 16:49:05.764: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-812'
-Sep 21 16:49:05.903: INFO: stderr: ""
-Sep 21 16:49:05.903: INFO: stdout: "update-demo-nautilus-8b46k update-demo-nautilus-nwl77 "
-Sep 21 16:49:05.903: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8b46k -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:06.035: INFO: stderr: ""
-Sep 21 16:49:06.035: INFO: stdout: ""
-Sep 21 16:49:06.036: INFO: update-demo-nautilus-8b46k is created but not running
-Sep 21 16:49:11.036: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-812'
-Sep 21 16:49:11.158: INFO: stderr: ""
-Sep 21 16:49:11.159: INFO: stdout: "update-demo-nautilus-8b46k update-demo-nautilus-nwl77 "
-Sep 21 16:49:11.159: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8b46k -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:11.291: INFO: stderr: ""
-Sep 21 16:49:11.291: INFO: stdout: "true"
-Sep 21 16:49:11.291: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8b46k -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:11.436: INFO: stderr: ""
-Sep 21 16:49:11.436: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Sep 21 16:49:11.436: INFO: validating pod update-demo-nautilus-8b46k
-Sep 21 16:49:11.484: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Sep 21 16:49:11.484: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Sep 21 16:49:11.484: INFO: update-demo-nautilus-8b46k is verified up and running
-Sep 21 16:49:11.484: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-nwl77 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:11.599: INFO: stderr: ""
-Sep 21 16:49:11.599: INFO: stdout: "true"
-Sep 21 16:49:11.599: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-nwl77 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:11.735: INFO: stderr: ""
-Sep 21 16:49:11.735: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Sep 21 16:49:11.735: INFO: validating pod update-demo-nautilus-nwl77
-Sep 21 16:49:11.798: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Sep 21 16:49:11.798: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Sep 21 16:49:11.798: INFO: update-demo-nautilus-nwl77 is verified up and running
-STEP: scaling down the replication controller
-Sep 21 16:49:11.803: INFO: scanned /root for discovery docs: 
-Sep 21 16:49:11.803: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=kubectl-812'
-Sep 21 16:49:13.087: INFO: stderr: ""
-Sep 21 16:49:13.087: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Sep 21 16:49:13.087: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-812'
-Sep 21 16:49:13.241: INFO: stderr: ""
-Sep 21 16:49:13.241: INFO: stdout: "update-demo-nautilus-8b46k update-demo-nautilus-nwl77 "
-STEP: Replicas for name=update-demo: expected=1 actual=2
-Sep 21 16:49:18.241: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-812'
-Sep 21 16:49:18.383: INFO: stderr: ""
-Sep 21 16:49:18.383: INFO: stdout: "update-demo-nautilus-8b46k "
-Sep 21 16:49:18.383: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8b46k -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:18.526: INFO: stderr: ""
-Sep 21 16:49:18.526: INFO: stdout: "true"
-Sep 21 16:49:18.527: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8b46k -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:18.656: INFO: stderr: ""
-Sep 21 16:49:18.656: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Sep 21 16:49:18.656: INFO: validating pod update-demo-nautilus-8b46k
-Sep 21 16:49:18.909: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Sep 21 16:49:18.909: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Sep 21 16:49:18.909: INFO: update-demo-nautilus-8b46k is verified up and running
-STEP: scaling up the replication controller
-Sep 21 16:49:18.913: INFO: scanned /root for discovery docs: 
-Sep 21 16:49:18.913: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=kubectl-812'
-Sep 21 16:49:20.399: INFO: stderr: ""
-Sep 21 16:49:20.399: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Sep 21 16:49:20.399: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-812'
-Sep 21 16:49:20.537: INFO: stderr: ""
-Sep 21 16:49:20.537: INFO: stdout: "update-demo-nautilus-8b46k update-demo-nautilus-r8jwl "
-Sep 21 16:49:20.537: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8b46k -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:20.708: INFO: stderr: ""
-Sep 21 16:49:20.708: INFO: stdout: "true"
-Sep 21 16:49:20.708: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8b46k -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:20.833: INFO: stderr: ""
-Sep 21 16:49:20.833: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Sep 21 16:49:20.833: INFO: validating pod update-demo-nautilus-8b46k
-Sep 21 16:49:20.873: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Sep 21 16:49:20.873: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Sep 21 16:49:20.873: INFO: update-demo-nautilus-8b46k is verified up and running
-Sep 21 16:49:20.873: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-r8jwl -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:21.005: INFO: stderr: ""
-Sep 21 16:49:21.005: INFO: stdout: ""
-Sep 21 16:49:21.005: INFO: update-demo-nautilus-r8jwl is created but not running
-Sep 21 16:49:26.005: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-812'
-Sep 21 16:49:26.142: INFO: stderr: ""
-Sep 21 16:49:26.142: INFO: stdout: "update-demo-nautilus-8b46k update-demo-nautilus-r8jwl "
-Sep 21 16:49:26.142: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8b46k -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:26.253: INFO: stderr: ""
-Sep 21 16:49:26.253: INFO: stdout: "true"
-Sep 21 16:49:26.253: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-8b46k -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:26.401: INFO: stderr: ""
-Sep 21 16:49:26.401: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Sep 21 16:49:26.401: INFO: validating pod update-demo-nautilus-8b46k
-Sep 21 16:49:26.442: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Sep 21 16:49:26.442: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Sep 21 16:49:26.442: INFO: update-demo-nautilus-8b46k is verified up and running
-Sep 21 16:49:26.442: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-r8jwl -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:26.569: INFO: stderr: ""
-Sep 21 16:49:26.569: INFO: stdout: "true"
-Sep 21 16:49:26.569: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods update-demo-nautilus-r8jwl -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-812'
-Sep 21 16:49:26.673: INFO: stderr: ""
-Sep 21 16:49:26.673: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Sep 21 16:49:26.673: INFO: validating pod update-demo-nautilus-r8jwl
-Sep 21 16:49:26.734: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Sep 21 16:49:26.734: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Sep 21 16:49:26.734: INFO: update-demo-nautilus-r8jwl is verified up and running
-STEP: using delete to clean up resources
-Sep 21 16:49:26.735: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete --grace-period=0 --force -f - --namespace=kubectl-812'
-Sep 21 16:49:26.922: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Sep 21 16:49:26.922: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n"
-Sep 21 16:49:26.922: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-812'
-Sep 21 16:49:27.118: INFO: stderr: "No resources found in kubectl-812 namespace.\n"
-Sep 21 16:49:27.119: INFO: stdout: ""
-Sep 21 16:49:27.119: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -l name=update-demo --namespace=kubectl-812 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
-Sep 21 16:49:27.242: INFO: stderr: ""
-Sep 21 16:49:27.242: INFO: stdout: "update-demo-nautilus-8b46k\nupdate-demo-nautilus-r8jwl\n"
-Sep 21 16:49:27.743: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-812'
-Sep 21 16:49:28.943: INFO: stderr: "No resources found in kubectl-812 namespace.\n"
-Sep 21 16:49:28.943: INFO: stdout: ""
-Sep 21 16:49:28.943: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -l name=update-demo --namespace=kubectl-812 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
-Sep 21 16:49:29.072: INFO: stderr: ""
-Sep 21 16:49:29.072: INFO: stdout: ""
-[AfterEach] [sig-cli] Kubectl client
+STEP: Creating pod pod-subpath-test-configmap-hd9n
+STEP: Creating a pod to test atomic-volume-subpath
+Oct 26 17:01:50.463: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-hd9n" in namespace "subpath-8482" to be "success or failure"
+Oct 26 17:01:50.478: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Pending", Reason="", readiness=false. Elapsed: 15.025579ms
+Oct 26 17:01:52.495: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031347793s
+Oct 26 17:01:54.518: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Running", Reason="", readiness=true. Elapsed: 4.054930058s
+Oct 26 17:01:56.535: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Running", Reason="", readiness=true. Elapsed: 6.071776515s
+Oct 26 17:01:58.551: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Running", Reason="", readiness=true. Elapsed: 8.087686488s
+Oct 26 17:02:00.572: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Running", Reason="", readiness=true. Elapsed: 10.108866413s
+Oct 26 17:02:02.587: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Running", Reason="", readiness=true. Elapsed: 12.12422936s
+Oct 26 17:02:04.604: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Running", Reason="", readiness=true. Elapsed: 14.140458686s
+Oct 26 17:02:06.620: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Running", Reason="", readiness=true. Elapsed: 16.157016973s
+Oct 26 17:02:08.638: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Running", Reason="", readiness=true. Elapsed: 18.174623778s
+Oct 26 17:02:10.701: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Running", Reason="", readiness=true. Elapsed: 20.238207192s
+Oct 26 17:02:12.718: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Running", Reason="", readiness=true. Elapsed: 22.254561612s
+Oct 26 17:02:14.736: INFO: Pod "pod-subpath-test-configmap-hd9n": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.272570643s
+STEP: Saw pod success
+Oct 26 17:02:14.736: INFO: Pod "pod-subpath-test-configmap-hd9n" satisfied condition "success or failure"
+Oct 26 17:02:14.752: INFO: Trying to get logs from node 10.72.119.74 pod pod-subpath-test-configmap-hd9n container test-container-subpath-configmap-hd9n: 
+STEP: delete the pod
+Oct 26 17:02:14.846: INFO: Waiting for pod pod-subpath-test-configmap-hd9n to disappear
+Oct 26 17:02:14.864: INFO: Pod pod-subpath-test-configmap-hd9n no longer exists
+STEP: Deleting pod pod-subpath-test-configmap-hd9n
+Oct 26 17:02:14.864: INFO: Deleting pod "pod-subpath-test-configmap-hd9n" in namespace "subpath-8482"
+[AfterEach] [sig-storage] Subpath
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:49:29.072: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-812" for this suite.
+Oct 26 17:02:14.880: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-8482" for this suite.
 
-• [SLOW TEST:24.274 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Update Demo
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:328
-    should scale a replication controller  [Conformance]
+• [SLOW TEST:24.721 seconds]
+[sig-storage] Subpath
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
+  Atomic writer volumes
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
+    should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
     /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Update Demo should scale a replication controller  [Conformance]","total":280,"completed":207,"skipped":3466,"failed":0}
+{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]","total":280,"completed":209,"skipped":3286,"failed":0}
 SSS
 ------------------------------
-[k8s.io] Container Runtime blackbox test on terminated container 
-  should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
+[k8s.io] Probing container 
+  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Runtime
+[BeforeEach] [k8s.io] Probing container
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:49:29.259: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-runtime
+Oct 26 17:02:14.912: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename container-probe
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] Probing container
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the container
-STEP: wait for the container to reach Succeeded
-STEP: get the container status
-STEP: the container should be terminated
-STEP: the termination message should be set
-Sep 21 16:49:31.858: INFO: Expected: &{OK} to match Container's Termination Message: OK --
-STEP: delete the container
-[AfterEach] [k8s.io] Container Runtime
+Oct 26 17:02:39.187: INFO: Container started at 2020-10-26 17:02:16 +0000 UTC, pod became ready at 2020-10-26 17:02:37 +0000 UTC
+[AfterEach] [k8s.io] Probing container
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:49:31.964: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-runtime-5576" for this suite.
-•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":280,"completed":208,"skipped":3469,"failed":0}
-SS
+Oct 26 17:02:39.188: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-1736" for this suite.
+
+• [SLOW TEST:24.314 seconds]
+[k8s.io] Probing container
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] [sig-node] PreStop 
-  should call prestop when killing a pod  [Conformance]
+{"msg":"PASSED [k8s.io] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]","total":280,"completed":210,"skipped":3289,"failed":0}
+SSS
+------------------------------
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates resource limits of pods that are allowed to run  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] [sig-node] PreStop
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:49:32.042: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename prestop
+Oct 26 17:02:39.227: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename sched-pred
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] [sig-node] PreStop
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pre_stop.go:172
-[It] should call prestop when killing a pod  [Conformance]
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
+Oct 26 17:02:39.394: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Oct 26 17:02:39.442: INFO: Waiting for terminating namespaces to be deleted...
+Oct 26 17:02:39.458: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.74 before test
+Oct 26 17:02:39.532: INFO: openshift-kube-proxy-7d26g from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: tuned-p4cpc from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: node-ca-4ppdl from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: tigera-operator-5c45c894b4-z9bnx from tigera-operator started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container tigera-operator ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: ibm-master-proxy-static-10.72.119.74 from kube-system started at 2020-10-26 14:34:26 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: 	Container pause ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: multus-cc44t from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: alertmanager-main-0 from openshift-monitoring started at 2020-10-26 16:09:57 +0000 UTC (3 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: multus-admission-controller-dnt9r from openshift-multus started at 2020-10-26 16:10:16 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: packageserver-5f76c64df7-jg7sj from openshift-operator-lifecycle-manager started at 2020-10-26 16:10:08 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container packageserver ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: vpn-5b455f6f88-64q7j from kube-system started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container vpn ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: calico-typha-7fd49c779d-46t4v from calico-system started at 2020-10-26 14:37:36 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: sonobuoy from sonobuoy started at 2020-10-26 15:59:53 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: sonobuoy-e2e-job-c6d919d638514ce4 from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container e2e ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: thanos-querier-84d7d58bc-c27ss from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (4 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: 	Container oauth-proxy ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: 	Container thanos-querier ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: grafana-f999495bf-t4tcv from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container grafana ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: 	Container grafana-proxy ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: telemeter-client-d5d9896c4-jpd8z from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (3 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: 	Container reload ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: 	Container telemeter-client ready: true, restart count 0
+Oct 26 17:02:39.532: INFO: ibmcloud-block-storage-driver-ws85j from kube-system started at 2020-10-26 14:34:32 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.532: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: node-exporter-ls6pf from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.533: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: dns-default-s9kvh from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.533: INFO: 	Container dns ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: prometheus-adapter-674456c5d6-p6bj5 from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.533: INFO: 	Container prometheus-adapter ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: console-5dd98d6f6f-zxw5x from openshift-console started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.533: INFO: 	Container console ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: calico-node-xlzf4 from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.533: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: cluster-samples-operator-848994bd88-rw4rq from openshift-cluster-samples-operator started at 2020-10-26 16:09:41 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.533: INFO: 	Container cluster-samples-operator ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: 	Container cluster-samples-operator-watch ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: test-webserver-a51736ea-1327-4690-990c-c60b0dd2f5fe from container-probe-1736 started at 2020-10-26 17:02:15 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.533: INFO: 	Container test-webserver ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: ibm-keepalived-watcher-49jbs from kube-system started at 2020-10-26 14:34:28 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.533: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: redhat-operators-5579777cd5-q5qm5 from openshift-marketplace started at 2020-10-26 16:39:28 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.533: INFO: 	Container redhat-operators ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-nbr6w from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.533: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Oct 26 17:02:39.533: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: service-ca-7d9db9c5b8-4w4cj from openshift-service-ca started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.533: INFO: 	Container service-ca-controller ready: true, restart count 0
+Oct 26 17:02:39.533: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.76 before test
+Oct 26 17:02:39.679: INFO: alertmanager-main-2 from openshift-monitoring started at 2020-10-26 14:44:28 +0000 UTC (3 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-2scb4 from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Oct 26 17:02:39.679: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: router-default-84766ff4fd-cftxw from openshift-ingress started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container router ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: ibmcloud-block-storage-driver-b724k from kube-system started at 2020-10-26 14:34:34 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: openshift-service-catalog-apiserver-operator-7cfbcc49dc-hrn66 from openshift-service-catalog-apiserver-operator started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container operator ready: true, restart count 1
+Oct 26 17:02:39.679: INFO: ibm-storage-watcher-77d6cc685c-bvg8q from kube-system started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: olm-operator-7cd6cbf579-zww2l from openshift-operator-lifecycle-manager started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container olm-operator ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: cluster-monitoring-operator-69649564f8-r4h9b from openshift-monitoring started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container cluster-monitoring-operator ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: multus-lbpcn from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: openshift-service-catalog-controller-manager-operator-d49bz89wp from openshift-service-catalog-controller-manager-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container operator ready: true, restart count 1
+Oct 26 17:02:39.679: INFO: multus-admission-controller-5j529 from openshift-multus started at 2020-10-26 14:35:50 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: node-exporter-zmwjz from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: node-ca-bpxm5 from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: ibm-cloud-provider-ip-158-176-78-170-5cdb49c95f-76fjw from ibm-system started at 2020-10-26 14:42:32 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container ibm-cloud-provider-ip-158-176-78-170 ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: ingress-operator-65878dcc8d-wdnrl from openshift-ingress-operator started at 2020-10-26 14:35:49 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container ingress-operator ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: downloads-74f6b6dcb6-r9ksv from openshift-console started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container download-server ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: catalog-operator-5db6d5849c-z2lgx from openshift-operator-lifecycle-manager started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container catalog-operator ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: calico-kube-controllers-cdc7f7f8f-jqklm from calico-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container calico-kube-controllers ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: downloads-74f6b6dcb6-vqzf9 from openshift-console started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container download-server ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: cluster-node-tuning-operator-575b98978c-zz7s9 from openshift-cluster-node-tuning-operator started at 2020-10-26 14:35:51 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container cluster-node-tuning-operator ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: calico-node-dhd7p from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: kube-storage-version-migrator-operator-86d479ffb5-9cmml from openshift-kube-storage-version-migrator-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container kube-storage-version-migrator-operator ready: true, restart count 1
+Oct 26 17:02:39.679: INFO: ibm-keepalived-watcher-qwg48 from kube-system started at 2020-10-26 14:34:30 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: ibmcloud-block-storage-plugin-5f486d686f-6brjv from kube-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container ibmcloud-block-storage-plugin-container ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: prometheus-k8s-0 from openshift-monitoring started at 2020-10-26 14:45:39 +0000 UTC (7 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container prometheus ready: true, restart count 1
+Oct 26 17:02:39.679: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container prometheus-proxy ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container thanos-sidecar ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: ibm-master-proxy-static-10.72.119.76 from kube-system started at 2020-10-26 14:34:28 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container pause ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: cluster-image-registry-operator-8444cfb976-gdjk6 from openshift-image-registry started at 2020-10-26 14:35:50 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container cluster-image-registry-operator ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container cluster-image-registry-operator-watch ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: cluster-storage-operator-9b57569c9-gpqd2 from openshift-cluster-storage-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container cluster-storage-operator ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: service-ca-operator-656f67898-svtt9 from openshift-service-ca-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container operator ready: true, restart count 1
+Oct 26 17:02:39.679: INFO: marketplace-operator-5df598b96b-6vr64 from openshift-marketplace started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container marketplace-operator ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: console-operator-5fc674d8d4-fjctd from openshift-console-operator started at 2020-10-26 14:35:51 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container console-operator ready: true, restart count 1
+Oct 26 17:02:39.679: INFO: tuned-s6flg from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: network-operator-86b67c6f7f-pcq2w from openshift-network-operator started at 2020-10-26 14:34:37 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container network-operator ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: dns-operator-58d86f779c-wbm98 from openshift-dns-operator started at 2020-10-26 14:35:49 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container dns-operator ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: calico-typha-7fd49c779d-766qb from calico-system started at 2020-10-26 14:37:36 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: dns-default-z8pn9 from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container dns ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: openshift-kube-proxy-74ss4 from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: ibm-file-plugin-944b59c94-k7mrm from kube-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.679: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
+Oct 26 17:02:39.679: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.82 before test
+Oct 26 17:02:39.801: INFO: calico-typha-7fd49c779d-nmzbk from calico-system started at 2020-10-26 14:35:39 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: test-k8s-e2e-pvg-master-verification from default started at 2020-10-26 14:37:46 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: registry-pvc-permissions-m9wwv from openshift-image-registry started at 2020-10-26 14:40:09 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container pvc-permissions ready: false, restart count 0
+Oct 26 17:02:39.801: INFO: prometheus-adapter-674456c5d6-kj77x from openshift-monitoring started at 2020-10-26 14:44:24 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container prometheus-adapter ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: image-registry-6db5c967cf-d4ljt from openshift-image-registry started at 2020-10-26 14:40:09 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container registry ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-x2lkl from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Oct 26 17:02:39.801: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: openshift-kube-proxy-z9sjt from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: openshift-state-metrics-d4d5d6f57-lszmv from openshift-monitoring started at 2020-10-26 14:36:34 +0000 UTC (3 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container openshift-state-metrics ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: node-exporter-4lfg7 from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: migrator-686fc6cc66-9wpj9 from openshift-kube-storage-version-migrator started at 2020-10-26 14:36:53 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container migrator ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: prometheus-operator-55f946f5d-tpnd8 from openshift-monitoring started at 2020-10-26 14:44:11 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container prometheus-operator ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: router-default-84766ff4fd-7dgrd from openshift-ingress started at 2020-10-26 14:38:40 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container router ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: multus-admission-controller-9rv7d from openshift-multus started at 2020-10-26 14:36:02 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: ibm-cloud-provider-ip-158-176-78-170-5cdb49c95f-89x4j from ibm-system started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container ibm-cloud-provider-ip-158-176-78-170 ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: packageserver-5f76c64df7-6nrvc from openshift-operator-lifecycle-manager started at 2020-10-26 16:10:11 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container packageserver ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: calico-node-c7wjs from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: dns-default-ggg6q from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container dns ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: prometheus-k8s-1 from openshift-monitoring started at 2020-10-26 14:45:29 +0000 UTC (7 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container prometheus ready: true, restart count 1
+Oct 26 17:02:39.801: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container prometheus-proxy ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container thanos-sidecar ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: console-5dd98d6f6f-xvz7c from openshift-console started at 2020-10-26 14:45:54 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container console ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: thanos-querier-84d7d58bc-mpnzq from openshift-monitoring started at 2020-10-26 14:45:18 +0000 UTC (4 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container oauth-proxy ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container thanos-querier ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: community-operators-f4df99f64-twf5k from openshift-marketplace started at 2020-10-26 14:38:39 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container community-operators ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: ibm-master-proxy-static-10.72.119.82 from kube-system started at 2020-10-26 14:34:29 +0000 UTC (2 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container pause ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: ibmcloud-block-storage-driver-sxlj2 from kube-system started at 2020-10-26 14:34:40 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: multus-hwkgh from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: kube-state-metrics-7bfb6f7864-vmgz9 from openshift-monitoring started at 2020-10-26 14:36:32 +0000 UTC (3 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container kube-state-metrics ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: node-ca-z564l from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: alertmanager-main-1 from openshift-monitoring started at 2020-10-26 14:44:39 +0000 UTC (3 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: ibm-keepalived-watcher-98n85 from kube-system started at 2020-10-26 14:34:32 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: tuned-h578f from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: certified-operators-69bb9f65cc-9mhpt from openshift-marketplace started at 2020-10-26 14:38:38 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container certified-operators ready: true, restart count 0
+Oct 26 17:02:39.801: INFO: redhat-marketplace-77d9b69f69-qb5s9 from openshift-marketplace started at 2020-10-26 14:38:39 +0000 UTC (1 container statuses recorded)
+Oct 26 17:02:39.801: INFO: 	Container redhat-marketplace ready: true, restart count 0
+[It] validates resource limits of pods that are allowed to run  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating server pod server in namespace prestop-7072
-STEP: Waiting for pods to come up.
-STEP: Creating tester pod tester in namespace prestop-7072
-STEP: Deleting pre-stop pod
-Sep 21 16:49:41.877: INFO: Saw: {
-	"Hostname": "server",
-	"Sent": null,
-	"Received": {
-		"prestop": 1
-	},
-	"Errors": null,
-	"Log": [
-		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.",
-		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up."
-	],
-	"StillContactingPeers": true
-}
-STEP: Deleting the server pod
-[AfterEach] [k8s.io] [sig-node] PreStop
+STEP: verifying the node has the label node 10.72.119.74
+STEP: verifying the node has the label node 10.72.119.76
+STEP: verifying the node has the label node 10.72.119.82
+Oct 26 17:02:40.020: INFO: Pod calico-kube-controllers-cdc7f7f8f-jqklm requesting resource cpu=0m on Node 10.72.119.76
+Oct 26 17:02:40.020: INFO: Pod calico-node-c7wjs requesting resource cpu=0m on Node 10.72.119.82
+Oct 26 17:02:40.020: INFO: Pod calico-node-dhd7p requesting resource cpu=0m on Node 10.72.119.76
+Oct 26 17:02:40.020: INFO: Pod calico-node-xlzf4 requesting resource cpu=0m on Node 10.72.119.74
+Oct 26 17:02:40.021: INFO: Pod calico-typha-7fd49c779d-46t4v requesting resource cpu=0m on Node 10.72.119.74
+Oct 26 17:02:40.021: INFO: Pod calico-typha-7fd49c779d-766qb requesting resource cpu=0m on Node 10.72.119.76
+Oct 26 17:02:40.021: INFO: Pod calico-typha-7fd49c779d-nmzbk requesting resource cpu=0m on Node 10.72.119.82
+Oct 26 17:02:40.021: INFO: Pod test-webserver-a51736ea-1327-4690-990c-c60b0dd2f5fe requesting resource cpu=0m on Node 10.72.119.74
+Oct 26 17:02:40.021: INFO: Pod test-k8s-e2e-pvg-master-verification requesting resource cpu=0m on Node 10.72.119.82
+Oct 26 17:02:40.021: INFO: Pod ibm-cloud-provider-ip-158-176-78-170-5cdb49c95f-76fjw requesting resource cpu=5m on Node 10.72.119.76
+Oct 26 17:02:40.021: INFO: Pod ibm-cloud-provider-ip-158-176-78-170-5cdb49c95f-89x4j requesting resource cpu=5m on Node 10.72.119.82
+Oct 26 17:02:40.021: INFO: Pod ibm-file-plugin-944b59c94-k7mrm requesting resource cpu=50m on Node 10.72.119.76
+Oct 26 17:02:40.022: INFO: Pod ibm-keepalived-watcher-49jbs requesting resource cpu=5m on Node 10.72.119.74
+Oct 26 17:02:40.022: INFO: Pod ibm-keepalived-watcher-98n85 requesting resource cpu=5m on Node 10.72.119.82
+Oct 26 17:02:40.022: INFO: Pod ibm-keepalived-watcher-qwg48 requesting resource cpu=5m on Node 10.72.119.76
+Oct 26 17:02:40.022: INFO: Pod ibm-master-proxy-static-10.72.119.74 requesting resource cpu=25m on Node 10.72.119.74
+Oct 26 17:02:40.022: INFO: Pod ibm-master-proxy-static-10.72.119.76 requesting resource cpu=25m on Node 10.72.119.76
+Oct 26 17:02:40.022: INFO: Pod ibm-master-proxy-static-10.72.119.82 requesting resource cpu=25m on Node 10.72.119.82
+Oct 26 17:02:40.022: INFO: Pod ibm-storage-watcher-77d6cc685c-bvg8q requesting resource cpu=50m on Node 10.72.119.76
+Oct 26 17:02:40.022: INFO: Pod ibmcloud-block-storage-driver-b724k requesting resource cpu=50m on Node 10.72.119.76
+Oct 26 17:02:40.023: INFO: Pod ibmcloud-block-storage-driver-sxlj2 requesting resource cpu=50m on Node 10.72.119.82
+Oct 26 17:02:40.023: INFO: Pod ibmcloud-block-storage-driver-ws85j requesting resource cpu=50m on Node 10.72.119.74
+Oct 26 17:02:40.023: INFO: Pod ibmcloud-block-storage-plugin-5f486d686f-6brjv requesting resource cpu=50m on Node 10.72.119.76
+Oct 26 17:02:40.023: INFO: Pod vpn-5b455f6f88-64q7j requesting resource cpu=5m on Node 10.72.119.74
+Oct 26 17:02:40.023: INFO: Pod cluster-node-tuning-operator-575b98978c-zz7s9 requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.023: INFO: Pod tuned-h578f requesting resource cpu=10m on Node 10.72.119.82
+Oct 26 17:02:40.023: INFO: Pod tuned-p4cpc requesting resource cpu=10m on Node 10.72.119.74
+Oct 26 17:02:40.023: INFO: Pod tuned-s6flg requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.024: INFO: Pod cluster-samples-operator-848994bd88-rw4rq requesting resource cpu=20m on Node 10.72.119.74
+Oct 26 17:02:40.024: INFO: Pod cluster-storage-operator-9b57569c9-gpqd2 requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.024: INFO: Pod console-operator-5fc674d8d4-fjctd requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.024: INFO: Pod console-5dd98d6f6f-xvz7c requesting resource cpu=10m on Node 10.72.119.82
+Oct 26 17:02:40.024: INFO: Pod console-5dd98d6f6f-zxw5x requesting resource cpu=10m on Node 10.72.119.74
+Oct 26 17:02:40.024: INFO: Pod downloads-74f6b6dcb6-r9ksv requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.024: INFO: Pod downloads-74f6b6dcb6-vqzf9 requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.024: INFO: Pod dns-operator-58d86f779c-wbm98 requesting resource cpu=20m on Node 10.72.119.76
+Oct 26 17:02:40.025: INFO: Pod dns-default-ggg6q requesting resource cpu=110m on Node 10.72.119.82
+Oct 26 17:02:40.025: INFO: Pod dns-default-s9kvh requesting resource cpu=110m on Node 10.72.119.74
+Oct 26 17:02:40.025: INFO: Pod dns-default-z8pn9 requesting resource cpu=110m on Node 10.72.119.76
+Oct 26 17:02:40.025: INFO: Pod cluster-image-registry-operator-8444cfb976-gdjk6 requesting resource cpu=20m on Node 10.72.119.76
+Oct 26 17:02:40.025: INFO: Pod image-registry-6db5c967cf-d4ljt requesting resource cpu=100m on Node 10.72.119.82
+Oct 26 17:02:40.025: INFO: Pod node-ca-4ppdl requesting resource cpu=10m on Node 10.72.119.74
+Oct 26 17:02:40.025: INFO: Pod node-ca-bpxm5 requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.025: INFO: Pod node-ca-z564l requesting resource cpu=10m on Node 10.72.119.82
+Oct 26 17:02:40.025: INFO: Pod ingress-operator-65878dcc8d-wdnrl requesting resource cpu=20m on Node 10.72.119.76
+Oct 26 17:02:40.026: INFO: Pod router-default-84766ff4fd-7dgrd requesting resource cpu=100m on Node 10.72.119.82
+Oct 26 17:02:40.026: INFO: Pod router-default-84766ff4fd-cftxw requesting resource cpu=100m on Node 10.72.119.76
+Oct 26 17:02:40.026: INFO: Pod openshift-kube-proxy-74ss4 requesting resource cpu=100m on Node 10.72.119.76
+Oct 26 17:02:40.026: INFO: Pod openshift-kube-proxy-7d26g requesting resource cpu=100m on Node 10.72.119.74
+Oct 26 17:02:40.026: INFO: Pod openshift-kube-proxy-z9sjt requesting resource cpu=100m on Node 10.72.119.82
+Oct 26 17:02:40.026: INFO: Pod kube-storage-version-migrator-operator-86d479ffb5-9cmml requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.026: INFO: Pod migrator-686fc6cc66-9wpj9 requesting resource cpu=100m on Node 10.72.119.82
+Oct 26 17:02:40.026: INFO: Pod certified-operators-69bb9f65cc-9mhpt requesting resource cpu=10m on Node 10.72.119.82
+Oct 26 17:02:40.027: INFO: Pod community-operators-f4df99f64-twf5k requesting resource cpu=10m on Node 10.72.119.82
+Oct 26 17:02:40.027: INFO: Pod marketplace-operator-5df598b96b-6vr64 requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.027: INFO: Pod redhat-marketplace-77d9b69f69-qb5s9 requesting resource cpu=10m on Node 10.72.119.82
+Oct 26 17:02:40.027: INFO: Pod redhat-operators-5579777cd5-q5qm5 requesting resource cpu=10m on Node 10.72.119.74
+Oct 26 17:02:40.027: INFO: Pod alertmanager-main-0 requesting resource cpu=6m on Node 10.72.119.74
+Oct 26 17:02:40.027: INFO: Pod alertmanager-main-1 requesting resource cpu=6m on Node 10.72.119.82
+Oct 26 17:02:40.027: INFO: Pod alertmanager-main-2 requesting resource cpu=6m on Node 10.72.119.76
+Oct 26 17:02:40.027: INFO: Pod cluster-monitoring-operator-69649564f8-r4h9b requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.027: INFO: Pod grafana-f999495bf-t4tcv requesting resource cpu=5m on Node 10.72.119.74
+Oct 26 17:02:40.028: INFO: Pod kube-state-metrics-7bfb6f7864-vmgz9 requesting resource cpu=4m on Node 10.72.119.82
+Oct 26 17:02:40.028: INFO: Pod node-exporter-4lfg7 requesting resource cpu=9m on Node 10.72.119.82
+Oct 26 17:02:40.028: INFO: Pod node-exporter-ls6pf requesting resource cpu=9m on Node 10.72.119.74
+Oct 26 17:02:40.028: INFO: Pod node-exporter-zmwjz requesting resource cpu=9m on Node 10.72.119.76
+Oct 26 17:02:40.028: INFO: Pod openshift-state-metrics-d4d5d6f57-lszmv requesting resource cpu=3m on Node 10.72.119.82
+Oct 26 17:02:40.028: INFO: Pod prometheus-adapter-674456c5d6-kj77x requesting resource cpu=1m on Node 10.72.119.82
+Oct 26 17:02:40.028: INFO: Pod prometheus-adapter-674456c5d6-p6bj5 requesting resource cpu=1m on Node 10.72.119.74
+Oct 26 17:02:40.028: INFO: Pod prometheus-k8s-0 requesting resource cpu=76m on Node 10.72.119.76
+Oct 26 17:02:40.029: INFO: Pod prometheus-k8s-1 requesting resource cpu=76m on Node 10.72.119.82
+Oct 26 17:02:40.029: INFO: Pod prometheus-operator-55f946f5d-tpnd8 requesting resource cpu=5m on Node 10.72.119.82
+Oct 26 17:02:40.029: INFO: Pod telemeter-client-d5d9896c4-jpd8z requesting resource cpu=3m on Node 10.72.119.74
+Oct 26 17:02:40.029: INFO: Pod thanos-querier-84d7d58bc-c27ss requesting resource cpu=8m on Node 10.72.119.74
+Oct 26 17:02:40.029: INFO: Pod thanos-querier-84d7d58bc-mpnzq requesting resource cpu=8m on Node 10.72.119.82
+Oct 26 17:02:40.029: INFO: Pod multus-admission-controller-5j529 requesting resource cpu=20m on Node 10.72.119.76
+Oct 26 17:02:40.029: INFO: Pod multus-admission-controller-9rv7d requesting resource cpu=20m on Node 10.72.119.82
+Oct 26 17:02:40.030: INFO: Pod multus-admission-controller-dnt9r requesting resource cpu=20m on Node 10.72.119.74
+Oct 26 17:02:40.030: INFO: Pod multus-cc44t requesting resource cpu=10m on Node 10.72.119.74
+Oct 26 17:02:40.030: INFO: Pod multus-hwkgh requesting resource cpu=10m on Node 10.72.119.82
+Oct 26 17:02:40.030: INFO: Pod multus-lbpcn requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.030: INFO: Pod network-operator-86b67c6f7f-pcq2w requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.030: INFO: Pod catalog-operator-5db6d5849c-z2lgx requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.030: INFO: Pod olm-operator-7cd6cbf579-zww2l requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.030: INFO: Pod packageserver-5f76c64df7-6nrvc requesting resource cpu=10m on Node 10.72.119.82
+Oct 26 17:02:40.031: INFO: Pod packageserver-5f76c64df7-jg7sj requesting resource cpu=10m on Node 10.72.119.74
+Oct 26 17:02:40.031: INFO: Pod service-ca-operator-656f67898-svtt9 requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.031: INFO: Pod service-ca-7d9db9c5b8-4w4cj requesting resource cpu=10m on Node 10.72.119.74
+Oct 26 17:02:40.031: INFO: Pod openshift-service-catalog-apiserver-operator-7cfbcc49dc-hrn66 requesting resource cpu=0m on Node 10.72.119.76
+Oct 26 17:02:40.031: INFO: Pod openshift-service-catalog-controller-manager-operator-d49bz89wp requesting resource cpu=10m on Node 10.72.119.76
+Oct 26 17:02:40.031: INFO: Pod sonobuoy requesting resource cpu=0m on Node 10.72.119.74
+Oct 26 17:02:40.031: INFO: Pod sonobuoy-e2e-job-c6d919d638514ce4 requesting resource cpu=0m on Node 10.72.119.74
+Oct 26 17:02:40.031: INFO: Pod sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-2scb4 requesting resource cpu=0m on Node 10.72.119.76
+Oct 26 17:02:40.031: INFO: Pod sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-nbr6w requesting resource cpu=0m on Node 10.72.119.74
+Oct 26 17:02:40.032: INFO: Pod sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-x2lkl requesting resource cpu=0m on Node 10.72.119.82
+Oct 26 17:02:40.032: INFO: Pod tigera-operator-5c45c894b4-z9bnx requesting resource cpu=100m on Node 10.72.119.74
+STEP: Starting Pods to consume most of the cluster CPU.
+Oct 26 17:02:40.032: INFO: Creating a pod which consumes cpu=2361m on Node 10.72.119.74
+Oct 26 17:02:40.091: INFO: Creating a pod which consumes cpu=2123m on Node 10.72.119.76
+Oct 26 17:02:40.134: INFO: Creating a pod which consumes cpu=2172m on Node 10.72.119.82
+STEP: Creating another pod that requires unavailable amount of CPU.
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-7ff16840-7587-4e0b-bb71-722b6877e83a.16419989f6daf3a5], Reason = [Scheduled], Message = [Successfully assigned sched-pred-6466/filler-pod-7ff16840-7587-4e0b-bb71-722b6877e83a to 10.72.119.76]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-7ff16840-7587-4e0b-bb71-722b6877e83a.1641998a3c4e41dd], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-7ff16840-7587-4e0b-bb71-722b6877e83a.1641998a4b51734e], Reason = [Created], Message = [Created container filler-pod-7ff16840-7587-4e0b-bb71-722b6877e83a]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-7ff16840-7587-4e0b-bb71-722b6877e83a.1641998a4e017915], Reason = [Started], Message = [Started container filler-pod-7ff16840-7587-4e0b-bb71-722b6877e83a]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-939d7abc-0df9-46d7-8396-2597d3c10f8f.16419989f92509e7], Reason = [Scheduled], Message = [Successfully assigned sched-pred-6466/filler-pod-939d7abc-0df9-46d7-8396-2597d3c10f8f to 10.72.119.82]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-939d7abc-0df9-46d7-8396-2597d3c10f8f.1641998a407999b5], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-939d7abc-0df9-46d7-8396-2597d3c10f8f.1641998a4e313bf2], Reason = [Created], Message = [Created container filler-pod-939d7abc-0df9-46d7-8396-2597d3c10f8f]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-939d7abc-0df9-46d7-8396-2597d3c10f8f.1641998a50b79d71], Reason = [Started], Message = [Started container filler-pod-939d7abc-0df9-46d7-8396-2597d3c10f8f]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-cdaf1d11-cb07-4218-b3d0-e1013c470fbe.16419989f46749c3], Reason = [Scheduled], Message = [Successfully assigned sched-pred-6466/filler-pod-cdaf1d11-cb07-4218-b3d0-e1013c470fbe to 10.72.119.74]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-cdaf1d11-cb07-4218-b3d0-e1013c470fbe.1641998a3c455627], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-cdaf1d11-cb07-4218-b3d0-e1013c470fbe.1641998a490121a4], Reason = [Created], Message = [Created container filler-pod-cdaf1d11-cb07-4218-b3d0-e1013c470fbe]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-cdaf1d11-cb07-4218-b3d0-e1013c470fbe.1641998a4b4f5461], Reason = [Started], Message = [Started container filler-pod-cdaf1d11-cb07-4218-b3d0-e1013c470fbe]
+STEP: Considering event: 
+Type = [Warning], Name = [additional-pod.1641998aef037b3e], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 Insufficient cpu.]
+STEP: removing the label node off the node 10.72.119.74
+STEP: verifying the node doesn't have the label node
+STEP: removing the label node off the node 10.72.119.76
+STEP: verifying the node doesn't have the label node
+STEP: removing the label node off the node 10.72.119.82
+STEP: verifying the node doesn't have the label node
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:49:41.928: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "prestop-7072" for this suite.
+Oct 26 17:02:45.534: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-6466" for this suite.
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
 
-• [SLOW TEST:9.957 seconds]
-[k8s.io] [sig-node] PreStop
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should call prestop when killing a pod  [Conformance]
+• [SLOW TEST:6.344 seconds]
+[sig-scheduling] SchedulerPredicates [Serial]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
+  validates resource limits of pods that are allowed to run  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] [sig-node] PreStop should call prestop when killing a pod  [Conformance]","total":280,"completed":209,"skipped":3471,"failed":0}
-SS
+{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run  [Conformance]","total":280,"completed":211,"skipped":3292,"failed":0}
+SSSSS
 ------------------------------
-[sig-network] Proxy version v1 
-  should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should unconditionally reject operations on fail closed webhook [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] version v1
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:49:42.000: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename proxy
+Oct 26 17:02:45.572: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Oct 26 17:02:46.685: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 17:02:48.732: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328566, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328566, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328566, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328566, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Oct 26 17:02:51.781: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should unconditionally reject operations on fail closed webhook [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:49:42.396: INFO: (0) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 94.108959ms)
-Sep 21 16:49:42.432: INFO: (1) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 35.49651ms)
-Sep 21 16:49:42.477: INFO: (2) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 45.014535ms)
-Sep 21 16:49:42.517: INFO: (3) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 40.379494ms)
-Sep 21 16:49:42.555: INFO: (4) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 38.382694ms)
-Sep 21 16:49:42.595: INFO: (5) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 39.353419ms)
-Sep 21 16:49:42.638: INFO: (6) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 43.221741ms)
-Sep 21 16:49:42.692: INFO: (7) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 54.354811ms)
-Sep 21 16:49:42.734: INFO: (8) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 41.380802ms)
-Sep 21 16:49:42.768: INFO: (9) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 34.00254ms)
-Sep 21 16:49:42.811: INFO: (10) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 43.234924ms)
-Sep 21 16:49:42.859: INFO: (11) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 47.718636ms)
-Sep 21 16:49:42.907: INFO: (12) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 47.521852ms)
-Sep 21 16:49:42.957: INFO: (13) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 50.262283ms)
-Sep 21 16:49:43.002: INFO: (14) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 45.22224ms)
-Sep 21 16:49:43.080: INFO: (15) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 77.621855ms)
-Sep 21 16:49:43.133: INFO: (16) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 52.827879ms)
-Sep 21 16:49:43.186: INFO: (17) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 53.232318ms)
-Sep 21 16:49:43.227: INFO: (18) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 40.899346ms)
-Sep 21 16:49:43.265: INFO: (19) /api/v1/nodes/10.241.51.149:10250/proxy/logs/: 
-anaconda/
-at/
-at-no-rotat... (200; 36.872566ms)
-[AfterEach] version v1
+STEP: Registering a webhook that server cannot talk to, with fail closed policy, via the AdmissionRegistration API
+STEP: create a namespace for the webhook
+STEP: create a configmap should be unconditionally rejected by the webhook
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:49:43.265: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "proxy-5972" for this suite.
-•{"msg":"PASSED [sig-network] Proxy version v1 should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]","total":280,"completed":210,"skipped":3473,"failed":0}
-SSSSSSSSSSSSSSS
+Oct 26 17:02:51.982: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-2557" for this suite.
+STEP: Destroying namespace "webhook-2557-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+
+• [SLOW TEST:6.684 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should unconditionally reject operations on fail closed webhook [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]","total":280,"completed":212,"skipped":3297,"failed":0}
+S
 ------------------------------
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should deny crd creation [Conformance]
+  patching/updating a validating webhook should work [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:49:43.335: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:02:52.256: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
@@ -10923,39 +9945,74 @@ STEP: Setting up server cert
 STEP: Create role binding to let webhook read extension-apiserver-authentication
 STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Sep 21 16:49:44.840: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303784, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303784, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303784, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303784, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:02:53.511: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328573, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328573, loc:(*time.Location)(0x7db5bc0)}}, Reason:"NewReplicaSetCreated", Message:"Created new replica set \"sample-webhook-deployment-5f65f8c764\""}, v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328573, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328573, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:02:55.529: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328573, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328573, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328573, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328573, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Sep 21 16:49:47.944: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should deny crd creation [Conformance]
+Oct 26 17:02:58.556: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] patching/updating a validating webhook should work [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a validating webhook configuration
+STEP: Creating a configMap that does not comply to the validation webhook rules
+STEP: Updating a validating webhook configuration's rules to not include the create operation
+STEP: Creating a configMap that does not comply to the validation webhook rules
+STEP: Patching a validating webhook configuration's rules to include the create operation
+STEP: Creating a configMap that does not comply to the validation webhook rules
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:02:58.860: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-9000" for this suite.
+STEP: Destroying namespace "webhook-9000-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+
+• [SLOW TEST:6.859 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  patching/updating a validating webhook should work [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance]","total":280,"completed":213,"skipped":3298,"failed":0}
+SSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  updates the published spec when one version gets renamed [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:02:59.116: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename crd-publish-openapi
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] updates the published spec when one version gets renamed [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering the crd webhook via the AdmissionRegistration API
-STEP: Creating a custom resource definition that should be denied by the webhook
-Sep 21 16:49:48.070: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+STEP: set up a multi version CRD
+Oct 26 17:02:59.302: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: rename a version
+STEP: check the new version name is served
+STEP: check the old version name is removed
+STEP: check the other version is not changed
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:49:48.147: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-3949" for this suite.
-STEP: Destroying namespace "webhook-3949-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+Oct 26 17:03:44.828: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-662" for this suite.
 
-• [SLOW TEST:5.537 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+• [SLOW TEST:45.761 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should deny crd creation [Conformance]
+  updates the published spec when one version gets renamed [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance]","total":280,"completed":211,"skipped":3488,"failed":0}
-SSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance]","total":280,"completed":214,"skipped":3310,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should be able to deny custom resource creation, update and deletion [Conformance]
+  should mutate custom resource with different stored version [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:49:48.875: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:03:44.878: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
@@ -10964,569 +10021,765 @@ STEP: Setting up server cert
 STEP: Create role binding to let webhook read extension-apiserver-authentication
 STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Sep 21 16:49:50.156: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Sep 21 16:49:52.219: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303790, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303790, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303790, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303790, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:03:45.755: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 17:03:47.804: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328625, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328625, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328625, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328625, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Sep 21 16:49:55.316: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should be able to deny custom resource creation, update and deletion [Conformance]
+Oct 26 17:03:50.871: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should mutate custom resource with different stored version [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:49:55.343: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Registering the custom resource webhook via the AdmissionRegistration API
-STEP: Creating a custom resource that should be denied by the webhook
-STEP: Creating a custom resource whose deletion would be denied by the webhook
-STEP: Updating the custom resource with disallowed data should be denied
-STEP: Deleting the custom resource should be denied
-STEP: Remove the offending key and value from the custom resource data
-STEP: Deleting the updated custom resource should be successful
+Oct 26 17:03:50.889: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Registering the mutating webhook for custom resource e2e-test-webhook-9520-crds.webhook.example.com via the AdmissionRegistration API
+STEP: Creating a custom resource while v1 is storage version
+STEP: Patching Custom Resource Definition to set v2 as storage
+STEP: Patching the custom resource while v2 is storage version
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:49:57.110: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-7797" for this suite.
-STEP: Destroying namespace "webhook-7797-markers" for this suite.
+Oct 26 17:03:52.462: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-9898" for this suite.
+STEP: Destroying namespace "webhook-9898-markers" for this suite.
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:8.617 seconds]
+• [SLOW TEST:7.834 seconds]
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should be able to deny custom resource creation, update and deletion [Conformance]
+  should mutate custom resource with different stored version [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance]","total":280,"completed":212,"skipped":3500,"failed":0}
-SSS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance]","total":280,"completed":215,"skipped":3348,"failed":0}
+SSSSSSSSSSSSSSS
 ------------------------------
-[sig-node] Downward API 
-  should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
+[sig-storage] ConfigMap 
+  binary data should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] Downward API
+[BeforeEach] [sig-storage] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:49:57.493: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
+Oct 26 17:03:52.715: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename configmap
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
+[It] binary data should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward api env vars
-Sep 21 16:49:57.822: INFO: Waiting up to 5m0s for pod "downward-api-035fdb3f-aeb7-48ad-8557-7b6dfd35bb9a" in namespace "downward-api-2842" to be "success or failure"
-Sep 21 16:49:57.843: INFO: Pod "downward-api-035fdb3f-aeb7-48ad-8557-7b6dfd35bb9a": Phase="Pending", Reason="", readiness=false. Elapsed: 20.615386ms
-Sep 21 16:49:59.868: INFO: Pod "downward-api-035fdb3f-aeb7-48ad-8557-7b6dfd35bb9a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.046342333s
-STEP: Saw pod success
-Sep 21 16:49:59.869: INFO: Pod "downward-api-035fdb3f-aeb7-48ad-8557-7b6dfd35bb9a" satisfied condition "success or failure"
-Sep 21 16:49:59.892: INFO: Trying to get logs from node 10.241.51.147 pod downward-api-035fdb3f-aeb7-48ad-8557-7b6dfd35bb9a container dapi-container: 
-STEP: delete the pod
-Sep 21 16:50:00.017: INFO: Waiting for pod downward-api-035fdb3f-aeb7-48ad-8557-7b6dfd35bb9a to disappear
-Sep 21 16:50:00.038: INFO: Pod downward-api-035fdb3f-aeb7-48ad-8557-7b6dfd35bb9a no longer exists
-[AfterEach] [sig-node] Downward API
+Oct 26 17:03:52.902: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
+STEP: Creating configMap with name configmap-test-upd-f8e6fb79-6626-4af4-a49a-4927ccea9a23
+STEP: Creating the pod
+STEP: Waiting for pod with text data
+STEP: Waiting for pod with binary data
+[AfterEach] [sig-storage] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:50:00.039: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-2842" for this suite.
-•{"msg":"PASSED [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]","total":280,"completed":213,"skipped":3503,"failed":0}
-SSSSSSS
+Oct 26 17:03:57.095: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-7517" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":216,"skipped":3363,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Container Runtime blackbox test on terminated container 
-  should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]
+[sig-apps] Daemon set [Serial] 
+  should rollback without unnecessary restarts [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Runtime
+[BeforeEach] [sig-apps] Daemon set [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:50:00.131: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-runtime
+Oct 26 17:03:57.133: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename daemonsets
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
+[It] should rollback without unnecessary restarts [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the container
-STEP: wait for the container to reach Succeeded
-STEP: get the container status
-STEP: the container should be terminated
-STEP: the termination message should be set
-Sep 21 16:50:02.584: INFO: Expected: &{DONE} to match Container's Termination Message: DONE --
-STEP: delete the container
-[AfterEach] [k8s.io] Container Runtime
+Oct 26 17:03:57.406: INFO: Create a RollingUpdate DaemonSet
+Oct 26 17:03:57.444: INFO: Check that daemon pods launch on every node of the cluster
+Oct 26 17:03:57.477: INFO: Number of nodes with available pods: 0
+Oct 26 17:03:57.477: INFO: Node 10.72.119.74 is running more than one daemon pod
+Oct 26 17:03:58.515: INFO: Number of nodes with available pods: 0
+Oct 26 17:03:58.515: INFO: Node 10.72.119.74 is running more than one daemon pod
+Oct 26 17:03:59.513: INFO: Number of nodes with available pods: 1
+Oct 26 17:03:59.513: INFO: Node 10.72.119.76 is running more than one daemon pod
+Oct 26 17:04:00.519: INFO: Number of nodes with available pods: 3
+Oct 26 17:04:00.519: INFO: Number of running nodes: 3, number of available pods: 3
+Oct 26 17:04:00.519: INFO: Update the DaemonSet to trigger a rollout
+Oct 26 17:04:00.616: INFO: Updating DaemonSet daemon-set
+Oct 26 17:04:07.691: INFO: Roll back the DaemonSet before rollout is complete
+Oct 26 17:04:07.728: INFO: Updating DaemonSet daemon-set
+Oct 26 17:04:07.728: INFO: Make sure DaemonSet rollback is complete
+Oct 26 17:04:07.754: INFO: Wrong image for pod: daemon-set-g4fj4. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent.
+Oct 26 17:04:07.754: INFO: Pod daemon-set-g4fj4 is not available
+Oct 26 17:04:08.799: INFO: Wrong image for pod: daemon-set-g4fj4. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent.
+Oct 26 17:04:08.799: INFO: Pod daemon-set-g4fj4 is not available
+Oct 26 17:04:09.801: INFO: Pod daemon-set-ztpgp is not available
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6098, will wait for the garbage collector to delete the pods
+Oct 26 17:04:09.963: INFO: Deleting DaemonSet.extensions daemon-set took: 38.556079ms
+Oct 26 17:04:10.263: INFO: Terminating DaemonSet.extensions daemon-set pods took: 300.418679ms
+Oct 26 17:04:19.581: INFO: Number of nodes with available pods: 0
+Oct 26 17:04:19.581: INFO: Number of running nodes: 0, number of available pods: 0
+Oct 26 17:04:19.599: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-6098/daemonsets","resourceVersion":"83473"},"items":null}
+
+Oct 26 17:04:19.614: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-6098/pods","resourceVersion":"83473"},"items":null}
+
+[AfterEach] [sig-apps] Daemon set [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:50:02.694: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-runtime-2961" for this suite.
-•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]","total":280,"completed":214,"skipped":3510,"failed":0}
-SSSSSSSSSS
+Oct 26 17:04:19.674: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "daemonsets-6098" for this suite.
+
+• [SLOW TEST:22.579 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should rollback without unnecessary restarts [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should update annotations on modification [NodeConformance] [Conformance]
+{"msg":"PASSED [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance]","total":280,"completed":217,"skipped":3410,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Kubelet when scheduling a busybox Pod with hostAliases 
+  should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
+[BeforeEach] [k8s.io] Kubelet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:50:02.773: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 17:04:19.713: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubelet-test
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should update annotations on modification [NodeConformance] [Conformance]
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating the pod
-Sep 21 16:50:07.946: INFO: Successfully updated pod "annotationupdatedabe6c21-d913-4c84-856d-fb673a4fc1e5"
-[AfterEach] [sig-storage] Projected downwardAPI
+[AfterEach] [k8s.io] Kubelet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:50:10.281: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-6113" for this suite.
-
-• [SLOW TEST:7.573 seconds]
-[sig-storage] Projected downwardAPI
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:34
-  should update annotations on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance]","total":280,"completed":215,"skipped":3520,"failed":0}
-SSSSSSSSSS
+Oct 26 17:04:22.010: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-6749" for this suite.
+•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox Pod with hostAliases should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":218,"skipped":3442,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Secrets 
-  should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute poststart exec hook properly [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
+[BeforeEach] [k8s.io] Container Lifecycle Hook
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:50:10.346: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename secrets
+Oct 26 17:04:22.055: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename container-lifecycle-hook
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute poststart exec hook properly [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-61449247-c627-4b69-b0fe-2fba6cf82ee5
-STEP: Creating a pod to test consume secrets
-Sep 21 16:50:11.040: INFO: Waiting up to 5m0s for pod "pod-secrets-14ca5665-81bf-4479-b4ec-022dbd078b58" in namespace "secrets-7418" to be "success or failure"
-Sep 21 16:50:11.061: INFO: Pod "pod-secrets-14ca5665-81bf-4479-b4ec-022dbd078b58": Phase="Pending", Reason="", readiness=false. Elapsed: 20.788817ms
-Sep 21 16:50:13.083: INFO: Pod "pod-secrets-14ca5665-81bf-4479-b4ec-022dbd078b58": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.042863242s
-STEP: Saw pod success
-Sep 21 16:50:13.083: INFO: Pod "pod-secrets-14ca5665-81bf-4479-b4ec-022dbd078b58" satisfied condition "success or failure"
-Sep 21 16:50:13.105: INFO: Trying to get logs from node 10.241.51.147 pod pod-secrets-14ca5665-81bf-4479-b4ec-022dbd078b58 container secret-volume-test: 
-STEP: delete the pod
-Sep 21 16:50:13.252: INFO: Waiting for pod pod-secrets-14ca5665-81bf-4479-b4ec-022dbd078b58 to disappear
-Sep 21 16:50:13.274: INFO: Pod pod-secrets-14ca5665-81bf-4479-b4ec-022dbd078b58 no longer exists
-[AfterEach] [sig-storage] Secrets
+STEP: create the pod with lifecycle hook
+STEP: check poststart hook
+STEP: delete the pod with lifecycle hook
+Oct 26 17:04:30.473: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Oct 26 17:04:30.488: INFO: Pod pod-with-poststart-exec-hook still exists
+Oct 26 17:04:32.489: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Oct 26 17:04:32.504: INFO: Pod pod-with-poststart-exec-hook still exists
+Oct 26 17:04:34.489: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Oct 26 17:04:34.507: INFO: Pod pod-with-poststart-exec-hook still exists
+Oct 26 17:04:36.489: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Oct 26 17:04:36.505: INFO: Pod pod-with-poststart-exec-hook still exists
+Oct 26 17:04:38.489: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Oct 26 17:04:38.506: INFO: Pod pod-with-poststart-exec-hook no longer exists
+[AfterEach] [k8s.io] Container Lifecycle Hook
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:50:13.274: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-7418" for this suite.
-STEP: Destroying namespace "secret-namespace-2774" for this suite.
-•{"msg":"PASSED [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]","total":280,"completed":216,"skipped":3530,"failed":0}
-SSSSSSSS
+Oct 26 17:04:38.506: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-3393" for this suite.
+
+• [SLOW TEST:16.489 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  when create a pod with lifecycle hook
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
+    should execute poststart exec hook properly [NodeConformance] [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide container's memory limit [NodeConformance] [Conformance]
+{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance]","total":280,"completed":219,"skipped":3480,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:50:13.361: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 17:04:38.547: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should provide container's memory limit [NodeConformance] [Conformance]
+[It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:50:13.690: INFO: Waiting up to 5m0s for pod "downwardapi-volume-4e723239-4ea4-4002-8960-e4bd54dbc441" in namespace "projected-8938" to be "success or failure"
-Sep 21 16:50:13.715: INFO: Pod "downwardapi-volume-4e723239-4ea4-4002-8960-e4bd54dbc441": Phase="Pending", Reason="", readiness=false. Elapsed: 25.489887ms
-Sep 21 16:50:15.738: INFO: Pod "downwardapi-volume-4e723239-4ea4-4002-8960-e4bd54dbc441": Phase="Pending", Reason="", readiness=false. Elapsed: 2.048241583s
-Sep 21 16:50:17.765: INFO: Pod "downwardapi-volume-4e723239-4ea4-4002-8960-e4bd54dbc441": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.075384337s
+STEP: Creating a pod to test emptydir 0644 on tmpfs
+Oct 26 17:04:38.743: INFO: Waiting up to 5m0s for pod "pod-f09e618a-9e02-4e38-8fc3-e95e731f3171" in namespace "emptydir-4437" to be "success or failure"
+Oct 26 17:04:38.759: INFO: Pod "pod-f09e618a-9e02-4e38-8fc3-e95e731f3171": Phase="Pending", Reason="", readiness=false. Elapsed: 15.181852ms
+Oct 26 17:04:40.775: INFO: Pod "pod-f09e618a-9e02-4e38-8fc3-e95e731f3171": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.031502356s
 STEP: Saw pod success
-Sep 21 16:50:17.765: INFO: Pod "downwardapi-volume-4e723239-4ea4-4002-8960-e4bd54dbc441" satisfied condition "success or failure"
-Sep 21 16:50:17.787: INFO: Trying to get logs from node 10.241.51.150 pod downwardapi-volume-4e723239-4ea4-4002-8960-e4bd54dbc441 container client-container: 
+Oct 26 17:04:40.775: INFO: Pod "pod-f09e618a-9e02-4e38-8fc3-e95e731f3171" satisfied condition "success or failure"
+Oct 26 17:04:40.791: INFO: Trying to get logs from node 10.72.119.74 pod pod-f09e618a-9e02-4e38-8fc3-e95e731f3171 container test-container: 
 STEP: delete the pod
-Sep 21 16:50:17.974: INFO: Waiting for pod downwardapi-volume-4e723239-4ea4-4002-8960-e4bd54dbc441 to disappear
-Sep 21 16:50:17.996: INFO: Pod downwardapi-volume-4e723239-4ea4-4002-8960-e4bd54dbc441 no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
+Oct 26 17:04:40.870: INFO: Waiting for pod pod-f09e618a-9e02-4e38-8fc3-e95e731f3171 to disappear
+Oct 26 17:04:40.888: INFO: Pod pod-f09e618a-9e02-4e38-8fc3-e95e731f3171 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:50:17.996: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-8938" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance]","total":280,"completed":217,"skipped":3538,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 17:04:40.888: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-4437" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":220,"skipped":3558,"failed":0}
+SSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl run --rm job 
-  should create a job from an image, then delete the job  [Conformance]
+[k8s.io] InitContainer [NodeConformance] 
+  should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:50:18.058: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubectl
+Oct 26 17:04:40.931: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename init-container
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[It] should create a job from an image, then delete the job  [Conformance]
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
+[It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: executing a command with run --rm and attach with stdin
-Sep 21 16:50:18.275: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=kubectl-1324 run e2e-test-rm-busybox-job --image=docker.io/library/busybox:1.29 --rm=true --generator=job/v1 --restart=OnFailure --attach=true --stdin -- sh -c cat && echo 'stdin closed''
-Sep 21 16:50:21.013: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\nIf you don't see a command prompt, try pressing enter.\n"
-Sep 21 16:50:21.013: INFO: stdout: "abcd1234stdin closed\njob.batch \"e2e-test-rm-busybox-job\" deleted\n"
-STEP: verifying the job e2e-test-rm-busybox-job was deleted
-[AfterEach] [sig-cli] Kubectl client
+STEP: creating the pod
+Oct 26 17:04:41.073: INFO: PodSpec: initContainers in spec.initContainers
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:50:23.077: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-1324" for this suite.
-
-• [SLOW TEST:5.085 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Kubectl run --rm job
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1944
-    should create a job from an image, then delete the job  [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Oct 26 17:04:44.534: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "init-container-8932" for this suite.
+•{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]","total":280,"completed":221,"skipped":3561,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Kubectl run --rm job should create a job from an image, then delete the job  [Conformance]","total":280,"completed":218,"skipped":3594,"failed":0}
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]
+[sig-api-machinery] ResourceQuota 
+  should verify ResourceQuota with terminating scopes. [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+[BeforeEach] [sig-api-machinery] ResourceQuota
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:50:23.143: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename webhook
+Oct 26 17:04:44.580: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename resourcequota
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Sep 21 16:50:24.273: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Sep 21 16:50:26.349: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303824, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303824, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303824, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303824, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Sep 21 16:50:29.433: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]
+[It] should verify ResourceQuota with terminating scopes. [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering a validating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API
-STEP: Registering a mutating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API
-STEP: Creating a dummy validating-webhook-configuration object
-STEP: Deleting the validating-webhook-configuration, which should be possible to remove
-STEP: Creating a dummy mutating-webhook-configuration object
-STEP: Deleting the mutating-webhook-configuration, which should be possible to remove
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+STEP: Creating a ResourceQuota with terminating scope
+STEP: Ensuring ResourceQuota status is calculated
+STEP: Creating a ResourceQuota with not terminating scope
+STEP: Ensuring ResourceQuota status is calculated
+STEP: Creating a long running pod
+STEP: Ensuring resource quota with not terminating scope captures the pod usage
+STEP: Ensuring resource quota with terminating scope ignored the pod usage
+STEP: Deleting the pod
+STEP: Ensuring resource quota status released the pod usage
+STEP: Creating a terminating pod
+STEP: Ensuring resource quota with terminating scope captures the pod usage
+STEP: Ensuring resource quota with not terminating scope ignored the pod usage
+STEP: Deleting the pod
+STEP: Ensuring resource quota status released the pod usage
+[AfterEach] [sig-api-machinery] ResourceQuota
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:50:30.174: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-2321" for this suite.
-STEP: Destroying namespace "webhook-2321-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+Oct 26 17:05:01.153: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-6924" for this suite.
 
-• [SLOW TEST:7.475 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+• [SLOW TEST:16.605 seconds]
+[sig-api-machinery] ResourceQuota
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]
+  should verify ResourceQuota with terminating scopes. [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]","total":280,"completed":219,"skipped":3594,"failed":0}
-S
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance]","total":280,"completed":222,"skipped":3586,"failed":0}
+SSSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide podname only [NodeConformance] [Conformance]
+[sig-network] Services 
+  should be able to change the type from ExternalName to NodePort [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
+[BeforeEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:50:30.618: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 17:05:01.185: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename services
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should provide podname only [NodeConformance] [Conformance]
+[BeforeEach] [sig-network] Services
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
+[It] should be able to change the type from ExternalName to NodePort [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:50:30.946: INFO: Waiting up to 5m0s for pod "downwardapi-volume-c073f2bb-860c-482a-aae9-95545b9e72df" in namespace "projected-4138" to be "success or failure"
-Sep 21 16:50:30.970: INFO: Pod "downwardapi-volume-c073f2bb-860c-482a-aae9-95545b9e72df": Phase="Pending", Reason="", readiness=false. Elapsed: 24.01725ms
-Sep 21 16:50:32.996: INFO: Pod "downwardapi-volume-c073f2bb-860c-482a-aae9-95545b9e72df": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.050599134s
-STEP: Saw pod success
-Sep 21 16:50:32.997: INFO: Pod "downwardapi-volume-c073f2bb-860c-482a-aae9-95545b9e72df" satisfied condition "success or failure"
-Sep 21 16:50:33.020: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-c073f2bb-860c-482a-aae9-95545b9e72df container client-container: 
-STEP: delete the pod
-Sep 21 16:50:33.175: INFO: Waiting for pod downwardapi-volume-c073f2bb-860c-482a-aae9-95545b9e72df to disappear
-Sep 21 16:50:33.413: INFO: Pod downwardapi-volume-c073f2bb-860c-482a-aae9-95545b9e72df no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
+STEP: creating a service externalname-service with the type=ExternalName in namespace services-4557
+STEP: changing the ExternalName service to type=NodePort
+STEP: creating replication controller externalname-service in namespace services-4557
+I1026 17:05:01.478808      26 runners.go:189] Created replication controller with name: externalname-service, namespace: services-4557, replica count: 2
+Oct 26 17:05:04.529: INFO: Creating new exec pod
+I1026 17:05:04.529572      26 runners.go:189] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+Oct 26 17:05:09.640: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-4557 execpodjg85x -- /bin/sh -x -c nc -zv -t -w 2 externalname-service 80'
+Oct 26 17:05:10.047: INFO: stderr: "+ nc -zv -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n"
+Oct 26 17:05:10.047: INFO: stdout: ""
+Oct 26 17:05:10.051: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-4557 execpodjg85x -- /bin/sh -x -c nc -zv -t -w 2 172.21.57.227 80'
+Oct 26 17:05:10.557: INFO: stderr: "+ nc -zv -t -w 2 172.21.57.227 80\nConnection to 172.21.57.227 80 port [tcp/http] succeeded!\n"
+Oct 26 17:05:10.557: INFO: stdout: ""
+Oct 26 17:05:10.557: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-4557 execpodjg85x -- /bin/sh -x -c nc -zv -t -w 2 10.72.119.76 32430'
+Oct 26 17:05:11.322: INFO: stderr: "+ nc -zv -t -w 2 10.72.119.76 32430\nConnection to 10.72.119.76 32430 port [tcp/32430] succeeded!\n"
+Oct 26 17:05:11.322: INFO: stdout: ""
+Oct 26 17:05:11.322: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-4557 execpodjg85x -- /bin/sh -x -c nc -zv -t -w 2 10.72.119.82 32430'
+Oct 26 17:05:11.825: INFO: stderr: "+ nc -zv -t -w 2 10.72.119.82 32430\nConnection to 10.72.119.82 32430 port [tcp/32430] succeeded!\n"
+Oct 26 17:05:11.825: INFO: stdout: ""
+Oct 26 17:05:11.825: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-4557 execpodjg85x -- /bin/sh -x -c nc -zv -t -w 2 158.176.65.138 32430'
+Oct 26 17:05:12.265: INFO: stderr: "+ nc -zv -t -w 2 158.176.65.138 32430\nConnection to 158.176.65.138 32430 port [tcp/32430] succeeded!\n"
+Oct 26 17:05:12.265: INFO: stdout: ""
+Oct 26 17:05:12.265: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-4557 execpodjg85x -- /bin/sh -x -c nc -zv -t -w 2 158.176.65.141 32430'
+Oct 26 17:05:12.654: INFO: stderr: "+ nc -zv -t -w 2 158.176.65.141 32430\nConnection to 158.176.65.141 32430 port [tcp/32430] succeeded!\n"
+Oct 26 17:05:12.654: INFO: stdout: ""
+Oct 26 17:05:12.654: INFO: Cleaning up the ExternalName to NodePort test service
+[AfterEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:50:33.413: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-4138" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance]","total":280,"completed":220,"skipped":3595,"failed":0}
+Oct 26 17:05:12.746: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-4557" for this suite.
+[AfterEach] [sig-network] Services
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
 
+• [SLOW TEST:11.602 seconds]
+[sig-network] Services
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should be able to change the type from ExternalName to NodePort [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-network] DNS 
-  should provide DNS for the cluster  [Conformance]
+{"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance]","total":280,"completed":223,"skipped":3590,"failed":0}
+SS
+------------------------------
+[sig-cli] Kubectl client Update Demo 
+  should scale a replication controller  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] DNS
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:50:33.487: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename dns
+Oct 26 17:05:12.787: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubectl
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide DNS for the cluster  [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
+[BeforeEach] Update Demo
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:330
+[It] should scale a replication controller  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-9763.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
+STEP: creating a replication controller
+Oct 26 17:05:12.936: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-1112'
+Oct 26 17:05:13.565: INFO: stderr: ""
+Oct 26 17:05:13.566: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Oct 26 17:05:13.566: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1112'
+Oct 26 17:05:13.740: INFO: stderr: ""
+Oct 26 17:05:13.740: INFO: stdout: "update-demo-nautilus-79sh2 update-demo-nautilus-99lrh "
+Oct 26 17:05:13.740: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-79sh2 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:13.894: INFO: stderr: ""
+Oct 26 17:05:13.894: INFO: stdout: ""
+Oct 26 17:05:13.894: INFO: update-demo-nautilus-79sh2 is created but not running
+Oct 26 17:05:18.894: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1112'
+Oct 26 17:05:19.044: INFO: stderr: ""
+Oct 26 17:05:19.044: INFO: stdout: "update-demo-nautilus-79sh2 update-demo-nautilus-99lrh "
+Oct 26 17:05:19.044: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-79sh2 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:19.203: INFO: stderr: ""
+Oct 26 17:05:19.203: INFO: stdout: "true"
+Oct 26 17:05:19.203: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-79sh2 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:19.343: INFO: stderr: ""
+Oct 26 17:05:19.343: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Oct 26 17:05:19.343: INFO: validating pod update-demo-nautilus-79sh2
+Oct 26 17:05:19.382: INFO: got data: {
+  "image": "nautilus.jpg"
+}
 
-STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-9763.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+Oct 26 17:05:19.382: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Oct 26 17:05:19.382: INFO: update-demo-nautilus-79sh2 is verified up and running
+Oct 26 17:05:19.382: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-99lrh -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:19.555: INFO: stderr: ""
+Oct 26 17:05:19.555: INFO: stdout: "true"
+Oct 26 17:05:19.555: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-99lrh -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:19.705: INFO: stderr: ""
+Oct 26 17:05:19.705: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Oct 26 17:05:19.705: INFO: validating pod update-demo-nautilus-99lrh
+Oct 26 17:05:19.738: INFO: got data: {
+  "image": "nautilus.jpg"
+}
 
-STEP: creating a pod to probe DNS
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Sep 21 16:50:38.632: INFO: DNS probes using dns-9763/dns-test-503a81a4-d850-40cd-97f6-9dbfc12e1e9c succeeded
+Oct 26 17:05:19.738: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Oct 26 17:05:19.738: INFO: update-demo-nautilus-99lrh is verified up and running
+STEP: scaling down the replication controller
+Oct 26 17:05:19.744: INFO: scanned /root for discovery docs: 
+Oct 26 17:05:19.744: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=kubectl-1112'
+Oct 26 17:05:20.997: INFO: stderr: ""
+Oct 26 17:05:20.997: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Oct 26 17:05:20.997: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1112'
+Oct 26 17:05:21.149: INFO: stderr: ""
+Oct 26 17:05:21.149: INFO: stdout: "update-demo-nautilus-79sh2 update-demo-nautilus-99lrh "
+STEP: Replicas for name=update-demo: expected=1 actual=2
+Oct 26 17:05:26.149: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1112'
+Oct 26 17:05:26.300: INFO: stderr: ""
+Oct 26 17:05:26.300: INFO: stdout: "update-demo-nautilus-79sh2 update-demo-nautilus-99lrh "
+STEP: Replicas for name=update-demo: expected=1 actual=2
+Oct 26 17:05:31.300: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1112'
+Oct 26 17:05:31.451: INFO: stderr: ""
+Oct 26 17:05:31.451: INFO: stdout: "update-demo-nautilus-99lrh "
+Oct 26 17:05:31.451: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-99lrh -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:31.607: INFO: stderr: ""
+Oct 26 17:05:31.607: INFO: stdout: "true"
+Oct 26 17:05:31.607: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-99lrh -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:31.776: INFO: stderr: ""
+Oct 26 17:05:31.776: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Oct 26 17:05:31.776: INFO: validating pod update-demo-nautilus-99lrh
+Oct 26 17:05:31.808: INFO: got data: {
+  "image": "nautilus.jpg"
+}
 
-STEP: deleting the pod
-[AfterEach] [sig-network] DNS
+Oct 26 17:05:31.808: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Oct 26 17:05:31.808: INFO: update-demo-nautilus-99lrh is verified up and running
+STEP: scaling up the replication controller
+Oct 26 17:05:31.812: INFO: scanned /root for discovery docs: 
+Oct 26 17:05:31.812: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=kubectl-1112'
+Oct 26 17:05:33.034: INFO: stderr: ""
+Oct 26 17:05:33.034: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Oct 26 17:05:33.034: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1112'
+Oct 26 17:05:33.181: INFO: stderr: ""
+Oct 26 17:05:33.181: INFO: stdout: "update-demo-nautilus-99lrh update-demo-nautilus-wx6qh "
+Oct 26 17:05:33.181: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-99lrh -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:33.333: INFO: stderr: ""
+Oct 26 17:05:33.333: INFO: stdout: "true"
+Oct 26 17:05:33.333: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-99lrh -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:33.482: INFO: stderr: ""
+Oct 26 17:05:33.482: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Oct 26 17:05:33.482: INFO: validating pod update-demo-nautilus-99lrh
+Oct 26 17:05:33.507: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Oct 26 17:05:33.507: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Oct 26 17:05:33.507: INFO: update-demo-nautilus-99lrh is verified up and running
+Oct 26 17:05:33.507: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-wx6qh -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:33.658: INFO: stderr: ""
+Oct 26 17:05:33.658: INFO: stdout: ""
+Oct 26 17:05:33.658: INFO: update-demo-nautilus-wx6qh is created but not running
+Oct 26 17:05:38.659: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1112'
+Oct 26 17:05:38.806: INFO: stderr: ""
+Oct 26 17:05:38.806: INFO: stdout: "update-demo-nautilus-99lrh update-demo-nautilus-wx6qh "
+Oct 26 17:05:38.806: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-99lrh -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:38.950: INFO: stderr: ""
+Oct 26 17:05:38.950: INFO: stdout: "true"
+Oct 26 17:05:38.950: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-99lrh -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:39.087: INFO: stderr: ""
+Oct 26 17:05:39.087: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Oct 26 17:05:39.087: INFO: validating pod update-demo-nautilus-99lrh
+Oct 26 17:05:39.112: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Oct 26 17:05:39.112: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Oct 26 17:05:39.112: INFO: update-demo-nautilus-99lrh is verified up and running
+Oct 26 17:05:39.112: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-wx6qh -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:39.253: INFO: stderr: ""
+Oct 26 17:05:39.253: INFO: stdout: "true"
+Oct 26 17:05:39.253: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-wx6qh -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1112'
+Oct 26 17:05:39.412: INFO: stderr: ""
+Oct 26 17:05:39.412: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Oct 26 17:05:39.412: INFO: validating pod update-demo-nautilus-wx6qh
+Oct 26 17:05:39.445: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Oct 26 17:05:39.445: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Oct 26 17:05:39.445: INFO: update-demo-nautilus-wx6qh is verified up and running
+STEP: using delete to clean up resources
+Oct 26 17:05:39.445: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 delete --grace-period=0 --force -f - --namespace=kubectl-1112'
+Oct 26 17:05:39.631: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Oct 26 17:05:39.631: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n"
+Oct 26 17:05:39.631: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-1112'
+Oct 26 17:05:39.831: INFO: stderr: "No resources found in kubectl-1112 namespace.\n"
+Oct 26 17:05:39.831: INFO: stdout: ""
+Oct 26 17:05:39.831: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -l name=update-demo --namespace=kubectl-1112 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Oct 26 17:05:39.985: INFO: stderr: ""
+Oct 26 17:05:39.985: INFO: stdout: ""
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:50:38.762: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-9763" for this suite.
+Oct 26 17:05:39.985: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-1112" for this suite.
 
-• [SLOW TEST:5.347 seconds]
-[sig-network] DNS
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should provide DNS for the cluster  [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:27.239 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Update Demo
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:328
+    should scale a replication controller  [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] DNS should provide DNS for the cluster  [Conformance]","total":280,"completed":221,"skipped":3595,"failed":0}
+{"msg":"PASSED [sig-cli] Kubectl client Update Demo should scale a replication controller  [Conformance]","total":280,"completed":224,"skipped":3592,"failed":0}
 SSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-node] ConfigMap 
-  should be consumable via the environment [NodeConformance] [Conformance]
+  should fail to create ConfigMap with empty key [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-node] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:50:38.836: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:05:40.026: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename configmap
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable via the environment [NodeConformance] [Conformance]
+[It] should fail to create ConfigMap with empty key [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap configmap-422/configmap-test-4c661ca2-5028-4b59-a6b0-210e5a1ad52f
-STEP: Creating a pod to test consume configMaps
-Sep 21 16:50:39.163: INFO: Waiting up to 5m0s for pod "pod-configmaps-c5da4972-c23f-4cf7-8110-a435ab2d82c1" in namespace "configmap-422" to be "success or failure"
-Sep 21 16:50:39.195: INFO: Pod "pod-configmaps-c5da4972-c23f-4cf7-8110-a435ab2d82c1": Phase="Pending", Reason="", readiness=false. Elapsed: 32.033275ms
-Sep 21 16:50:41.227: INFO: Pod "pod-configmaps-c5da4972-c23f-4cf7-8110-a435ab2d82c1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064143096s
-Sep 21 16:50:43.253: INFO: Pod "pod-configmaps-c5da4972-c23f-4cf7-8110-a435ab2d82c1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.089977142s
-STEP: Saw pod success
-Sep 21 16:50:43.253: INFO: Pod "pod-configmaps-c5da4972-c23f-4cf7-8110-a435ab2d82c1" satisfied condition "success or failure"
-Sep 21 16:50:43.273: INFO: Trying to get logs from node 10.241.51.147 pod pod-configmaps-c5da4972-c23f-4cf7-8110-a435ab2d82c1 container env-test: 
-STEP: delete the pod
-Sep 21 16:50:43.693: INFO: Waiting for pod pod-configmaps-c5da4972-c23f-4cf7-8110-a435ab2d82c1 to disappear
-Sep 21 16:50:43.719: INFO: Pod pod-configmaps-c5da4972-c23f-4cf7-8110-a435ab2d82c1 no longer exists
+STEP: Creating configMap that has name configmap-test-emptyKey-e1cfdb4c-3fb5-461a-b159-bb1face06ab1
 [AfterEach] [sig-node] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:50:43.720: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-422" for this suite.
-•{"msg":"PASSED [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance]","total":280,"completed":222,"skipped":3612,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
-  should execute prestop exec hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:50:43.790: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-lifecycle-hook
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] when create a pod with lifecycle hook
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
-STEP: create the container to handle the HTTPGet hook request.
-[It] should execute prestop exec hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the pod with lifecycle hook
-STEP: delete the pod with lifecycle hook
-Sep 21 16:50:52.414: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Sep 21 16:50:52.442: INFO: Pod pod-with-prestop-exec-hook still exists
-Sep 21 16:50:54.442: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Sep 21 16:50:54.466: INFO: Pod pod-with-prestop-exec-hook still exists
-Sep 21 16:50:56.442: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Sep 21 16:50:56.463: INFO: Pod pod-with-prestop-exec-hook still exists
-Sep 21 16:50:58.442: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Sep 21 16:50:58.464: INFO: Pod pod-with-prestop-exec-hook no longer exists
-STEP: check prestop hook
-[AfterEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:50:58.521: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-lifecycle-hook-8993" for this suite.
-
-• [SLOW TEST:14.801 seconds]
-[k8s.io] Container Lifecycle Hook
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  when create a pod with lifecycle hook
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
-    should execute prestop exec hook properly [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance]","total":280,"completed":223,"skipped":3659,"failed":0}
-SSSSSSSSSSSSS
+Oct 26 17:05:40.202: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-7649" for this suite.
+•{"msg":"PASSED [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance]","total":280,"completed":225,"skipped":3609,"failed":0}
+SSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition 
-  listing custom resource definition objects works  [Conformance]
+[sig-storage] HostPath 
+  should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+[BeforeEach] [sig-storage] HostPath
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:50:58.591: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename custom-resource-definition
+Oct 26 17:05:40.239: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename hostpath
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] listing custom resource definition objects works  [Conformance]
+[BeforeEach] [sig-storage] HostPath
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37
+[It] should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:50:58.803: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+STEP: Creating a pod to test hostPath mode
+Oct 26 17:05:40.469: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "hostpath-333" to be "success or failure"
+Oct 26 17:05:40.490: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 21.296245ms
+Oct 26 17:05:42.507: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03819682s
+Oct 26 17:05:44.527: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.0583525s
+STEP: Saw pod success
+Oct 26 17:05:44.527: INFO: Pod "pod-host-path-test" satisfied condition "success or failure"
+Oct 26 17:05:44.546: INFO: Trying to get logs from node 10.72.119.74 pod pod-host-path-test container test-container-1: 
+STEP: delete the pod
+Oct 26 17:05:44.656: INFO: Waiting for pod pod-host-path-test to disappear
+Oct 26 17:05:44.675: INFO: Pod pod-host-path-test no longer exists
+[AfterEach] [sig-storage] HostPath
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:51:01.670: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "custom-resource-definition-5318" for this suite.
-•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works  [Conformance]","total":280,"completed":224,"skipped":3672,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 17:05:44.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "hostpath-333" for this suite.
+•{"msg":"PASSED [sig-storage] HostPath should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":226,"skipped":3624,"failed":0}
+SSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+[sig-api-machinery] Garbage collector 
+  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
+[BeforeEach] [sig-api-machinery] Garbage collector
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:51:01.751: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename configmap
+Oct 26 17:05:44.718: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename gc
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+[It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name configmap-test-volume-map-aa06cb7f-6812-422c-b154-397193c2507b
-STEP: Creating a pod to test consume configMaps
-Sep 21 16:51:02.367: INFO: Waiting up to 5m0s for pod "pod-configmaps-499ed98a-f66b-4fbd-81d8-ba9ee980412f" in namespace "configmap-2682" to be "success or failure"
-Sep 21 16:51:02.391: INFO: Pod "pod-configmaps-499ed98a-f66b-4fbd-81d8-ba9ee980412f": Phase="Pending", Reason="", readiness=false. Elapsed: 23.887419ms
-Sep 21 16:51:04.415: INFO: Pod "pod-configmaps-499ed98a-f66b-4fbd-81d8-ba9ee980412f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.047908375s
-Sep 21 16:51:06.438: INFO: Pod "pod-configmaps-499ed98a-f66b-4fbd-81d8-ba9ee980412f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.071250537s
-STEP: Saw pod success
-Sep 21 16:51:06.438: INFO: Pod "pod-configmaps-499ed98a-f66b-4fbd-81d8-ba9ee980412f" satisfied condition "success or failure"
-Sep 21 16:51:06.459: INFO: Trying to get logs from node 10.241.51.147 pod pod-configmaps-499ed98a-f66b-4fbd-81d8-ba9ee980412f container configmap-volume-test: 
-STEP: delete the pod
-Sep 21 16:51:06.668: INFO: Waiting for pod pod-configmaps-499ed98a-f66b-4fbd-81d8-ba9ee980412f to disappear
-Sep 21 16:51:06.690: INFO: Pod pod-configmaps-499ed98a-f66b-4fbd-81d8-ba9ee980412f no longer exists
-[AfterEach] [sig-storage] ConfigMap
+STEP: create the deployment
+STEP: Wait for the Deployment to create new ReplicaSet
+STEP: delete the deployment
+STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the rs
+STEP: Gathering metrics
+Oct 26 17:06:15.595: INFO: For apiserver_request_total:
+For apiserver_request_latency_seconds:
+For apiserver_init_events_total:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
+
+W1026 17:06:15.595602      26 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+[AfterEach] [sig-api-machinery] Garbage collector
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:51:06.690: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-2682" for this suite.
+Oct 26 17:06:15.595: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-8626" for this suite.
 
-• [SLOW TEST:5.017 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+• [SLOW TEST:30.918 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":225,"skipped":3698,"failed":0}
-S
+{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]","total":280,"completed":227,"skipped":3628,"failed":0}
+SSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Container Runtime blackbox test on terminated container 
-  should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
+[sig-network] Proxy version v1 
+  should proxy logs on node using proxy subresource  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Runtime
+[BeforeEach] version v1
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:51:06.768: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-runtime
+Oct 26 17:06:15.638: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename proxy
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
+[It] should proxy logs on node using proxy subresource  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the container
-STEP: wait for the container to reach Failed
-STEP: get the container status
-STEP: the container should be terminated
-STEP: the termination message should be set
-Sep 21 16:51:10.371: INFO: Expected: &{DONE} to match Container's Termination Message: DONE --
-STEP: delete the container
-[AfterEach] [k8s.io] Container Runtime
+Oct 26 17:06:15.874: INFO: (0) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 58.083591ms)
+Oct 26 17:06:15.916: INFO: (1) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 41.290084ms)
+Oct 26 17:06:15.940: INFO: (2) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 24.301686ms)
+Oct 26 17:06:15.964: INFO: (3) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 23.538359ms)
+Oct 26 17:06:15.989: INFO: (4) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 25.272474ms)
+Oct 26 17:06:16.011: INFO: (5) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 21.8865ms)
+Oct 26 17:06:16.034: INFO: (6) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 22.549337ms)
+Oct 26 17:06:16.061: INFO: (7) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 27.206084ms)
+Oct 26 17:06:16.087: INFO: (8) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 25.059638ms)
+Oct 26 17:06:16.110: INFO: (9) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 23.397682ms)
+Oct 26 17:06:16.132: INFO: (10) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 21.919263ms)
+Oct 26 17:06:16.158: INFO: (11) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 25.70119ms)
+Oct 26 17:06:16.180: INFO: (12) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 21.781773ms)
+Oct 26 17:06:16.204: INFO: (13) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 24.463685ms)
+Oct 26 17:06:16.243: INFO: (14) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 38.504494ms)
+Oct 26 17:06:16.265: INFO: (15) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 21.808096ms)
+Oct 26 17:06:16.303: INFO: (16) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 38.234361ms)
+Oct 26 17:06:16.336: INFO: (17) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 32.77899ms)
+Oct 26 17:06:16.359: INFO: (18) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 22.910751ms)
+Oct 26 17:06:16.382: INFO: (19) /api/v1/nodes/10.72.119.82/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 22.748327ms)
+[AfterEach] version v1
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:51:10.504: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-runtime-8235" for this suite.
-•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":280,"completed":226,"skipped":3699,"failed":0}
-SSSSSSSS
+Oct 26 17:06:16.383: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "proxy-1204" for this suite.
+•{"msg":"PASSED [sig-network] Proxy version v1 should proxy logs on node using proxy subresource  [Conformance]","total":280,"completed":228,"skipped":3649,"failed":0}
+SSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl label 
-  should update the label on a resource  [Conformance]
+[sig-storage] Downward API volume 
+  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:51:10.572: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubectl
+Oct 26 17:06:16.424: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename downward-api
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[BeforeEach] Kubectl label
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1382
-STEP: creating the pod
-Sep 21 16:51:10.800: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 create -f - --namespace=kubectl-2524'
-Sep 21 16:51:11.481: INFO: stderr: ""
-Sep 21 16:51:11.481: INFO: stdout: "pod/pause created\n"
-Sep 21 16:51:11.481: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause]
-Sep 21 16:51:11.481: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-2524" to be "running and ready"
-Sep 21 16:51:11.502: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 21.584881ms
-Sep 21 16:51:13.525: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 2.044099586s
-Sep 21 16:51:15.547: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 4.066515851s
-Sep 21 16:51:15.547: INFO: Pod "pause" satisfied condition "running and ready"
-Sep 21 16:51:15.547: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause]
-[It] should update the label on a resource  [Conformance]
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: adding the label testing-label with value testing-label-value to a pod
-Sep 21 16:51:15.547: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 label pods pause testing-label=testing-label-value --namespace=kubectl-2524'
-Sep 21 16:51:15.733: INFO: stderr: ""
-Sep 21 16:51:15.733: INFO: stdout: "pod/pause labeled\n"
-STEP: verifying the pod has the label testing-label with the value testing-label-value
-Sep 21 16:51:15.733: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pod pause -L testing-label --namespace=kubectl-2524'
-Sep 21 16:51:15.842: INFO: stderr: ""
-Sep 21 16:51:15.842: INFO: stdout: "NAME    READY   STATUS    RESTARTS   AGE   TESTING-LABEL\npause   1/1     Running   0          4s    testing-label-value\n"
-STEP: removing the label testing-label of a pod
-Sep 21 16:51:15.842: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 label pods pause testing-label- --namespace=kubectl-2524'
-Sep 21 16:51:16.039: INFO: stderr: ""
-Sep 21 16:51:16.039: INFO: stdout: "pod/pause labeled\n"
-STEP: verifying the pod doesn't have the label testing-label
-Sep 21 16:51:16.039: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pod pause -L testing-label --namespace=kubectl-2524'
-Sep 21 16:51:16.178: INFO: stderr: ""
-Sep 21 16:51:16.178: INFO: stdout: "NAME    READY   STATUS    RESTARTS   AGE   TESTING-LABEL\npause   1/1     Running   0          5s    \n"
-[AfterEach] Kubectl label
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1389
-STEP: using delete to clean up resources
-Sep 21 16:51:16.178: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete --grace-period=0 --force -f - --namespace=kubectl-2524'
-Sep 21 16:51:16.324: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Sep 21 16:51:16.324: INFO: stdout: "pod \"pause\" force deleted\n"
-Sep 21 16:51:16.324: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get rc,svc -l name=pause --no-headers --namespace=kubectl-2524'
-Sep 21 16:51:16.489: INFO: stderr: "No resources found in kubectl-2524 namespace.\n"
-Sep 21 16:51:16.489: INFO: stdout: ""
-Sep 21 16:51:16.490: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 get pods -l name=pause --namespace=kubectl-2524 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
-Sep 21 16:51:16.611: INFO: stderr: ""
-Sep 21 16:51:16.611: INFO: stdout: ""
-[AfterEach] [sig-cli] Kubectl client
+STEP: Creating a pod to test downward API volume plugin
+Oct 26 17:06:16.653: INFO: Waiting up to 5m0s for pod "downwardapi-volume-78ee0998-8aaf-42b0-bf4c-07e15028fafe" in namespace "downward-api-806" to be "success or failure"
+Oct 26 17:06:16.669: INFO: Pod "downwardapi-volume-78ee0998-8aaf-42b0-bf4c-07e15028fafe": Phase="Pending", Reason="", readiness=false. Elapsed: 16.162994ms
+Oct 26 17:06:18.687: INFO: Pod "downwardapi-volume-78ee0998-8aaf-42b0-bf4c-07e15028fafe": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033890072s
+Oct 26 17:06:20.704: INFO: Pod "downwardapi-volume-78ee0998-8aaf-42b0-bf4c-07e15028fafe": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.050640861s
+STEP: Saw pod success
+Oct 26 17:06:20.704: INFO: Pod "downwardapi-volume-78ee0998-8aaf-42b0-bf4c-07e15028fafe" satisfied condition "success or failure"
+Oct 26 17:06:20.728: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-78ee0998-8aaf-42b0-bf4c-07e15028fafe container client-container: 
+STEP: delete the pod
+Oct 26 17:06:20.827: INFO: Waiting for pod downwardapi-volume-78ee0998-8aaf-42b0-bf4c-07e15028fafe to disappear
+Oct 26 17:06:20.843: INFO: Pod downwardapi-volume-78ee0998-8aaf-42b0-bf4c-07e15028fafe no longer exists
+[AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:51:16.611: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-2524" for this suite.
-
-• [SLOW TEST:6.106 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Kubectl label
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1379
-    should update the label on a resource  [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Kubectl label should update the label on a resource  [Conformance]","total":280,"completed":227,"skipped":3707,"failed":0}
-SSSSSSSSSSSSS
+Oct 26 17:06:20.843: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-806" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":229,"skipped":3670,"failed":0}
+SSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl api-versions 
-  should check if v1 is in available api versions  [Conformance]
+[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] 
+  should be able to convert a non homogeneous list of CRs [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:51:16.679: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubectl
+Oct 26 17:06:20.892: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename crd-webhook
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[It] should check if v1 is in available api versions  [Conformance]
+[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:125
+STEP: Setting up server cert
+STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication
+STEP: Deploying the custom resource conversion webhook pod
+STEP: Wait for the deployment to be ready
+Oct 26 17:06:21.808: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set
+Oct 26 17:06:23.864: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328781, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328781, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328781, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328781, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-78dcf5dd84\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Oct 26 17:06:26.917: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1
+[It] should be able to convert a non homogeneous list of CRs [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: validating api versions
-Sep 21 16:51:20.292: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 api-versions'
-Sep 21 16:51:20.391: INFO: stderr: ""
-Sep 21 16:51:20.391: INFO: stdout: "admissionregistration.k8s.io/v1\nadmissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps.openshift.io/v1\napps/v1\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nauthorization.openshift.io/v1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\nbuild.openshift.io/v1\ncertificates.k8s.io/v1beta1\ncloudcredential.openshift.io/v1\nconfig.openshift.io/v1\nconsole.openshift.io/v1\ncoordination.k8s.io/v1\ncoordination.k8s.io/v1beta1\ncrd.projectcalico.org/v1\ndiscovery.k8s.io/v1beta1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nibm.com/v1alpha1\nimage.openshift.io/v1\nimageregistry.operator.openshift.io/v1\ningress.operator.openshift.io/v1\nk8s.cni.cncf.io/v1\nmetal3.io/v1alpha1\nmetrics.k8s.io/v1beta1\nmigration.k8s.io/v1alpha1\nmonitoring.coreos.com/v1\nnetwork.operator.openshift.io/v1\nnetworking.k8s.io/v1\nnetworking.k8s.io/v1beta1\nnode.k8s.io/v1beta1\noauth.openshift.io/v1\noperator.openshift.io/v1\noperator.openshift.io/v1alpha1\noperator.tigera.io/v1\noperators.coreos.com/v1\noperators.coreos.com/v1alpha1\noperators.coreos.com/v1alpha2\noperators.coreos.com/v2\npackages.operators.coreos.com/v1\npolicy/v1beta1\nproject.openshift.io/v1\nquota.openshift.io/v1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nroute.openshift.io/v1\nsamples.operator.openshift.io/v1\nscheduling.k8s.io/v1\nscheduling.k8s.io/v1beta1\nsecurity.openshift.io/v1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\ntemplate.openshift.io/v1\ntuned.openshift.io/v1\nuser.openshift.io/v1\nv1\nwhereabouts.cni.cncf.io/v1alpha1\n"
-[AfterEach] [sig-cli] Kubectl client
+Oct 26 17:06:26.935: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Creating a v1 custom resource
+STEP: Create a v2 custom resource
+STEP: List CRs in v1
+STEP: List CRs in v2
+[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:51:20.391: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-3534" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions  [Conformance]","total":280,"completed":228,"skipped":3720,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 17:06:28.528: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-webhook-629" for this suite.
+[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:136
+
+• [SLOW TEST:7.915 seconds]
+[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should be able to convert a non homogeneous list of CRs [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance]","total":280,"completed":230,"skipped":3677,"failed":0}
+SSSSSS
 ------------------------------
 [sig-api-machinery] Garbage collector 
   should delete pods created by rc when not orphaning [Conformance]
@@ -11534,7 +10787,7 @@ SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 [BeforeEach] [sig-api-machinery] Garbage collector
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:51:20.506: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:06:28.814: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename gc
 STEP: Waiting for a default service account to be provisioned in namespace
 [It] should delete pods created by rc when not orphaning [Conformance]
@@ -11542,9 +10795,8 @@ STEP: Waiting for a default service account to be provisioned in namespace
 STEP: create the rc
 STEP: delete the rc
 STEP: wait for all pods to be garbage collected
-STEP: expected 0 pods, got 2 pods
 STEP: Gathering metrics
-Sep 21 16:51:36.191: INFO: For apiserver_request_total:
+Oct 26 17:06:39.170: INFO: For apiserver_request_total:
 For apiserver_request_latency_seconds:
 For apiserver_init_events_total:
 For garbage_collector_attempt_to_delete_queue_latency:
@@ -11567,355 +10819,509 @@ For function_duration_seconds:
 For errors_total:
 For evicted_pods_total:
 
+W1026 17:06:39.170192      26 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
 [AfterEach] [sig-api-machinery] Garbage collector
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:51:36.191: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-W0921 16:51:36.191274      24 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-STEP: Destroying namespace "gc-1799" for this suite.
+Oct 26 17:06:39.170: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-6445" for this suite.
 
-• [SLOW TEST:15.761 seconds]
+• [SLOW TEST:10.395 seconds]
 [sig-api-machinery] Garbage collector
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
   should delete pods created by rc when not orphaning [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance]","total":280,"completed":229,"skipped":3770,"failed":0}
-SSSSSSSS
-------------------------------
-[k8s.io] Security Context When creating a container with runAsUser 
-  should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:51:36.267: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename security-context-test
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39
-[It] should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:51:36.802: INFO: Waiting up to 5m0s for pod "busybox-user-65534-a957bd8d-6f85-4444-823e-60ad872fa581" in namespace "security-context-test-9412" to be "success or failure"
-Sep 21 16:51:36.827: INFO: Pod "busybox-user-65534-a957bd8d-6f85-4444-823e-60ad872fa581": Phase="Pending", Reason="", readiness=false. Elapsed: 24.309701ms
-Sep 21 16:51:38.855: INFO: Pod "busybox-user-65534-a957bd8d-6f85-4444-823e-60ad872fa581": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.05295183s
-Sep 21 16:51:38.855: INFO: Pod "busybox-user-65534-a957bd8d-6f85-4444-823e-60ad872fa581" satisfied condition "success or failure"
-[AfterEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:51:38.855: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "security-context-test-9412" for this suite.
-•{"msg":"PASSED [k8s.io] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":230,"skipped":3778,"failed":0}
-SSSSS
+{"msg":"PASSED [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance]","total":280,"completed":231,"skipped":3683,"failed":0}
+SSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide podname only [NodeConformance] [Conformance]
+[sig-api-machinery] Secrets 
+  should be consumable from pods in env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
+[BeforeEach] [sig-api-machinery] Secrets
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:51:38.933: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
+Oct 26 17:06:39.212: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename secrets
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should provide podname only [NodeConformance] [Conformance]
+[It] should be consumable from pods in env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:51:39.351: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2a04af61-57af-4f30-9855-ea1cb48d06dd" in namespace "downward-api-561" to be "success or failure"
-Sep 21 16:51:39.377: INFO: Pod "downwardapi-volume-2a04af61-57af-4f30-9855-ea1cb48d06dd": Phase="Pending", Reason="", readiness=false. Elapsed: 25.63903ms
-Sep 21 16:51:41.403: INFO: Pod "downwardapi-volume-2a04af61-57af-4f30-9855-ea1cb48d06dd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.051978595s
+STEP: Creating secret with name secret-test-5a90b084-ba48-4d6b-aa94-466b599e4cf4
+STEP: Creating a pod to test consume secrets
+Oct 26 17:06:39.445: INFO: Waiting up to 5m0s for pod "pod-secrets-e5811581-cfc0-44a3-b25a-afb886417558" in namespace "secrets-26" to be "success or failure"
+Oct 26 17:06:39.463: INFO: Pod "pod-secrets-e5811581-cfc0-44a3-b25a-afb886417558": Phase="Pending", Reason="", readiness=false. Elapsed: 18.028263ms
+Oct 26 17:06:41.484: INFO: Pod "pod-secrets-e5811581-cfc0-44a3-b25a-afb886417558": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.039043837s
 STEP: Saw pod success
-Sep 21 16:51:41.403: INFO: Pod "downwardapi-volume-2a04af61-57af-4f30-9855-ea1cb48d06dd" satisfied condition "success or failure"
-Sep 21 16:51:41.431: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-2a04af61-57af-4f30-9855-ea1cb48d06dd container client-container: 
+Oct 26 17:06:41.484: INFO: Pod "pod-secrets-e5811581-cfc0-44a3-b25a-afb886417558" satisfied condition "success or failure"
+Oct 26 17:06:41.499: INFO: Trying to get logs from node 10.72.119.74 pod pod-secrets-e5811581-cfc0-44a3-b25a-afb886417558 container secret-env-test: 
 STEP: delete the pod
-Sep 21 16:51:41.669: INFO: Waiting for pod downwardapi-volume-2a04af61-57af-4f30-9855-ea1cb48d06dd to disappear
-Sep 21 16:51:41.915: INFO: Pod downwardapi-volume-2a04af61-57af-4f30-9855-ea1cb48d06dd no longer exists
-[AfterEach] [sig-storage] Downward API volume
+Oct 26 17:06:41.582: INFO: Waiting for pod pod-secrets-e5811581-cfc0-44a3-b25a-afb886417558 to disappear
+Oct 26 17:06:41.597: INFO: Pod pod-secrets-e5811581-cfc0-44a3-b25a-afb886417558 no longer exists
+[AfterEach] [sig-api-machinery] Secrets
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:51:41.916: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-561" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance]","total":280,"completed":231,"skipped":3783,"failed":0}
-SSSSSSS
+Oct 26 17:06:41.597: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-26" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance]","total":280,"completed":232,"skipped":3700,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  works for multiple CRDs of same group and version but different kinds [Conformance]
+[k8s.io] [sig-node] Events 
+  should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+[BeforeEach] [k8s.io] [sig-node] Events
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:51:41.999: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename crd-publish-openapi
+Oct 26 17:06:41.636: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename events
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] works for multiple CRDs of same group and version but different kinds [Conformance]
+[It] should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: CRs in the same group and version but different kinds (two CRDs) show up in OpenAPI documentation
-Sep 21 16:51:42.221: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 16:51:50.826: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+STEP: retrieving the pod
+Oct 26 17:06:45.910: INFO: &Pod{ObjectMeta:{send-events-9585422b-ec80-405c-93b5-cd9bf0dcc2c6  events-6005 /api/v1/namespaces/events-6005/pods/send-events-9585422b-ec80-405c-93b5-cd9bf0dcc2c6 23040f06-b9c3-4949-8e86-55b32509c3fd 85517 0 2020-10-26 17:06:41 +0000 UTC   map[name:foo time:772014635] map[cni.projectcalico.org/podIP:172.30.194.77/32 cni.projectcalico.org/podIPs:172.30.194.77/32 k8s.v1.cni.cncf.io/networks-status:[{
+    "name": "k8s-pod-network",
+    "ips": [
+        "172.30.194.77"
+    ],
+    "dns": {}
+}] openshift.io/scc:anyuid] [] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-w2zll,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-w2zll,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:p,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[serve-hostname],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:,HostPort:0,ContainerPort:80,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-w2zll,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.74,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:&SELinuxOptions{User:,Role:,Type:,Level:s0:c64,c54,},RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:06:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:06:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:06:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:06:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.74,PodIP:172.30.194.77,StartTime:2020-10-26 17:06:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:p,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 17:06:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:gcr.io/kubernetes-e2e-test-images/agnhost@sha256:02b9e10ea1f7439a03d9bc3ee16d984172ca7eac1818792ad4d721c5c8f72ff4,ContainerID:cri-o://b58603fc4831449fc4e86b8b437155d356ca68fd96d47128f3ab250ead773ae5,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.194.77,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+
+STEP: checking for scheduler event about the pod
+Oct 26 17:06:47.922: INFO: Saw scheduler event for our pod.
+STEP: checking for kubelet event about the pod
+Oct 26 17:06:49.936: INFO: Saw kubelet event for our pod.
+STEP: deleting the pod
+[AfterEach] [k8s.io] [sig-node] Events
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:52:23.093: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-4082" for this suite.
+Oct 26 17:06:49.974: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "events-6005" for this suite.
 
-• [SLOW TEST:41.376 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for multiple CRDs of same group and version but different kinds [Conformance]
+• [SLOW TEST:8.376 seconds]
+[k8s.io] [sig-node] Events
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance]","total":280,"completed":232,"skipped":3790,"failed":0}
-SSSSSS
+{"msg":"PASSED [k8s.io] [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]","total":280,"completed":233,"skipped":3739,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-network] Services 
-  should be able to change the type from ClusterIP to ExternalName [Conformance]
+  should serve multiport endpoints from pods  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:52:23.374: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:06:50.014: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename services
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should be able to change the type from ClusterIP to ExternalName [Conformance]
+[It] should serve multiport endpoints from pods  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a service clusterip-service with the type=ClusterIP in namespace services-8215
-STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service
-STEP: creating service externalsvc in namespace services-8215
-STEP: creating replication controller externalsvc in namespace services-8215
-I0921 16:52:23.949226      24 runners.go:189] Created replication controller with name: externalsvc, namespace: services-8215, replica count: 2
-I0921 16:52:27.000075      24 runners.go:189] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-STEP: changing the ClusterIP service to type=ExternalName
-Sep 21 16:52:27.142: INFO: Creating new exec pod
-Sep 21 16:52:29.519: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-8215 execpodgs4s2 -- /bin/sh -x -c nslookup clusterip-service'
-Sep 21 16:52:29.954: INFO: stderr: "+ nslookup clusterip-service\n"
-Sep 21 16:52:29.954: INFO: stdout: "Server:\t\t172.21.0.10\nAddress:\t172.21.0.10#53\n\nclusterip-service.services-8215.svc.cluster.local\tcanonical name = externalsvc.services-8215.svc.cluster.local.\nName:\texternalsvc.services-8215.svc.cluster.local\nAddress: 172.21.158.68\n\n"
-STEP: deleting ReplicationController externalsvc in namespace services-8215, will wait for the garbage collector to delete the pods
-Sep 21 16:52:30.097: INFO: Deleting ReplicationController externalsvc took: 69.043934ms
-Sep 21 16:52:30.597: INFO: Terminating ReplicationController externalsvc pods took: 500.252822ms
-Sep 21 16:52:37.459: INFO: Cleaning up the ClusterIP to ExternalName test service
+STEP: creating service multi-endpoint-test in namespace services-8215
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8215 to expose endpoints map[]
+Oct 26 17:06:50.220: INFO: successfully validated that service multi-endpoint-test in namespace services-8215 exposes endpoints map[] (18.888268ms elapsed)
+STEP: Creating pod pod1 in namespace services-8215
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8215 to expose endpoints map[pod1:[100]]
+Oct 26 17:06:52.405: INFO: successfully validated that service multi-endpoint-test in namespace services-8215 exposes endpoints map[pod1:[100]] (2.124382892s elapsed)
+STEP: Creating pod pod2 in namespace services-8215
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8215 to expose endpoints map[pod1:[100] pod2:[101]]
+Oct 26 17:06:55.662: INFO: successfully validated that service multi-endpoint-test in namespace services-8215 exposes endpoints map[pod1:[100] pod2:[101]] (3.217251144s elapsed)
+STEP: Deleting pod pod1 in namespace services-8215
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8215 to expose endpoints map[pod2:[101]]
+Oct 26 17:06:55.725: INFO: successfully validated that service multi-endpoint-test in namespace services-8215 exposes endpoints map[pod2:[101]] (34.719604ms elapsed)
+STEP: Deleting pod pod2 in namespace services-8215
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8215 to expose endpoints map[]
+Oct 26 17:06:55.775: INFO: successfully validated that service multi-endpoint-test in namespace services-8215 exposes endpoints map[] (18.914212ms elapsed)
 [AfterEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:52:37.587: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+Oct 26 17:06:55.856: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
 STEP: Destroying namespace "services-8215" for this suite.
 [AfterEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
 
-• [SLOW TEST:14.286 seconds]
+• [SLOW TEST:5.904 seconds]
 [sig-network] Services
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should be able to change the type from ClusterIP to ExternalName [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance]","total":280,"completed":233,"skipped":3796,"failed":0}
-SSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-api-machinery] Garbage collector 
-  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:52:37.662: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename gc
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the rc1
-STEP: create the rc2
-STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well
-STEP: delete the rc simpletest-rc-to-be-deleted
-STEP: wait for the rc to be deleted
-STEP: Gathering metrics
-W0921 16:52:48.379598      24 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-Sep 21 16:52:48.379: INFO: For apiserver_request_total:
-For apiserver_request_latency_seconds:
-For apiserver_init_events_total:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
-
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:52:48.379: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-4981" for this suite.
-
-• [SLOW TEST:10.786 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]","total":280,"completed":234,"skipped":3815,"failed":0}
-SSSS
-------------------------------
-[sig-storage] Downward API volume 
-  should update labels on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:52:48.448: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should update labels on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating the pod
-Sep 21 16:52:53.657: INFO: Successfully updated pod "labelsupdateaba1795c-1034-44d4-abf9-087d9683521f"
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:52:56.033: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-4595" for this suite.
-
-• [SLOW TEST:7.653 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:35
-  should update labels on modification [NodeConformance] [Conformance]
+  should serve multiport endpoints from pods  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance]","total":280,"completed":235,"skipped":3819,"failed":0}
-SSSSSSSSSSSSS
+{"msg":"PASSED [sig-network] Services should serve multiport endpoints from pods  [Conformance]","total":280,"completed":234,"skipped":3780,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-storage] Projected configMap 
-  should be consumable from pods in volume [NodeConformance] [Conformance]
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:52:56.102: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:06:55.919: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name projected-configmap-test-volume-812b1841-17f4-4f1d-a22b-e7dd0ee3dbe6
+STEP: Creating configMap with name projected-configmap-test-volume-map-9a868e6b-754d-44e4-b154-601a4a08f8a1
 STEP: Creating a pod to test consume configMaps
-Sep 21 16:52:56.488: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-a8f58771-2e70-42ac-ae6f-bdc76a69d0a2" in namespace "projected-8802" to be "success or failure"
-Sep 21 16:52:56.520: INFO: Pod "pod-projected-configmaps-a8f58771-2e70-42ac-ae6f-bdc76a69d0a2": Phase="Pending", Reason="", readiness=false. Elapsed: 32.516414ms
-Sep 21 16:52:58.769: INFO: Pod "pod-projected-configmaps-a8f58771-2e70-42ac-ae6f-bdc76a69d0a2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.281010846s
+Oct 26 17:06:56.165: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-381a08d2-2206-462d-a645-50eaaad774b1" in namespace "projected-4859" to be "success or failure"
+Oct 26 17:06:56.187: INFO: Pod "pod-projected-configmaps-381a08d2-2206-462d-a645-50eaaad774b1": Phase="Pending", Reason="", readiness=false. Elapsed: 21.531147ms
+Oct 26 17:06:58.204: INFO: Pod "pod-projected-configmaps-381a08d2-2206-462d-a645-50eaaad774b1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.038735017s
+Oct 26 17:07:00.225: INFO: Pod "pod-projected-configmaps-381a08d2-2206-462d-a645-50eaaad774b1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.060003061s
 STEP: Saw pod success
-Sep 21 16:52:58.769: INFO: Pod "pod-projected-configmaps-a8f58771-2e70-42ac-ae6f-bdc76a69d0a2" satisfied condition "success or failure"
-Sep 21 16:52:58.799: INFO: Trying to get logs from node 10.241.51.147 pod pod-projected-configmaps-a8f58771-2e70-42ac-ae6f-bdc76a69d0a2 container projected-configmap-volume-test: 
+Oct 26 17:07:00.225: INFO: Pod "pod-projected-configmaps-381a08d2-2206-462d-a645-50eaaad774b1" satisfied condition "success or failure"
+Oct 26 17:07:00.242: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-configmaps-381a08d2-2206-462d-a645-50eaaad774b1 container projected-configmap-volume-test: 
 STEP: delete the pod
-Sep 21 16:52:58.948: INFO: Waiting for pod pod-projected-configmaps-a8f58771-2e70-42ac-ae6f-bdc76a69d0a2 to disappear
-Sep 21 16:52:58.969: INFO: Pod pod-projected-configmaps-a8f58771-2e70-42ac-ae6f-bdc76a69d0a2 no longer exists
+Oct 26 17:07:00.362: INFO: Waiting for pod pod-projected-configmaps-381a08d2-2206-462d-a645-50eaaad774b1 to disappear
+Oct 26 17:07:00.382: INFO: Pod pod-projected-configmaps-381a08d2-2206-462d-a645-50eaaad774b1 no longer exists
 [AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:52:58.970: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-8802" for this suite.
-•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":236,"skipped":3832,"failed":0}
-SSSSSSSSSSSS
+Oct 26 17:07:00.382: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-4859" for this suite.
+•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":235,"skipped":3845,"failed":0}
+SSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Docker Containers 
-  should be able to override the image's default command and arguments [NodeConformance] [Conformance]
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates that NodeSelector is respected if matching  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Docker Containers
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:52:59.050: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename containers
+Oct 26 17:07:00.421: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename sched-pred
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to override the image's default command and arguments [NodeConformance] [Conformance]
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
+Oct 26 17:07:00.626: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Oct 26 17:07:00.689: INFO: Waiting for terminating namespaces to be deleted...
+Oct 26 17:07:00.711: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.74 before test
+Oct 26 17:07:00.780: INFO: calico-node-xlzf4 from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: cluster-samples-operator-848994bd88-rw4rq from openshift-cluster-samples-operator started at 2020-10-26 16:09:41 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container cluster-samples-operator ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container cluster-samples-operator-watch ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: pod1 from services-8215 started at 2020-10-26 17:06:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container pause ready: false, restart count 0
+Oct 26 17:07:00.781: INFO: ibm-keepalived-watcher-49jbs from kube-system started at 2020-10-26 14:34:28 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: redhat-operators-5579777cd5-q5qm5 from openshift-marketplace started at 2020-10-26 16:39:28 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container redhat-operators ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-nbr6w from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Oct 26 17:07:00.781: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: service-ca-7d9db9c5b8-4w4cj from openshift-service-ca started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container service-ca-controller ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: openshift-kube-proxy-7d26g from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: tuned-p4cpc from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: node-ca-4ppdl from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: send-events-9585422b-ec80-405c-93b5-cd9bf0dcc2c6 from events-6005 started at 2020-10-26 17:06:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container p ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: vpn-5b455f6f88-64q7j from kube-system started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container vpn ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: tigera-operator-5c45c894b4-z9bnx from tigera-operator started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container tigera-operator ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: ibm-master-proxy-static-10.72.119.74 from kube-system started at 2020-10-26 14:34:26 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container pause ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: multus-cc44t from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: alertmanager-main-0 from openshift-monitoring started at 2020-10-26 16:09:57 +0000 UTC (3 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: multus-admission-controller-dnt9r from openshift-multus started at 2020-10-26 16:10:16 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: packageserver-5f76c64df7-jg7sj from openshift-operator-lifecycle-manager started at 2020-10-26 16:10:08 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container packageserver ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: telemeter-client-d5d9896c4-jpd8z from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (3 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container reload ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container telemeter-client ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: calico-typha-7fd49c779d-46t4v from calico-system started at 2020-10-26 14:37:36 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: sonobuoy from sonobuoy started at 2020-10-26 15:59:53 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: sonobuoy-e2e-job-c6d919d638514ce4 from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container e2e ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: thanos-querier-84d7d58bc-c27ss from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (4 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container oauth-proxy ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container thanos-querier ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: grafana-f999495bf-t4tcv from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container grafana ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container grafana-proxy ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: console-5dd98d6f6f-zxw5x from openshift-console started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container console ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: ibmcloud-block-storage-driver-ws85j from kube-system started at 2020-10-26 14:34:32 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: node-exporter-ls6pf from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: dns-default-s9kvh from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container dns ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: prometheus-adapter-674456c5d6-p6bj5 from openshift-monitoring started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.781: INFO: 	Container prometheus-adapter ready: true, restart count 0
+Oct 26 17:07:00.781: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.76 before test
+Oct 26 17:07:00.953: INFO: openshift-kube-proxy-74ss4 from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.954: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 17:07:00.954: INFO: ibm-file-plugin-944b59c94-k7mrm from kube-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.954: INFO: 	Container ibm-file-plugin-container ready: true, restart count 0
+Oct 26 17:07:00.954: INFO: calico-typha-7fd49c779d-766qb from calico-system started at 2020-10-26 14:37:36 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.954: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 17:07:00.954: INFO: dns-default-z8pn9 from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.954: INFO: 	Container dns ready: true, restart count 0
+Oct 26 17:07:00.954: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 17:07:00.954: INFO: ibmcloud-block-storage-driver-b724k from kube-system started at 2020-10-26 14:34:34 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.954: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 17:07:00.954: INFO: openshift-service-catalog-apiserver-operator-7cfbcc49dc-hrn66 from openshift-service-catalog-apiserver-operator started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.954: INFO: 	Container operator ready: true, restart count 1
+Oct 26 17:07:00.954: INFO: alertmanager-main-2 from openshift-monitoring started at 2020-10-26 14:44:28 +0000 UTC (3 container statuses recorded)
+Oct 26 17:07:00.954: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 17:07:00.954: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 17:07:00.954: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 17:07:00.954: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-2scb4 from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.954: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Oct 26 17:07:00.954: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 17:07:00.954: INFO: router-default-84766ff4fd-cftxw from openshift-ingress started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.954: INFO: 	Container router ready: true, restart count 0
+Oct 26 17:07:00.955: INFO: multus-lbpcn from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.955: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 17:07:00.955: INFO: openshift-service-catalog-controller-manager-operator-d49bz89wp from openshift-service-catalog-controller-manager-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.955: INFO: 	Container operator ready: true, restart count 1
+Oct 26 17:07:00.955: INFO: ibm-storage-watcher-77d6cc685c-bvg8q from kube-system started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.955: INFO: 	Container ibm-storage-watcher-container ready: true, restart count 0
+Oct 26 17:07:00.955: INFO: olm-operator-7cd6cbf579-zww2l from openshift-operator-lifecycle-manager started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.955: INFO: 	Container olm-operator ready: true, restart count 0
+Oct 26 17:07:00.955: INFO: cluster-monitoring-operator-69649564f8-r4h9b from openshift-monitoring started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.955: INFO: 	Container cluster-monitoring-operator ready: true, restart count 0
+Oct 26 17:07:00.955: INFO: ibm-cloud-provider-ip-158-176-78-170-5cdb49c95f-76fjw from ibm-system started at 2020-10-26 14:42:32 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.955: INFO: 	Container ibm-cloud-provider-ip-158-176-78-170 ready: true, restart count 0
+Oct 26 17:07:00.955: INFO: ingress-operator-65878dcc8d-wdnrl from openshift-ingress-operator started at 2020-10-26 14:35:49 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.955: INFO: 	Container ingress-operator ready: true, restart count 0
+Oct 26 17:07:00.955: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:00.955: INFO: downloads-74f6b6dcb6-r9ksv from openshift-console started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.955: INFO: 	Container download-server ready: true, restart count 0
+Oct 26 17:07:00.955: INFO: multus-admission-controller-5j529 from openshift-multus started at 2020-10-26 14:35:50 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.955: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:00.955: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 17:07:00.956: INFO: node-exporter-zmwjz from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.956: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:00.956: INFO: 	Container node-exporter ready: true, restart count 0
+Oct 26 17:07:00.956: INFO: node-ca-bpxm5 from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.956: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 17:07:00.956: INFO: cluster-node-tuning-operator-575b98978c-zz7s9 from openshift-cluster-node-tuning-operator started at 2020-10-26 14:35:51 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.956: INFO: 	Container cluster-node-tuning-operator ready: true, restart count 0
+Oct 26 17:07:00.956: INFO: calico-node-dhd7p from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.956: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 17:07:00.956: INFO: kube-storage-version-migrator-operator-86d479ffb5-9cmml from openshift-kube-storage-version-migrator-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.956: INFO: 	Container kube-storage-version-migrator-operator ready: true, restart count 1
+Oct 26 17:07:00.956: INFO: catalog-operator-5db6d5849c-z2lgx from openshift-operator-lifecycle-manager started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.956: INFO: 	Container catalog-operator ready: true, restart count 0
+Oct 26 17:07:00.956: INFO: calico-kube-controllers-cdc7f7f8f-jqklm from calico-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.956: INFO: 	Container calico-kube-controllers ready: true, restart count 0
+Oct 26 17:07:00.956: INFO: downloads-74f6b6dcb6-vqzf9 from openshift-console started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.956: INFO: 	Container download-server ready: true, restart count 0
+Oct 26 17:07:00.956: INFO: ibm-keepalived-watcher-qwg48 from kube-system started at 2020-10-26 14:34:30 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.956: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 17:07:00.956: INFO: ibmcloud-block-storage-plugin-5f486d686f-6brjv from kube-system started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.956: INFO: 	Container ibmcloud-block-storage-plugin-container ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: ibm-master-proxy-static-10.72.119.76 from kube-system started at 2020-10-26 14:34:28 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.957: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: 	Container pause ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: cluster-image-registry-operator-8444cfb976-gdjk6 from openshift-image-registry started at 2020-10-26 14:35:50 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.957: INFO: 	Container cluster-image-registry-operator ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: 	Container cluster-image-registry-operator-watch ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: prometheus-k8s-0 from openshift-monitoring started at 2020-10-26 14:45:39 +0000 UTC (7 container statuses recorded)
+Oct 26 17:07:00.957: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: 	Container prometheus ready: true, restart count 1
+Oct 26 17:07:00.957: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: 	Container prometheus-proxy ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: 	Container thanos-sidecar ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: tuned-s6flg from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.957: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: network-operator-86b67c6f7f-pcq2w from openshift-network-operator started at 2020-10-26 14:34:37 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.957: INFO: 	Container network-operator ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: dns-operator-58d86f779c-wbm98 from openshift-dns-operator started at 2020-10-26 14:35:49 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:00.957: INFO: 	Container dns-operator ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:00.957: INFO: cluster-storage-operator-9b57569c9-gpqd2 from openshift-cluster-storage-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.958: INFO: 	Container cluster-storage-operator ready: true, restart count 0
+Oct 26 17:07:00.958: INFO: service-ca-operator-656f67898-svtt9 from openshift-service-ca-operator started at 2020-10-26 14:35:49 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.958: INFO: 	Container operator ready: true, restart count 1
+Oct 26 17:07:00.958: INFO: marketplace-operator-5df598b96b-6vr64 from openshift-marketplace started at 2020-10-26 14:35:50 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.958: INFO: 	Container marketplace-operator ready: true, restart count 0
+Oct 26 17:07:00.958: INFO: console-operator-5fc674d8d4-fjctd from openshift-console-operator started at 2020-10-26 14:35:51 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:00.958: INFO: 	Container console-operator ready: true, restart count 1
+Oct 26 17:07:00.958: INFO: 
+Logging pods the kubelet thinks is on node 10.72.119.82 before test
+Oct 26 17:07:01.048: INFO: prometheus-operator-55f946f5d-tpnd8 from openshift-monitoring started at 2020-10-26 14:44:11 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.048: INFO: 	Container prometheus-operator ready: true, restart count 0
+Oct 26 17:07:01.048: INFO: router-default-84766ff4fd-7dgrd from openshift-ingress started at 2020-10-26 14:38:40 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.048: INFO: 	Container router ready: true, restart count 0
+Oct 26 17:07:01.048: INFO: multus-admission-controller-9rv7d from openshift-multus started at 2020-10-26 14:36:02 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:01.048: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:01.048: INFO: 	Container multus-admission-controller ready: true, restart count 0
+Oct 26 17:07:01.048: INFO: ibm-cloud-provider-ip-158-176-78-170-5cdb49c95f-89x4j from ibm-system started at 2020-10-26 16:09:41 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.048: INFO: 	Container ibm-cloud-provider-ip-158-176-78-170 ready: true, restart count 0
+Oct 26 17:07:01.048: INFO: console-5dd98d6f6f-xvz7c from openshift-console started at 2020-10-26 14:45:54 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.048: INFO: 	Container console ready: true, restart count 0
+Oct 26 17:07:01.048: INFO: packageserver-5f76c64df7-6nrvc from openshift-operator-lifecycle-manager started at 2020-10-26 16:10:11 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.048: INFO: 	Container packageserver ready: true, restart count 0
+Oct 26 17:07:01.048: INFO: calico-node-c7wjs from calico-system started at 2020-10-26 14:35:40 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.048: INFO: 	Container calico-node ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: dns-default-ggg6q from openshift-dns started at 2020-10-26 14:38:14 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:01.049: INFO: 	Container dns ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: 	Container dns-node-resolver ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: prometheus-k8s-1 from openshift-monitoring started at 2020-10-26 14:45:29 +0000 UTC (7 container statuses recorded)
+Oct 26 17:07:01.049: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: 	Container prometheus ready: true, restart count 1
+Oct 26 17:07:01.049: INFO: 	Container prometheus-config-reloader ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: 	Container prometheus-proxy ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: 	Container rules-configmap-reloader ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: 	Container thanos-sidecar ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: kube-state-metrics-7bfb6f7864-vmgz9 from openshift-monitoring started at 2020-10-26 14:36:32 +0000 UTC (3 container statuses recorded)
+Oct 26 17:07:01.049: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: 	Container kube-state-metrics ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: thanos-querier-84d7d58bc-mpnzq from openshift-monitoring started at 2020-10-26 14:45:18 +0000 UTC (4 container statuses recorded)
+Oct 26 17:07:01.049: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: 	Container oauth-proxy ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: 	Container prom-label-proxy ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: 	Container thanos-querier ready: true, restart count 0
+Oct 26 17:07:01.049: INFO: community-operators-f4df99f64-twf5k from openshift-marketplace started at 2020-10-26 14:38:39 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.050: INFO: 	Container community-operators ready: true, restart count 0
+Oct 26 17:07:01.050: INFO: ibm-master-proxy-static-10.72.119.82 from kube-system started at 2020-10-26 14:34:29 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:01.050: INFO: 	Container ibm-master-proxy-static ready: true, restart count 0
+Oct 26 17:07:01.050: INFO: 	Container pause ready: true, restart count 0
+Oct 26 17:07:01.050: INFO: ibmcloud-block-storage-driver-sxlj2 from kube-system started at 2020-10-26 14:34:40 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.050: INFO: 	Container ibmcloud-block-storage-driver-container ready: true, restart count 0
+Oct 26 17:07:01.050: INFO: multus-hwkgh from openshift-multus started at 2020-10-26 14:35:04 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.050: INFO: 	Container kube-multus ready: true, restart count 0
+Oct 26 17:07:01.050: INFO: node-ca-z564l from openshift-image-registry started at 2020-10-26 14:38:21 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.050: INFO: 	Container node-ca ready: true, restart count 0
+Oct 26 17:07:01.050: INFO: redhat-marketplace-77d9b69f69-qb5s9 from openshift-marketplace started at 2020-10-26 14:38:39 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.050: INFO: 	Container redhat-marketplace ready: true, restart count 0
+Oct 26 17:07:01.050: INFO: alertmanager-main-1 from openshift-monitoring started at 2020-10-26 14:44:39 +0000 UTC (3 container statuses recorded)
+Oct 26 17:07:01.050: INFO: 	Container alertmanager ready: true, restart count 0
+Oct 26 17:07:01.050: INFO: 	Container alertmanager-proxy ready: true, restart count 0
+Oct 26 17:07:01.050: INFO: 	Container config-reloader ready: true, restart count 0
+Oct 26 17:07:01.050: INFO: ibm-keepalived-watcher-98n85 from kube-system started at 2020-10-26 14:34:32 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.050: INFO: 	Container keepalived-watcher ready: true, restart count 0
+Oct 26 17:07:01.050: INFO: tuned-h578f from openshift-cluster-node-tuning-operator started at 2020-10-26 14:36:31 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.050: INFO: 	Container tuned ready: true, restart count 0
+Oct 26 17:07:01.050: INFO: certified-operators-69bb9f65cc-9mhpt from openshift-marketplace started at 2020-10-26 14:38:38 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.051: INFO: 	Container certified-operators ready: true, restart count 0
+Oct 26 17:07:01.051: INFO: pod2 from services-8215 started at 2020-10-26 17:06:52 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.051: INFO: 	Container pause ready: false, restart count 0
+Oct 26 17:07:01.051: INFO: calico-typha-7fd49c779d-nmzbk from calico-system started at 2020-10-26 14:35:39 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.051: INFO: 	Container calico-typha ready: true, restart count 0
+Oct 26 17:07:01.051: INFO: test-k8s-e2e-pvg-master-verification from default started at 2020-10-26 14:37:46 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.051: INFO: 	Container test-k8s-e2e-pvg-master-verification ready: true, restart count 0
+Oct 26 17:07:01.051: INFO: registry-pvc-permissions-m9wwv from openshift-image-registry started at 2020-10-26 14:40:09 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.051: INFO: 	Container pvc-permissions ready: false, restart count 0
+Oct 26 17:07:01.051: INFO: migrator-686fc6cc66-9wpj9 from openshift-kube-storage-version-migrator started at 2020-10-26 14:36:53 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.051: INFO: 	Container migrator ready: true, restart count 0
+Oct 26 17:07:01.051: INFO: prometheus-adapter-674456c5d6-kj77x from openshift-monitoring started at 2020-10-26 14:44:24 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.051: INFO: 	Container prometheus-adapter ready: true, restart count 0
+Oct 26 17:07:01.051: INFO: image-registry-6db5c967cf-d4ljt from openshift-image-registry started at 2020-10-26 14:40:09 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.051: INFO: 	Container registry ready: true, restart count 0
+Oct 26 17:07:01.051: INFO: sonobuoy-systemd-logs-daemon-set-1900f6b6edd84c23-x2lkl from sonobuoy started at 2020-10-26 16:00:03 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:01.051: INFO: 	Container sonobuoy-worker ready: true, restart count 1
+Oct 26 17:07:01.051: INFO: 	Container systemd-logs ready: true, restart count 0
+Oct 26 17:07:01.051: INFO: openshift-kube-proxy-z9sjt from openshift-kube-proxy started at 2020-10-26 14:35:10 +0000 UTC (1 container statuses recorded)
+Oct 26 17:07:01.051: INFO: 	Container kube-proxy ready: true, restart count 0
+Oct 26 17:07:01.052: INFO: openshift-state-metrics-d4d5d6f57-lszmv from openshift-monitoring started at 2020-10-26 14:36:34 +0000 UTC (3 container statuses recorded)
+Oct 26 17:07:01.052: INFO: 	Container kube-rbac-proxy-main ready: true, restart count 0
+Oct 26 17:07:01.052: INFO: 	Container kube-rbac-proxy-self ready: true, restart count 0
+Oct 26 17:07:01.052: INFO: 	Container openshift-state-metrics ready: true, restart count 0
+Oct 26 17:07:01.052: INFO: node-exporter-4lfg7 from openshift-monitoring started at 2020-10-26 14:36:35 +0000 UTC (2 container statuses recorded)
+Oct 26 17:07:01.052: INFO: 	Container kube-rbac-proxy ready: true, restart count 0
+Oct 26 17:07:01.052: INFO: 	Container node-exporter ready: true, restart count 0
+[It] validates that NodeSelector is respected if matching  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test override all
-Sep 21 16:52:59.490: INFO: Waiting up to 5m0s for pod "client-containers-0e586203-2e96-42cc-a66c-e5065475f05f" in namespace "containers-7680" to be "success or failure"
-Sep 21 16:52:59.521: INFO: Pod "client-containers-0e586203-2e96-42cc-a66c-e5065475f05f": Phase="Pending", Reason="", readiness=false. Elapsed: 31.703549ms
-Sep 21 16:53:01.545: INFO: Pod "client-containers-0e586203-2e96-42cc-a66c-e5065475f05f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.055429909s
-STEP: Saw pod success
-Sep 21 16:53:01.545: INFO: Pod "client-containers-0e586203-2e96-42cc-a66c-e5065475f05f" satisfied condition "success or failure"
-Sep 21 16:53:01.571: INFO: Trying to get logs from node 10.241.51.147 pod client-containers-0e586203-2e96-42cc-a66c-e5065475f05f container test-container: 
-STEP: delete the pod
-Sep 21 16:53:01.731: INFO: Waiting for pod client-containers-0e586203-2e96-42cc-a66c-e5065475f05f to disappear
-Sep 21 16:53:01.759: INFO: Pod client-containers-0e586203-2e96-42cc-a66c-e5065475f05f no longer exists
-[AfterEach] [k8s.io] Docker Containers
+STEP: Trying to launch a pod without a label to get a node which can launch it.
+STEP: Explicitly delete pod here to free the resource it takes.
+STEP: Trying to apply a random label on the found node.
+STEP: verifying the node has the label kubernetes.io/e2e-c7696a58-dad3-4c3a-9a47-2fef628b1538 42
+STEP: Trying to relaunch the pod, now with labels.
+STEP: removing the label kubernetes.io/e2e-c7696a58-dad3-4c3a-9a47-2fef628b1538 off the node 10.72.119.74
+STEP: verifying the node doesn't have the label kubernetes.io/e2e-c7696a58-dad3-4c3a-9a47-2fef628b1538
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:53:01.759: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "containers-7680" for this suite.
-•{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance]","total":280,"completed":237,"skipped":3844,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 17:07:07.431: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-3306" for this suite.
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
+
+• [SLOW TEST:7.053 seconds]
+[sig-scheduling] SchedulerPredicates [Serial]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
+  validates that NodeSelector is respected if matching  [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-cli] Kubectl client Kubectl run job 
-  should create a job from an image when restart is OnFailure  [Conformance]
+{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching  [Conformance]","total":280,"completed":236,"skipped":3863,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir wrapper volumes 
+  should not conflict [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
+[BeforeEach] [sig-storage] EmptyDir wrapper volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:53:01.844: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubectl
+Oct 26 17:07:07.475: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir-wrapper
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
-[BeforeEach] Kubectl run job
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1788
-[It] should create a job from an image when restart is OnFailure  [Conformance]
+[It] should not conflict [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: running the image docker.io/library/httpd:2.4.38-alpine
-Sep 21 16:53:02.140: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 run e2e-test-httpd-job --restart=OnFailure --generator=job/v1 --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-4642'
-Sep 21 16:53:02.289: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-Sep 21 16:53:02.290: INFO: stdout: "job.batch/e2e-test-httpd-job created\n"
-STEP: verifying the job e2e-test-httpd-job was created
-[AfterEach] Kubectl run job
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1793
-Sep 21 16:53:02.339: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 delete jobs e2e-test-httpd-job --namespace=kubectl-4642'
-Sep 21 16:53:02.505: INFO: stderr: ""
-Sep 21 16:53:02.505: INFO: stdout: "job.batch \"e2e-test-httpd-job\" deleted\n"
-[AfterEach] [sig-cli] Kubectl client
+STEP: Cleaning up the secret
+STEP: Cleaning up the configmap
+STEP: Cleaning up the pod
+[AfterEach] [sig-storage] EmptyDir wrapper volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:53:02.505: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-4642" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run job should create a job from an image when restart is OnFailure  [Conformance]","total":280,"completed":238,"skipped":3893,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 17:07:09.917: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-wrapper-1186" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance]","total":280,"completed":237,"skipped":3885,"failed":0}
+SSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Pods 
-  should be submitted and removed [NodeConformance] [Conformance]
+[sig-storage] Secrets 
+  should be consumable from pods in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Pods
+[BeforeEach] [sig-storage] Secrets
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:53:02.595: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename pods
+Oct 26 17:07:09.955: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename secrets
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
-[It] should be submitted and removed [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-STEP: setting up watch
-STEP: submitting the pod to kubernetes
-Sep 21 16:53:02.938: INFO: observed the pod list
-STEP: verifying the pod is in kubernetes
-STEP: verifying pod creation was observed
-STEP: deleting the pod gracefully
-STEP: verifying the kubelet observed the termination notice
-STEP: verifying pod deletion was observed
-[AfterEach] [k8s.io] Pods
+STEP: Creating secret with name secret-test-a65fc234-1f7c-4416-bf82-1101a2547bbb
+STEP: Creating a pod to test consume secrets
+Oct 26 17:07:10.190: INFO: Waiting up to 5m0s for pod "pod-secrets-5518509e-f6fa-4b49-a43d-9c73d7dc826f" in namespace "secrets-421" to be "success or failure"
+Oct 26 17:07:10.208: INFO: Pod "pod-secrets-5518509e-f6fa-4b49-a43d-9c73d7dc826f": Phase="Pending", Reason="", readiness=false. Elapsed: 17.426254ms
+Oct 26 17:07:12.225: INFO: Pod "pod-secrets-5518509e-f6fa-4b49-a43d-9c73d7dc826f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034773122s
+Oct 26 17:07:14.242: INFO: Pod "pod-secrets-5518509e-f6fa-4b49-a43d-9c73d7dc826f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.051295859s
+STEP: Saw pod success
+Oct 26 17:07:14.242: INFO: Pod "pod-secrets-5518509e-f6fa-4b49-a43d-9c73d7dc826f" satisfied condition "success or failure"
+Oct 26 17:07:14.259: INFO: Trying to get logs from node 10.72.119.74 pod pod-secrets-5518509e-f6fa-4b49-a43d-9c73d7dc826f container secret-volume-test: 
+STEP: delete the pod
+Oct 26 17:07:14.347: INFO: Waiting for pod pod-secrets-5518509e-f6fa-4b49-a43d-9c73d7dc826f to disappear
+Oct 26 17:07:14.362: INFO: Pod pod-secrets-5518509e-f6fa-4b49-a43d-9c73d7dc826f no longer exists
+[AfterEach] [sig-storage] Secrets
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:53:17.361: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-1227" for this suite.
-
-• [SLOW TEST:14.827 seconds]
-[k8s.io] Pods
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should be submitted and removed [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [k8s.io] Pods should be submitted and removed [NodeConformance] [Conformance]","total":280,"completed":239,"skipped":3922,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 17:07:14.362: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-421" for this suite.
+•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":238,"skipped":3900,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-api-machinery] Aggregator 
   Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
@@ -11923,70 +11329,161 @@ SSSSSSSSSSSSSSSSSSSSSSS
 [BeforeEach] [sig-api-machinery] Aggregator
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:53:17.422: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:07:14.408: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename aggregator
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-api-machinery] Aggregator
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:76
-Sep 21 16:53:17.614: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:07:14.588: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 [It] Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Registering the sample API server.
-Sep 21 16:53:18.601: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:53:20.621: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:53:22.631: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:53:24.621: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:53:26.629: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:53:28.620: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736303998, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:53:32.431: INFO: Waited 1.773513289s for the sample-apiserver to be ready to handle requests.
+Oct 26 17:07:15.587: INFO: deployment "sample-apiserver-deployment" doesn't have the required revision set
+Oct 26 17:07:17.773: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:07:19.791: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:07:21.789: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:07:23.793: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:07:25.792: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:07:27.790: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:07:29.797: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:07:31.790: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:07:33.789: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328852, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328835, loc:(*time.Location)(0x7db5bc0)}}, Reason:"NewReplicaSetAvailable", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" has successfully progressed."}, v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328853, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328853, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:07:36.029: INFO: Waited 219.989791ms for the sample-apiserver to be ready to handle requests.
 [AfterEach] [sig-api-machinery] Aggregator
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:67
 [AfterEach] [sig-api-machinery] Aggregator
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:53:34.775: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "aggregator-2596" for this suite.
+Oct 26 17:07:38.145: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "aggregator-2135" for this suite.
 
-• [SLOW TEST:17.496 seconds]
+• [SLOW TEST:23.842 seconds]
 [sig-api-machinery] Aggregator
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
   Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Aggregator Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]","total":280,"completed":240,"skipped":3945,"failed":0}
-SSSSSS
+{"msg":"PASSED [sig-api-machinery] Aggregator Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]","total":280,"completed":239,"skipped":3930,"failed":0}
+SSSSS
 ------------------------------
-[sig-api-machinery] ResourceQuota 
-  should create a ResourceQuota and capture the life of a secret. [Conformance]
+[sig-storage] ConfigMap 
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
+[BeforeEach] [sig-storage] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:53:34.918: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename resourcequota
+Oct 26 17:07:38.250: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename configmap
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should create a ResourceQuota and capture the life of a secret. [Conformance]
+[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Discovering how many secrets are in namespace by default
-STEP: Counting existing ResourceQuota
-STEP: Creating a ResourceQuota
-STEP: Ensuring resource quota status is calculated
-STEP: Creating a Secret
-STEP: Ensuring resource quota status captures secret creation
-STEP: Deleting a secret
-STEP: Ensuring resource quota status released usage
-[AfterEach] [sig-api-machinery] ResourceQuota
+Oct 26 17:07:38.422: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
+STEP: Creating configMap with name cm-test-opt-del-407a44dc-2212-4c73-8902-acadbdb2f235
+STEP: Creating configMap with name cm-test-opt-upd-c30494db-58f5-4b3b-a4d3-d74b343de428
+STEP: Creating the pod
+STEP: Deleting configmap cm-test-opt-del-407a44dc-2212-4c73-8902-acadbdb2f235
+STEP: Updating configmap cm-test-opt-upd-c30494db-58f5-4b3b-a4d3-d74b343de428
+STEP: Creating configMap with name cm-test-opt-create-0959162c-e87b-4921-a23e-a76c9d0ba957
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:53:52.573: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-7618" for this suite.
+Oct 26 17:07:44.895: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-4137" for this suite.
 
-• [SLOW TEST:17.710 seconds]
-[sig-api-machinery] ResourceQuota
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should create a ResourceQuota and capture the life of a secret. [Conformance]
+• [SLOW TEST:6.688 seconds]
+[sig-storage] ConfigMap
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance]","total":280,"completed":241,"skipped":3951,"failed":0}
-SSSSSS
+{"msg":"PASSED [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":240,"skipped":3935,"failed":0}
+SSSSSSSSSSSSSSSSS
+------------------------------
+[sig-network] DNS 
+  should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] DNS
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:07:44.945: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename dns
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a test headless service
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-3620 A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-3620;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-3620 A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-3620;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-3620.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-3620.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-3620.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-3620.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-3620.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-3620.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-3620.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-3620.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-3620.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-3620.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-3620.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-3620.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3620.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 130.97.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.97.130_udp@PTR;check="$$(dig +tcp +noall +answer +search 130.97.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.97.130_tcp@PTR;sleep 1; done
+
+STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-3620 A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-3620;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-3620 A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-3620;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-3620.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-3620.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-3620.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-3620.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-3620.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-3620.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-3620.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-3620.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-3620.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-3620.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-3620.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-3620.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3620.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 130.97.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.97.130_udp@PTR;check="$$(dig +tcp +noall +answer +search 130.97.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.97.130_tcp@PTR;sleep 1; done
+
+STEP: creating a pod to probe DNS
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Oct 26 17:07:49.292: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.319: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.343: INFO: Unable to read wheezy_udp@dns-test-service.dns-3620 from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.366: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3620 from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.388: INFO: Unable to read wheezy_udp@dns-test-service.dns-3620.svc from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.414: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3620.svc from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.439: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3620.svc from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.463: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3620.svc from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.646: INFO: Unable to read jessie_udp@dns-test-service from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.685: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.709: INFO: Unable to read jessie_udp@dns-test-service.dns-3620 from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.742: INFO: Unable to read jessie_tcp@dns-test-service.dns-3620 from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.771: INFO: Unable to read jessie_udp@dns-test-service.dns-3620.svc from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.805: INFO: Unable to read jessie_tcp@dns-test-service.dns-3620.svc from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.837: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3620.svc from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:49.862: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3620.svc from pod dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29: the server could not find the requested resource (get pods dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29)
+Oct 26 17:07:50.013: INFO: Lookups using dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-3620 wheezy_tcp@dns-test-service.dns-3620 wheezy_udp@dns-test-service.dns-3620.svc wheezy_tcp@dns-test-service.dns-3620.svc wheezy_udp@_http._tcp.dns-test-service.dns-3620.svc wheezy_tcp@_http._tcp.dns-test-service.dns-3620.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-3620 jessie_tcp@dns-test-service.dns-3620 jessie_udp@dns-test-service.dns-3620.svc jessie_tcp@dns-test-service.dns-3620.svc jessie_udp@_http._tcp.dns-test-service.dns-3620.svc jessie_tcp@_http._tcp.dns-test-service.dns-3620.svc]
+
+Oct 26 17:07:55.832: INFO: DNS probes using dns-3620/dns-test-b490d67c-51f7-4a5c-a388-5cf4cdb41f29 succeeded
+
+STEP: deleting the pod
+STEP: deleting the test service
+STEP: deleting the test headless service
+[AfterEach] [sig-network] DNS
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:07:56.035: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-3620" for this suite.
+
+• [SLOW TEST:11.134 seconds]
+[sig-network] DNS
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]","total":280,"completed":241,"skipped":3952,"failed":0}
+SSSSSSS
+------------------------------
+[sig-storage] Projected secret 
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:07:56.082: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating projection with secret that has name projected-secret-test-9d4510da-0de1-4565-8f6f-5112a4e73f28
+STEP: Creating a pod to test consume secrets
+Oct 26 17:07:56.332: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-406ba670-6dba-4952-8660-a66206e2518e" in namespace "projected-1387" to be "success or failure"
+Oct 26 17:07:56.351: INFO: Pod "pod-projected-secrets-406ba670-6dba-4952-8660-a66206e2518e": Phase="Pending", Reason="", readiness=false. Elapsed: 18.38172ms
+Oct 26 17:07:58.370: INFO: Pod "pod-projected-secrets-406ba670-6dba-4952-8660-a66206e2518e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.037764344s
+Oct 26 17:08:00.386: INFO: Pod "pod-projected-secrets-406ba670-6dba-4952-8660-a66206e2518e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.054022164s
+STEP: Saw pod success
+Oct 26 17:08:00.386: INFO: Pod "pod-projected-secrets-406ba670-6dba-4952-8660-a66206e2518e" satisfied condition "success or failure"
+Oct 26 17:08:00.402: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-secrets-406ba670-6dba-4952-8660-a66206e2518e container projected-secret-volume-test: 
+STEP: delete the pod
+Oct 26 17:08:00.491: INFO: Waiting for pod pod-projected-secrets-406ba670-6dba-4952-8660-a66206e2518e to disappear
+Oct 26 17:08:00.505: INFO: Pod pod-projected-secrets-406ba670-6dba-4952-8660-a66206e2518e no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:08:00.505: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-1387" for this suite.
+•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":242,"skipped":3959,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-apps] Deployment 
   RecreateDeployment should delete old pods and create new ones [Conformance]
@@ -11994,377 +11491,425 @@ SSSSSS
 [BeforeEach] [sig-apps] Deployment
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:53:52.628: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:08:00.548: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename deployment
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-apps] Deployment
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
 [It] RecreateDeployment should delete old pods and create new ones [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:53:52.872: INFO: Creating deployment "test-recreate-deployment"
-Sep 21 16:53:52.896: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1
-Sep 21 16:53:52.960: INFO: deployment "test-recreate-deployment" doesn't have the required revision set
-Sep 21 16:53:54.999: INFO: Waiting deployment "test-recreate-deployment" to complete
-Sep 21 16:53:55.037: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736304032, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736304032, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736304033, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736304032, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-799c574856\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 16:53:57.061: INFO: Triggering a new rollout for deployment "test-recreate-deployment"
-Sep 21 16:53:57.113: INFO: Updating deployment test-recreate-deployment
-Sep 21 16:53:57.113: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods
+Oct 26 17:08:00.736: INFO: Creating deployment "test-recreate-deployment"
+Oct 26 17:08:00.766: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1
+Oct 26 17:08:00.805: INFO: deployment "test-recreate-deployment" doesn't have the required revision set
+Oct 26 17:08:02.838: INFO: Waiting deployment "test-recreate-deployment" to complete
+Oct 26 17:08:02.861: INFO: Triggering a new rollout for deployment "test-recreate-deployment"
+Oct 26 17:08:02.901: INFO: Updating deployment test-recreate-deployment
+Oct 26 17:08:02.901: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods
 [AfterEach] [sig-apps] Deployment
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
-Sep 21 16:53:57.389: INFO: Deployment "test-recreate-deployment":
-&Deployment{ObjectMeta:{test-recreate-deployment  deployment-6818 /apis/apps/v1/namespaces/deployment-6818/deployments/test-recreate-deployment 8be7cf57-8c59-405b-888d-e7f9ff08d2ae 89062 2 2020-09-21 16:53:52 +0000 UTC   map[name:sample-pod-3] map[deployment.kubernetes.io/revision:2] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod-3] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc005b3e098  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2020-09-21 16:53:57 +0000 UTC,LastTransitionTime:2020-09-21 16:53:57 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "test-recreate-deployment-5f94c574ff" is progressing.,LastUpdateTime:2020-09-21 16:53:57 +0000 UTC,LastTransitionTime:2020-09-21 16:53:52 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},}
-
-Sep 21 16:53:57.410: INFO: New ReplicaSet "test-recreate-deployment-5f94c574ff" of Deployment "test-recreate-deployment":
-&ReplicaSet{ObjectMeta:{test-recreate-deployment-5f94c574ff  deployment-6818 /apis/apps/v1/namespaces/deployment-6818/replicasets/test-recreate-deployment-5f94c574ff ddc02b22-15c5-4462-8b02-76b13071b1fe 89060 1 2020-09-21 16:53:57 +0000 UTC   map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-recreate-deployment 8be7cf57-8c59-405b-888d-e7f9ff08d2ae 0xc00961fd87 0xc00961fd88}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 5f94c574ff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00961fde8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
-Sep 21 16:53:57.410: INFO: All old ReplicaSets of Deployment "test-recreate-deployment":
-Sep 21 16:53:57.410: INFO: &ReplicaSet{ObjectMeta:{test-recreate-deployment-799c574856  deployment-6818 /apis/apps/v1/namespaces/deployment-6818/replicasets/test-recreate-deployment-799c574856 d4f8d75e-a7ed-4a53-9885-569d0bf0beac 89049 2 2020-09-21 16:53:52 +0000 UTC   map[name:sample-pod-3 pod-template-hash:799c574856] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-recreate-deployment 8be7cf57-8c59-405b-888d-e7f9ff08d2ae 0xc00961fe57 0xc00961fe58}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 799c574856,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod-3 pod-template-hash:799c574856] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00961fec8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
-Sep 21 16:53:57.435: INFO: Pod "test-recreate-deployment-5f94c574ff-hvdpx" is not available:
-&Pod{ObjectMeta:{test-recreate-deployment-5f94c574ff-hvdpx test-recreate-deployment-5f94c574ff- deployment-6818 /api/v1/namespaces/deployment-6818/pods/test-recreate-deployment-5f94c574ff-hvdpx ea95a085-e841-4993-b690-2bdc4b800a9b 89063 0 2020-09-21 16:53:57 +0000 UTC   map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[openshift.io/scc:privileged] [{apps/v1 ReplicaSet test-recreate-deployment-5f94c574ff ddc02b22-15c5-4462-8b02-76b13071b1fe 0xc005b3e467 0xc005b3e468}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-94lwt,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-94lwt,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-94lwt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.241.51.147,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-xm56h,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:53:57 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:53:57 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:53:57 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-09-21 16:53:57 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.241.51.147,PodIP:,StartTime:2020-09-21 16:53:57 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Oct 26 17:08:03.122: INFO: Deployment "test-recreate-deployment":
+&Deployment{ObjectMeta:{test-recreate-deployment  deployment-246 /apis/apps/v1/namespaces/deployment-246/deployments/test-recreate-deployment e2d6451e-5d03-4dfd-a0b1-c86a6a47bd27 86888 2 2020-10-26 17:08:00 +0000 UTC   map[name:sample-pod-3] map[deployment.kubernetes.io/revision:2] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod-3] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc004a00998  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2020-10-26 17:08:03 +0000 UTC,LastTransitionTime:2020-10-26 17:08:03 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "test-recreate-deployment-5f94c574ff" is progressing.,LastUpdateTime:2020-10-26 17:08:03 +0000 UTC,LastTransitionTime:2020-10-26 17:08:00 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},}
+
+Oct 26 17:08:03.138: INFO: New ReplicaSet "test-recreate-deployment-5f94c574ff" of Deployment "test-recreate-deployment":
+&ReplicaSet{ObjectMeta:{test-recreate-deployment-5f94c574ff  deployment-246 /apis/apps/v1/namespaces/deployment-246/replicasets/test-recreate-deployment-5f94c574ff 3ce09160-f10e-4112-9cb2-6ce1bb471bff 86886 1 2020-10-26 17:08:02 +0000 UTC   map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-recreate-deployment e2d6451e-5d03-4dfd-a0b1-c86a6a47bd27 0xc000d84ea7 0xc000d84ea8}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 5f94c574ff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc000d84f08  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
+Oct 26 17:08:03.138: INFO: All old ReplicaSets of Deployment "test-recreate-deployment":
+Oct 26 17:08:03.138: INFO: &ReplicaSet{ObjectMeta:{test-recreate-deployment-799c574856  deployment-246 /apis/apps/v1/namespaces/deployment-246/replicasets/test-recreate-deployment-799c574856 843c4fbc-c718-4d45-8034-4e2b52fa28f5 86877 2 2020-10-26 17:08:00 +0000 UTC   map[name:sample-pod-3 pod-template-hash:799c574856] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-recreate-deployment e2d6451e-5d03-4dfd-a0b1-c86a6a47bd27 0xc000d84f77 0xc000d84f78}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 799c574856,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod-3 pod-template-hash:799c574856] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc000d84fe8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
+Oct 26 17:08:03.154: INFO: Pod "test-recreate-deployment-5f94c574ff-z9bpr" is not available:
+&Pod{ObjectMeta:{test-recreate-deployment-5f94c574ff-z9bpr test-recreate-deployment-5f94c574ff- deployment-246 /api/v1/namespaces/deployment-246/pods/test-recreate-deployment-5f94c574ff-z9bpr 0a8a9da3-937f-4ae6-8d85-2cf547e84b61 86889 0 2020-10-26 17:08:03 +0000 UTC   map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[openshift.io/scc:privileged] [{apps/v1 ReplicaSet test-recreate-deployment-5f94c574ff 3ce09160-f10e-4112-9cb2-6ce1bb471bff 0xc000d857b7 0xc000d857b8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qlvsn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qlvsn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qlvsn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.74,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-mc9rd,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:08:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:08:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:08:03 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:08:03 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.74,PodIP:,StartTime:2020-10-26 17:08:03 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
 [AfterEach] [sig-apps] Deployment
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:53:57.435: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "deployment-6818" for this suite.
-•{"msg":"PASSED [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance]","total":280,"completed":242,"skipped":3957,"failed":0}
-SSSSSSSSSSSSSS
+Oct 26 17:08:03.154: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-246" for this suite.
+•{"msg":"PASSED [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance]","total":280,"completed":243,"skipped":3998,"failed":0}
+SSSSSSSSSSS
+------------------------------
+[sig-node] ConfigMap 
+  should be consumable via the environment [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-node] ConfigMap
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:08:03.194: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename configmap
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable via the environment [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap configmap-3471/configmap-test-83980493-4fec-4297-aa04-de3b051f6ad5
+STEP: Creating a pod to test consume configMaps
+Oct 26 17:08:03.427: INFO: Waiting up to 5m0s for pod "pod-configmaps-41981c97-bd28-4c6d-b458-ac77b9dc82e3" in namespace "configmap-3471" to be "success or failure"
+Oct 26 17:08:03.446: INFO: Pod "pod-configmaps-41981c97-bd28-4c6d-b458-ac77b9dc82e3": Phase="Pending", Reason="", readiness=false. Elapsed: 19.273091ms
+Oct 26 17:08:05.463: INFO: Pod "pod-configmaps-41981c97-bd28-4c6d-b458-ac77b9dc82e3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.036219469s
+Oct 26 17:08:07.479: INFO: Pod "pod-configmaps-41981c97-bd28-4c6d-b458-ac77b9dc82e3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.052475171s
+STEP: Saw pod success
+Oct 26 17:08:07.480: INFO: Pod "pod-configmaps-41981c97-bd28-4c6d-b458-ac77b9dc82e3" satisfied condition "success or failure"
+Oct 26 17:08:07.497: INFO: Trying to get logs from node 10.72.119.74 pod pod-configmaps-41981c97-bd28-4c6d-b458-ac77b9dc82e3 container env-test: 
+STEP: delete the pod
+Oct 26 17:08:07.593: INFO: Waiting for pod pod-configmaps-41981c97-bd28-4c6d-b458-ac77b9dc82e3 to disappear
+Oct 26 17:08:07.609: INFO: Pod pod-configmaps-41981c97-bd28-4c6d-b458-ac77b9dc82e3 no longer exists
+[AfterEach] [sig-node] ConfigMap
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:08:07.609: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-3471" for this suite.
+•{"msg":"PASSED [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance]","total":280,"completed":244,"skipped":4009,"failed":0}
+SSSSSSSS
+------------------------------
+[k8s.io] Pods 
+  should support retrieving logs from the container over websockets [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:08:07.650: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename pods
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Pods
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
+[It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Oct 26 17:08:07.779: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+[AfterEach] [k8s.io] Pods
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:08:09.944: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-2372" for this suite.
+•{"msg":"PASSED [k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]","total":280,"completed":245,"skipped":4017,"failed":0}
+SSSSSSSSSSSS
+------------------------------
+[sig-storage] ConfigMap 
+  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:08:09.987: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename configmap
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name configmap-test-volume-1081f092-c3f9-49d3-878b-1b7e4a8b7de1
+STEP: Creating a pod to test consume configMaps
+Oct 26 17:08:10.206: INFO: Waiting up to 5m0s for pod "pod-configmaps-718a80ee-c4ab-4672-bd08-0408e0cb2f20" in namespace "configmap-267" to be "success or failure"
+Oct 26 17:08:10.229: INFO: Pod "pod-configmaps-718a80ee-c4ab-4672-bd08-0408e0cb2f20": Phase="Pending", Reason="", readiness=false. Elapsed: 22.721074ms
+Oct 26 17:08:12.252: INFO: Pod "pod-configmaps-718a80ee-c4ab-4672-bd08-0408e0cb2f20": Phase="Pending", Reason="", readiness=false. Elapsed: 2.045266413s
+Oct 26 17:08:14.274: INFO: Pod "pod-configmaps-718a80ee-c4ab-4672-bd08-0408e0cb2f20": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.06717937s
+STEP: Saw pod success
+Oct 26 17:08:14.274: INFO: Pod "pod-configmaps-718a80ee-c4ab-4672-bd08-0408e0cb2f20" satisfied condition "success or failure"
+Oct 26 17:08:14.292: INFO: Trying to get logs from node 10.72.119.74 pod pod-configmaps-718a80ee-c4ab-4672-bd08-0408e0cb2f20 container configmap-volume-test: 
+STEP: delete the pod
+Oct 26 17:08:14.392: INFO: Waiting for pod pod-configmaps-718a80ee-c4ab-4672-bd08-0408e0cb2f20 to disappear
+Oct 26 17:08:14.407: INFO: Pod pod-configmaps-718a80ee-c4ab-4672-bd08-0408e0cb2f20 no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:08:14.407: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-267" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":280,"completed":246,"skipped":4029,"failed":0}
+
 ------------------------------
 [sig-network] DNS 
-  should provide DNS for ExternalName services [Conformance]
+  should provide DNS for services  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-network] DNS
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:53:57.503: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:08:14.448: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename dns
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide DNS for ExternalName services [Conformance]
+[It] should provide DNS for services  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a test externalName service
-STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-9869.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-9869.svc.cluster.local; sleep 1; done
+STEP: Creating a test headless service
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-5373.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-5373.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-5373.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-5373.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-5373.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-5373.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-5373.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-5373.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-5373.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-5373.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-5373.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-5373.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5373.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 98.247.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.247.98_udp@PTR;check="$$(dig +tcp +noall +answer +search 98.247.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.247.98_tcp@PTR;sleep 1; done
 
-STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-9869.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-9869.svc.cluster.local; sleep 1; done
+STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-5373.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-5373.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-5373.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-5373.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-5373.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-5373.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-5373.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-5373.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-5373.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-5373.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-5373.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-5373.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5373.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 98.247.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.247.98_udp@PTR;check="$$(dig +tcp +noall +answer +search 98.247.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.247.98_tcp@PTR;sleep 1; done
 
 STEP: creating a pod to probe DNS
 STEP: submitting the pod to kubernetes
 STEP: retrieving the pod
 STEP: looking for the results for each expected name from probers
-Sep 21 16:54:02.007: INFO: DNS probes using dns-test-21886b60-40d0-4bcb-84e1-61ae9f2feddd succeeded
-
-STEP: deleting the pod
-STEP: changing the externalName to bar.example.com
-STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-9869.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-9869.svc.cluster.local; sleep 1; done
-
-STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-9869.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-9869.svc.cluster.local; sleep 1; done
-
-STEP: creating a second pod to probe DNS
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Sep 21 16:54:04.412: INFO: File wheezy_udp@dns-test-service-3.dns-9869.svc.cluster.local from pod  dns-9869/dns-test-e65569d5-a209-42db-96da-7762c2809ce1 contains 'foo.example.com.
-' instead of 'bar.example.com.'
-Sep 21 16:54:04.450: INFO: File jessie_udp@dns-test-service-3.dns-9869.svc.cluster.local from pod  dns-9869/dns-test-e65569d5-a209-42db-96da-7762c2809ce1 contains 'foo.example.com.
-' instead of 'bar.example.com.'
-Sep 21 16:54:04.450: INFO: Lookups using dns-9869/dns-test-e65569d5-a209-42db-96da-7762c2809ce1 failed for: [wheezy_udp@dns-test-service-3.dns-9869.svc.cluster.local jessie_udp@dns-test-service-3.dns-9869.svc.cluster.local]
-
-Sep 21 16:54:09.530: INFO: DNS probes using dns-test-e65569d5-a209-42db-96da-7762c2809ce1 succeeded
-
-STEP: deleting the pod
-STEP: changing the service to type=ClusterIP
-STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-9869.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-9869.svc.cluster.local; sleep 1; done
-
-STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-9869.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-9869.svc.cluster.local; sleep 1; done
+Oct 26 17:08:19.491: INFO: Unable to read wheezy_udp@dns-test-service.dns-5373.svc.cluster.local from pod dns-5373/dns-test-db9e7086-1aa6-426d-9f45-597f7ed63612: the server could not find the requested resource (get pods dns-test-db9e7086-1aa6-426d-9f45-597f7ed63612)
+Oct 26 17:08:19.538: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-5373.svc.cluster.local from pod dns-5373/dns-test-db9e7086-1aa6-426d-9f45-597f7ed63612: the server could not find the requested resource (get pods dns-test-db9e7086-1aa6-426d-9f45-597f7ed63612)
+Oct 26 17:08:19.772: INFO: Unable to read jessie_udp@dns-test-service.dns-5373.svc.cluster.local from pod dns-5373/dns-test-db9e7086-1aa6-426d-9f45-597f7ed63612: the server could not find the requested resource (get pods dns-test-db9e7086-1aa6-426d-9f45-597f7ed63612)
+Oct 26 17:08:19.994: INFO: Lookups using dns-5373/dns-test-db9e7086-1aa6-426d-9f45-597f7ed63612 failed for: [wheezy_udp@dns-test-service.dns-5373.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-5373.svc.cluster.local jessie_udp@dns-test-service.dns-5373.svc.cluster.local]
 
-STEP: creating a third pod to probe DNS
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Sep 21 16:54:13.943: INFO: DNS probes using dns-test-64b84f08-1f97-409b-89a3-df98c7d5aed4 succeeded
+Oct 26 17:08:25.512: INFO: DNS probes using dns-5373/dns-test-db9e7086-1aa6-426d-9f45-597f7ed63612 succeeded
 
 STEP: deleting the pod
-STEP: deleting the test externalName service
+STEP: deleting the test service
+STEP: deleting the test headless service
 [AfterEach] [sig-network] DNS
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:54:14.345: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-9869" for this suite.
+Oct 26 17:08:25.718: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-5373" for this suite.
 
-• [SLOW TEST:16.904 seconds]
+• [SLOW TEST:11.312 seconds]
 [sig-network] DNS
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should provide DNS for ExternalName services [Conformance]
+  should provide DNS for services  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] DNS should provide DNS for ExternalName services [Conformance]","total":280,"completed":243,"skipped":3971,"failed":0}
-SSS
+{"msg":"PASSED [sig-network] DNS should provide DNS for services  [Conformance]","total":280,"completed":247,"skipped":4029,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-node] Downward API 
-  should provide host IP as an env var [NodeConformance] [Conformance]
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:08:25.761: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename configmap
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name configmap-test-volume-8c8d6382-e92f-4252-97c0-185abd087e80
+STEP: Creating a pod to test consume configMaps
+Oct 26 17:08:25.978: INFO: Waiting up to 5m0s for pod "pod-configmaps-880577e0-15c1-4c0e-9114-de80ba83b63a" in namespace "configmap-5399" to be "success or failure"
+Oct 26 17:08:25.995: INFO: Pod "pod-configmaps-880577e0-15c1-4c0e-9114-de80ba83b63a": Phase="Pending", Reason="", readiness=false. Elapsed: 16.569254ms
+Oct 26 17:08:28.011: INFO: Pod "pod-configmaps-880577e0-15c1-4c0e-9114-de80ba83b63a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.032483843s
+STEP: Saw pod success
+Oct 26 17:08:28.011: INFO: Pod "pod-configmaps-880577e0-15c1-4c0e-9114-de80ba83b63a" satisfied condition "success or failure"
+Oct 26 17:08:28.029: INFO: Trying to get logs from node 10.72.119.74 pod pod-configmaps-880577e0-15c1-4c0e-9114-de80ba83b63a container configmap-volume-test: 
+STEP: delete the pod
+Oct 26 17:08:28.119: INFO: Waiting for pod pod-configmaps-880577e0-15c1-4c0e-9114-de80ba83b63a to disappear
+Oct 26 17:08:28.134: INFO: Pod pod-configmaps-880577e0-15c1-4c0e-9114-de80ba83b63a no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:08:28.134: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-5399" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":248,"skipped":4086,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected secret 
+  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:08:28.175: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating projection with secret that has name projected-secret-test-c1980e95-5309-4c6f-9584-781c042724fd
+STEP: Creating a pod to test consume secrets
+Oct 26 17:08:28.410: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-716b5c80-deb5-4415-832d-669a4e5936d8" in namespace "projected-9815" to be "success or failure"
+Oct 26 17:08:28.439: INFO: Pod "pod-projected-secrets-716b5c80-deb5-4415-832d-669a4e5936d8": Phase="Pending", Reason="", readiness=false. Elapsed: 28.765034ms
+Oct 26 17:08:30.457: INFO: Pod "pod-projected-secrets-716b5c80-deb5-4415-832d-669a4e5936d8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.046939079s
+Oct 26 17:08:32.473: INFO: Pod "pod-projected-secrets-716b5c80-deb5-4415-832d-669a4e5936d8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.063104756s
+STEP: Saw pod success
+Oct 26 17:08:32.473: INFO: Pod "pod-projected-secrets-716b5c80-deb5-4415-832d-669a4e5936d8" satisfied condition "success or failure"
+Oct 26 17:08:32.496: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-secrets-716b5c80-deb5-4415-832d-669a4e5936d8 container projected-secret-volume-test: 
+STEP: delete the pod
+Oct 26 17:08:32.587: INFO: Waiting for pod pod-projected-secrets-716b5c80-deb5-4415-832d-669a4e5936d8 to disappear
+Oct 26 17:08:32.608: INFO: Pod pod-projected-secrets-716b5c80-deb5-4415-832d-669a4e5936d8 no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:08:32.609: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9815" for this suite.
+•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":249,"skipped":4117,"failed":0}
+SSSSSSS
+------------------------------
+[sig-storage] Secrets 
+  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Secrets
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:08:32.651: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename secrets
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating secret with name secret-test-ba341073-95c1-412a-b2ec-fda91021c3db
+STEP: Creating a pod to test consume secrets
+Oct 26 17:08:32.928: INFO: Waiting up to 5m0s for pod "pod-secrets-1a2b791b-fe9b-45d8-9d2b-3adf2767b207" in namespace "secrets-6376" to be "success or failure"
+Oct 26 17:08:32.945: INFO: Pod "pod-secrets-1a2b791b-fe9b-45d8-9d2b-3adf2767b207": Phase="Pending", Reason="", readiness=false. Elapsed: 15.974698ms
+Oct 26 17:08:34.962: INFO: Pod "pod-secrets-1a2b791b-fe9b-45d8-9d2b-3adf2767b207": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.032983594s
+STEP: Saw pod success
+Oct 26 17:08:34.962: INFO: Pod "pod-secrets-1a2b791b-fe9b-45d8-9d2b-3adf2767b207" satisfied condition "success or failure"
+Oct 26 17:08:34.986: INFO: Trying to get logs from node 10.72.119.74 pod pod-secrets-1a2b791b-fe9b-45d8-9d2b-3adf2767b207 container secret-volume-test: 
+STEP: delete the pod
+Oct 26 17:08:35.075: INFO: Waiting for pod pod-secrets-1a2b791b-fe9b-45d8-9d2b-3adf2767b207 to disappear
+Oct 26 17:08:35.089: INFO: Pod pod-secrets-1a2b791b-fe9b-45d8-9d2b-3adf2767b207 no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:08:35.089: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-6376" for this suite.
+•{"msg":"PASSED [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":280,"completed":250,"skipped":4124,"failed":0}
+SSSS
+------------------------------
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] Downward API
+[BeforeEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:54:14.407: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
+Oct 26 17:08:35.140: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide host IP as an env var [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward api env vars
-Sep 21 16:54:14.723: INFO: Waiting up to 5m0s for pod "downward-api-48081ce9-021a-4a48-8a37-1b694a48cf03" in namespace "downward-api-4440" to be "success or failure"
-Sep 21 16:54:14.742: INFO: Pod "downward-api-48081ce9-021a-4a48-8a37-1b694a48cf03": Phase="Pending", Reason="", readiness=false. Elapsed: 18.67302ms
-Sep 21 16:54:16.766: INFO: Pod "downward-api-48081ce9-021a-4a48-8a37-1b694a48cf03": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.042075544s
+STEP: Creating configMap with name projected-configmap-test-volume-map-1465bc87-217a-4e18-a94d-a45f4dd4d1c6
+STEP: Creating a pod to test consume configMaps
+Oct 26 17:08:36.363: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-48985acd-2018-43e1-948e-8f87ee7b7d41" in namespace "projected-9903" to be "success or failure"
+Oct 26 17:08:36.379: INFO: Pod "pod-projected-configmaps-48985acd-2018-43e1-948e-8f87ee7b7d41": Phase="Pending", Reason="", readiness=false. Elapsed: 16.299457ms
+Oct 26 17:08:38.394: INFO: Pod "pod-projected-configmaps-48985acd-2018-43e1-948e-8f87ee7b7d41": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03100417s
+Oct 26 17:08:40.411: INFO: Pod "pod-projected-configmaps-48985acd-2018-43e1-948e-8f87ee7b7d41": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.047392913s
 STEP: Saw pod success
-Sep 21 16:54:16.766: INFO: Pod "downward-api-48081ce9-021a-4a48-8a37-1b694a48cf03" satisfied condition "success or failure"
-Sep 21 16:54:16.787: INFO: Trying to get logs from node 10.241.51.147 pod downward-api-48081ce9-021a-4a48-8a37-1b694a48cf03 container dapi-container: 
+Oct 26 17:08:40.411: INFO: Pod "pod-projected-configmaps-48985acd-2018-43e1-948e-8f87ee7b7d41" satisfied condition "success or failure"
+Oct 26 17:08:40.425: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-configmaps-48985acd-2018-43e1-948e-8f87ee7b7d41 container projected-configmap-volume-test: 
 STEP: delete the pod
-Sep 21 16:54:16.923: INFO: Waiting for pod downward-api-48081ce9-021a-4a48-8a37-1b694a48cf03 to disappear
-Sep 21 16:54:16.944: INFO: Pod downward-api-48081ce9-021a-4a48-8a37-1b694a48cf03 no longer exists
-[AfterEach] [sig-node] Downward API
+Oct 26 17:08:40.511: INFO: Waiting for pod pod-projected-configmaps-48985acd-2018-43e1-948e-8f87ee7b7d41 to disappear
+Oct 26 17:08:40.525: INFO: Pod pod-projected-configmaps-48985acd-2018-43e1-948e-8f87ee7b7d41 no longer exists
+[AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:54:16.944: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-4440" for this suite.
-•{"msg":"PASSED [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance]","total":280,"completed":244,"skipped":3974,"failed":0}
+Oct 26 17:08:40.525: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9903" for this suite.
+
+• [SLOW TEST:5.426 seconds]
+[sig-storage] Projected configMap
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
+  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":251,"skipped":4128,"failed":0}
 SSSSSSSS
 ------------------------------
-[sig-storage] Projected secret 
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should mutate configmap [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:54:17.003: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 17:08:40.567: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Oct 26 17:08:41.392: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 17:08:43.485: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328921, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328921, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328921, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328921, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Oct 26 17:08:46.532: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should mutate configmap [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:54:17.230: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
-STEP: Creating secret with name s-test-opt-del-6f734163-413d-40ab-9950-2082bc56e196
-STEP: Creating secret with name s-test-opt-upd-edb8ed72-84f8-4934-98c9-3d3e12aba73f
-STEP: Creating the pod
-STEP: Deleting secret s-test-opt-del-6f734163-413d-40ab-9950-2082bc56e196
-STEP: Updating secret s-test-opt-upd-edb8ed72-84f8-4934-98c9-3d3e12aba73f
-STEP: Creating secret with name s-test-opt-create-0f61a410-0871-49d6-8a55-f134c8d565b2
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] Projected secret
+STEP: Registering the mutating configmap webhook via the AdmissionRegistration API
+STEP: create a configmap that should be updated by the webhook
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:55:50.414: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-3697" for this suite.
+Oct 26 17:08:46.665: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-1880" for this suite.
+STEP: Destroying namespace "webhook-1880-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:93.456 seconds]
-[sig-storage] Projected secret
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
+• [SLOW TEST:6.361 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should mutate configmap [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":245,"skipped":3982,"failed":0}
-SSSS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance]","total":280,"completed":252,"skipped":4136,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] Daemon set [Serial] 
-  should rollback without unnecessary restarts [Conformance]
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Daemon set [Serial]
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:55:50.459: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename daemonsets
+Oct 26 17:08:46.928: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
-[It] should rollback without unnecessary restarts [Conformance]
+[It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:55:50.806: INFO: Create a RollingUpdate DaemonSet
-Sep 21 16:55:50.826: INFO: Check that daemon pods launch on every node of the cluster
-Sep 21 16:55:50.876: INFO: Number of nodes with available pods: 0
-Sep 21 16:55:50.876: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:55:51.917: INFO: Number of nodes with available pods: 0
-Sep 21 16:55:51.917: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:55:52.916: INFO: Number of nodes with available pods: 2
-Sep 21 16:55:52.916: INFO: Node 10.241.51.150 is running more than one daemon pod
-Sep 21 16:55:53.915: INFO: Number of nodes with available pods: 3
-Sep 21 16:55:53.915: INFO: Number of running nodes: 3, number of available pods: 3
-Sep 21 16:55:53.915: INFO: Update the DaemonSet to trigger a rollout
-Sep 21 16:55:54.053: INFO: Updating DaemonSet daemon-set
-Sep 21 16:56:08.142: INFO: Roll back the DaemonSet before rollout is complete
-Sep 21 16:56:08.187: INFO: Updating DaemonSet daemon-set
-Sep 21 16:56:08.187: INFO: Make sure DaemonSet rollback is complete
-Sep 21 16:56:08.208: INFO: Wrong image for pod: daemon-set-h7kvt. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent.
-Sep 21 16:56:08.208: INFO: Pod daemon-set-h7kvt is not available
-Sep 21 16:56:09.251: INFO: Wrong image for pod: daemon-set-h7kvt. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent.
-Sep 21 16:56:09.251: INFO: Pod daemon-set-h7kvt is not available
-Sep 21 16:56:10.252: INFO: Pod daemon-set-5jwzb is not available
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-5045, will wait for the garbage collector to delete the pods
-Sep 21 16:56:10.439: INFO: Deleting DaemonSet.extensions daemon-set took: 41.484248ms
-Sep 21 16:56:10.540: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.380708ms
-Sep 21 16:56:19.667: INFO: Number of nodes with available pods: 0
-Sep 21 16:56:19.667: INFO: Number of running nodes: 0, number of available pods: 0
-Sep 21 16:56:19.687: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-5045/daemonsets","resourceVersion":"90196"},"items":null}
-
-Sep 21 16:56:19.710: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-5045/pods","resourceVersion":"90196"},"items":null}
-
-[AfterEach] [sig-apps] Daemon set [Serial]
+STEP: Creating a pod to test emptydir 0666 on tmpfs
+Oct 26 17:08:48.152: INFO: Waiting up to 5m0s for pod "pod-a80ff234-ecbe-40a8-aa7a-286bb2b9a572" in namespace "emptydir-646" to be "success or failure"
+Oct 26 17:08:48.171: INFO: Pod "pod-a80ff234-ecbe-40a8-aa7a-286bb2b9a572": Phase="Pending", Reason="", readiness=false. Elapsed: 18.142096ms
+Oct 26 17:08:50.187: INFO: Pod "pod-a80ff234-ecbe-40a8-aa7a-286bb2b9a572": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.034384472s
+STEP: Saw pod success
+Oct 26 17:08:50.187: INFO: Pod "pod-a80ff234-ecbe-40a8-aa7a-286bb2b9a572" satisfied condition "success or failure"
+Oct 26 17:08:50.202: INFO: Trying to get logs from node 10.72.119.74 pod pod-a80ff234-ecbe-40a8-aa7a-286bb2b9a572 container test-container: 
+STEP: delete the pod
+Oct 26 17:08:50.291: INFO: Waiting for pod pod-a80ff234-ecbe-40a8-aa7a-286bb2b9a572 to disappear
+Oct 26 17:08:50.306: INFO: Pod pod-a80ff234-ecbe-40a8-aa7a-286bb2b9a572 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:56:19.800: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "daemonsets-5045" for this suite.
-
-• [SLOW TEST:29.403 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should rollback without unnecessary restarts [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance]","total":280,"completed":246,"skipped":3986,"failed":0}
-SSSSSSSSSSSSSSSSSSS
+Oct 26 17:08:50.307: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-646" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":253,"skipped":4176,"failed":0}
+S
 ------------------------------
-[sig-apps] Daemon set [Serial] 
-  should update pod when spec was updated and update strategy is RollingUpdate [Conformance]
+[k8s.io] Variable Expansion 
+  should allow composing env vars into new env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Daemon set [Serial]
+[BeforeEach] [k8s.io] Variable Expansion
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:56:19.862: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename daemonsets
+Oct 26 17:08:50.347: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename var-expansion
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
-[It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]
+[It] should allow composing env vars into new env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:56:20.454: INFO: Creating simple daemon set daemon-set
-STEP: Check that daemon pods launch on every node of the cluster.
-Sep 21 16:56:20.536: INFO: Number of nodes with available pods: 0
-Sep 21 16:56:20.536: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:56:21.590: INFO: Number of nodes with available pods: 0
-Sep 21 16:56:21.590: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:56:22.587: INFO: Number of nodes with available pods: 1
-Sep 21 16:56:22.587: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:56:23.579: INFO: Number of nodes with available pods: 3
-Sep 21 16:56:23.579: INFO: Number of running nodes: 3, number of available pods: 3
-STEP: Update daemon pods image.
-STEP: Check that daemon pods images are updated.
-Sep 21 16:56:23.784: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:23.784: INFO: Wrong image for pod: daemon-set-dnppx. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:23.784: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:24.840: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:24.840: INFO: Wrong image for pod: daemon-set-dnppx. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:24.840: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:25.836: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:25.836: INFO: Wrong image for pod: daemon-set-dnppx. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:25.836: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:26.845: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:26.845: INFO: Wrong image for pod: daemon-set-dnppx. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:26.845: INFO: Pod daemon-set-dnppx is not available
-Sep 21 16:56:26.845: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:27.835: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:27.836: INFO: Wrong image for pod: daemon-set-dnppx. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:27.836: INFO: Pod daemon-set-dnppx is not available
-Sep 21 16:56:27.836: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:28.837: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:28.837: INFO: Wrong image for pod: daemon-set-dnppx. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:28.837: INFO: Pod daemon-set-dnppx is not available
-Sep 21 16:56:28.837: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:29.838: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:29.838: INFO: Pod daemon-set-88stx is not available
-Sep 21 16:56:29.838: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:30.845: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:30.845: INFO: Pod daemon-set-88stx is not available
-Sep 21 16:56:30.845: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:32.049: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:32.049: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:32.844: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:32.844: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:33.836: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:33.836: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:33.836: INFO: Pod daemon-set-hcqq4 is not available
-Sep 21 16:56:34.838: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:34.838: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:34.838: INFO: Pod daemon-set-hcqq4 is not available
-Sep 21 16:56:35.835: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:35.835: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:35.835: INFO: Pod daemon-set-hcqq4 is not available
-Sep 21 16:56:36.835: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:36.835: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:36.835: INFO: Pod daemon-set-hcqq4 is not available
-Sep 21 16:56:37.841: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:37.841: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:37.841: INFO: Pod daemon-set-hcqq4 is not available
-Sep 21 16:56:38.836: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:38.836: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:38.836: INFO: Pod daemon-set-hcqq4 is not available
-Sep 21 16:56:39.834: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:39.834: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:39.834: INFO: Pod daemon-set-hcqq4 is not available
-Sep 21 16:56:40.835: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:40.835: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:40.835: INFO: Pod daemon-set-hcqq4 is not available
-Sep 21 16:56:41.831: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:41.831: INFO: Wrong image for pod: daemon-set-hcqq4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:41.831: INFO: Pod daemon-set-hcqq4 is not available
-Sep 21 16:56:42.843: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:42.843: INFO: Pod daemon-set-xjxqb is not available
-Sep 21 16:56:43.833: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:43.833: INFO: Pod daemon-set-xjxqb is not available
-Sep 21 16:56:44.832: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:45.835: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:46.837: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:46.837: INFO: Pod daemon-set-4b8pq is not available
-Sep 21 16:56:47.842: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:47.842: INFO: Pod daemon-set-4b8pq is not available
-Sep 21 16:56:48.836: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:48.836: INFO: Pod daemon-set-4b8pq is not available
-Sep 21 16:56:49.838: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:49.838: INFO: Pod daemon-set-4b8pq is not available
-Sep 21 16:56:50.838: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:50.838: INFO: Pod daemon-set-4b8pq is not available
-Sep 21 16:56:51.841: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:51.841: INFO: Pod daemon-set-4b8pq is not available
-Sep 21 16:56:52.836: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:52.836: INFO: Pod daemon-set-4b8pq is not available
-Sep 21 16:56:53.834: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:53.834: INFO: Pod daemon-set-4b8pq is not available
-Sep 21 16:56:54.833: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:54.833: INFO: Pod daemon-set-4b8pq is not available
-Sep 21 16:56:55.833: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:55.833: INFO: Pod daemon-set-4b8pq is not available
-Sep 21 16:56:56.836: INFO: Wrong image for pod: daemon-set-4b8pq. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Sep 21 16:56:56.836: INFO: Pod daemon-set-4b8pq is not available
-Sep 21 16:56:57.836: INFO: Pod daemon-set-lxxtp is not available
-STEP: Check that daemon pods are still running on every node of the cluster.
-Sep 21 16:56:57.903: INFO: Number of nodes with available pods: 2
-Sep 21 16:56:57.903: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:56:58.951: INFO: Number of nodes with available pods: 3
-Sep 21 16:56:58.951: INFO: Number of running nodes: 3, number of available pods: 3
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-2528, will wait for the garbage collector to delete the pods
-Sep 21 16:56:59.143: INFO: Deleting DaemonSet.extensions daemon-set took: 41.821204ms
-Sep 21 16:56:59.643: INFO: Terminating DaemonSet.extensions daemon-set pods took: 500.335765ms
-Sep 21 16:57:12.265: INFO: Number of nodes with available pods: 0
-Sep 21 16:57:12.265: INFO: Number of running nodes: 0, number of available pods: 0
-Sep 21 16:57:12.281: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-2528/daemonsets","resourceVersion":"90658"},"items":null}
-
-Sep 21 16:57:12.303: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-2528/pods","resourceVersion":"90658"},"items":null}
-
-[AfterEach] [sig-apps] Daemon set [Serial]
+STEP: Creating a pod to test env composition
+Oct 26 17:08:50.565: INFO: Waiting up to 5m0s for pod "var-expansion-ef4d3dbf-6088-45b1-8bca-243b72ce996d" in namespace "var-expansion-1041" to be "success or failure"
+Oct 26 17:08:50.584: INFO: Pod "var-expansion-ef4d3dbf-6088-45b1-8bca-243b72ce996d": Phase="Pending", Reason="", readiness=false. Elapsed: 19.260597ms
+Oct 26 17:08:52.599: INFO: Pod "var-expansion-ef4d3dbf-6088-45b1-8bca-243b72ce996d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034456652s
+Oct 26 17:08:54.616: INFO: Pod "var-expansion-ef4d3dbf-6088-45b1-8bca-243b72ce996d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.051122858s
+STEP: Saw pod success
+Oct 26 17:08:54.616: INFO: Pod "var-expansion-ef4d3dbf-6088-45b1-8bca-243b72ce996d" satisfied condition "success or failure"
+Oct 26 17:08:54.634: INFO: Trying to get logs from node 10.72.119.74 pod var-expansion-ef4d3dbf-6088-45b1-8bca-243b72ce996d container dapi-container: 
+STEP: delete the pod
+Oct 26 17:08:54.743: INFO: Waiting for pod var-expansion-ef4d3dbf-6088-45b1-8bca-243b72ce996d to disappear
+Oct 26 17:08:54.759: INFO: Pod var-expansion-ef4d3dbf-6088-45b1-8bca-243b72ce996d no longer exists
+[AfterEach] [k8s.io] Variable Expansion
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:57:12.418: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "daemonsets-2528" for this suite.
-
-• [SLOW TEST:52.631 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should update pod when spec was updated and update strategy is RollingUpdate [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Oct 26 17:08:54.759: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "var-expansion-1041" for this suite.
+•{"msg":"PASSED [k8s.io] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance]","total":280,"completed":254,"skipped":4177,"failed":0}
+SS
 ------------------------------
-{"msg":"PASSED [sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]","total":280,"completed":247,"skipped":4005,"failed":0}
-SSSSSSSSSS
+[sig-storage] Downward API volume 
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:08:54.812: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename downward-api
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Oct 26 17:08:55.031: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9bdfddec-1389-4d43-b855-7c1a1c916791" in namespace "downward-api-9952" to be "success or failure"
+Oct 26 17:08:55.048: INFO: Pod "downwardapi-volume-9bdfddec-1389-4d43-b855-7c1a1c916791": Phase="Pending", Reason="", readiness=false. Elapsed: 16.765664ms
+Oct 26 17:08:57.065: INFO: Pod "downwardapi-volume-9bdfddec-1389-4d43-b855-7c1a1c916791": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.033709176s
+STEP: Saw pod success
+Oct 26 17:08:57.065: INFO: Pod "downwardapi-volume-9bdfddec-1389-4d43-b855-7c1a1c916791" satisfied condition "success or failure"
+Oct 26 17:08:57.087: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-9bdfddec-1389-4d43-b855-7c1a1c916791 container client-container: 
+STEP: delete the pod
+Oct 26 17:08:57.176: INFO: Waiting for pod downwardapi-volume-9bdfddec-1389-4d43-b855-7c1a1c916791 to disappear
+Oct 26 17:08:57.193: INFO: Pod downwardapi-volume-9bdfddec-1389-4d43-b855-7c1a1c916791 no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:08:57.193: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-9952" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance]","total":280,"completed":255,"skipped":4179,"failed":0}
+SSSSSSSSSSSSSS
 ------------------------------
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  listing mutating webhooks should work [Conformance]
+  listing validating webhooks should work [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:57:12.495: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:08:57.232: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
@@ -12373,1236 +11918,1140 @@ STEP: Setting up server cert
 STEP: Create role binding to let webhook read extension-apiserver-authentication
 STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Sep 21 16:57:14.417: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 17:08:58.329: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Sep 21 16:57:17.544: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] listing mutating webhooks should work [Conformance]
+Oct 26 17:09:01.415: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] listing validating webhooks should work [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Listing all of the created validation webhooks
-STEP: Creating a configMap that should be mutated
+STEP: Creating a configMap that does not comply to the validation webhook rules
 STEP: Deleting the collection of validation webhooks
-STEP: Creating a configMap that should not be mutated
+STEP: Creating a configMap that does not comply to the validation webhook rules
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:57:18.771: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-1457" for this suite.
-STEP: Destroying namespace "webhook-1457-markers" for this suite.
+Oct 26 17:09:02.193: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-3289" for this suite.
+STEP: Destroying namespace "webhook-3289-markers" for this suite.
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:6.741 seconds]
+• [SLOW TEST:5.254 seconds]
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  listing mutating webhooks should work [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance]","total":280,"completed":248,"skipped":4015,"failed":0}
-[sig-api-machinery] Garbage collector 
-  should not be blocked by dependency circle [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:57:19.235: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename gc
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not be blocked by dependency circle [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:57:19.812: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"4b09ceda-f6ad-460f-a369-eee77c75ca31", Controller:(*bool)(0xc007e517f6), BlockOwnerDeletion:(*bool)(0xc007e517f7)}}
-Sep 21 16:57:19.874: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"a7e40a9d-f7cd-47a0-b1d0-c5be0376ce36", Controller:(*bool)(0xc00a34c002), BlockOwnerDeletion:(*bool)(0xc00a34c003)}}
-Sep 21 16:57:19.906: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"64145d71-4e9f-4a92-a71e-cb71fef11c9d", Controller:(*bool)(0xc005c9ca86), BlockOwnerDeletion:(*bool)(0xc005c9ca87)}}
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:57:24.985: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-429" for this suite.
-
-• [SLOW TEST:5.826 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should not be blocked by dependency circle [Conformance]
+  listing validating webhooks should work [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]","total":280,"completed":249,"skipped":4015,"failed":0}
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance]","total":280,"completed":256,"skipped":4193,"failed":0}
 SSSSSSSSSS
 ------------------------------
-[sig-apps] ReplicationController 
-  should adopt matching pods on creation [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:57:25.061: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename replication-controller
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should adopt matching pods on creation [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Given a Pod with a 'name' label pod-adoption is created
-STEP: When a replication controller with a matching selector is created
-STEP: Then the orphan pod is adopted
-[AfterEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:57:28.540: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replication-controller-9316" for this suite.
-•{"msg":"PASSED [sig-apps] ReplicationController should adopt matching pods on creation [Conformance]","total":280,"completed":250,"skipped":4025,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-api-machinery] Garbage collector 
-  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Garbage collector
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:57:28.608: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename gc
+Oct 26 17:09:02.487: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename webhook
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Oct 26 17:09:03.514: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Oct 26 17:09:05.561: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328943, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328943, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328943, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739328943, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Oct 26 17:09:08.609: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the deployment
-STEP: Wait for the Deployment to create new ReplicaSet
-STEP: delete the deployment
-STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the rs
-STEP: Gathering metrics
-Sep 21 16:57:59.171: INFO: For apiserver_request_total:
-For apiserver_request_latency_seconds:
-For apiserver_init_events_total:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
-
-W0921 16:57:59.171209      24 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-[AfterEach] [sig-api-machinery] Garbage collector
+STEP: Registering a validating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API
+STEP: Registering a mutating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API
+STEP: Creating a dummy validating-webhook-configuration object
+STEP: Deleting the validating-webhook-configuration, which should be possible to remove
+STEP: Creating a dummy mutating-webhook-configuration object
+STEP: Deleting the mutating-webhook-configuration, which should be possible to remove
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:57:59.171: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-6370" for this suite.
+Oct 26 17:09:08.948: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-7338" for this suite.
+STEP: Destroying namespace "webhook-7338-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:30.636 seconds]
-[sig-api-machinery] Garbage collector
+• [SLOW TEST:6.717 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
+  should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]","total":280,"completed":251,"skipped":4069,"failed":0}
-SSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]","total":280,"completed":257,"skipped":4203,"failed":0}
+SSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+[sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial] 
+  evicts pods with minTolerationSeconds [Disruptive] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
+[BeforeEach] [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:57:59.245: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 17:09:09.204: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename taint-multiple-pods
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+[BeforeEach] [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/taints.go:344
+Oct 26 17:09:09.392: INFO: Waiting up to 1m0s for all nodes to be ready
+Oct 26 17:10:09.531: INFO: Waiting for terminating namespaces to be deleted...
+[It] evicts pods with minTolerationSeconds [Disruptive] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating projection with secret that has name projected-secret-test-map-f077160d-0002-4415-bf56-26b58622cffb
-STEP: Creating a pod to test consume secrets
-Sep 21 16:57:59.662: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-7cb46d4f-454e-4e27-a035-b9af1311d275" in namespace "projected-4114" to be "success or failure"
-Sep 21 16:57:59.687: INFO: Pod "pod-projected-secrets-7cb46d4f-454e-4e27-a035-b9af1311d275": Phase="Pending", Reason="", readiness=false. Elapsed: 25.237632ms
-Sep 21 16:58:01.711: INFO: Pod "pod-projected-secrets-7cb46d4f-454e-4e27-a035-b9af1311d275": Phase="Pending", Reason="", readiness=false. Elapsed: 2.049356249s
-Sep 21 16:58:03.740: INFO: Pod "pod-projected-secrets-7cb46d4f-454e-4e27-a035-b9af1311d275": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.078201251s
-STEP: Saw pod success
-Sep 21 16:58:03.740: INFO: Pod "pod-projected-secrets-7cb46d4f-454e-4e27-a035-b9af1311d275" satisfied condition "success or failure"
-Sep 21 16:58:03.770: INFO: Trying to get logs from node 10.241.51.147 pod pod-projected-secrets-7cb46d4f-454e-4e27-a035-b9af1311d275 container projected-secret-volume-test: 
-STEP: delete the pod
-Sep 21 16:58:03.971: INFO: Waiting for pod pod-projected-secrets-7cb46d4f-454e-4e27-a035-b9af1311d275 to disappear
-Sep 21 16:58:03.991: INFO: Pod pod-projected-secrets-7cb46d4f-454e-4e27-a035-b9af1311d275 no longer exists
-[AfterEach] [sig-storage] Projected secret
+Oct 26 17:10:09.551: INFO: Starting informer...
+STEP: Starting pods...
+Oct 26 17:10:09.854: INFO: Pod1 is running on 10.72.119.74. Tainting Node
+Oct 26 17:10:14.162: INFO: Pod2 is running on 10.72.119.74. Tainting Node
+STEP: Trying to apply a taint on the Node
+STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute
+STEP: Waiting for Pod1 and Pod2 to be deleted
+Oct 26 17:10:26.764: INFO: Noticed Pod "taint-eviction-b1" gets evicted.
+Oct 26 17:10:46.672: INFO: Noticed Pod "taint-eviction-b2" gets evicted.
+STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute
+[AfterEach] [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:58:03.991: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-4114" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":252,"skipped":4079,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSS
+Oct 26 17:10:46.724: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "taint-multiple-pods-7157" for this suite.
+
+• [SLOW TEST:97.563 seconds]
+[sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
+  evicts pods with minTolerationSeconds [Disruptive] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] ConfigMap 
-  updates should be reflected in volume [NodeConformance] [Conformance]
+{"msg":"PASSED [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial] evicts pods with minTolerationSeconds [Disruptive] [Conformance]","total":280,"completed":258,"skipped":4214,"failed":0}
+SSSSSSSSSSS
+------------------------------
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
+[BeforeEach] [sig-apps] StatefulSet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:58:04.075: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename configmap
+Oct 26 17:10:46.768: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename statefulset
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] updates should be reflected in volume [NodeConformance] [Conformance]
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
+STEP: Creating service test in namespace statefulset-3075
+[It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:58:04.355: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
-STEP: Creating configMap with name configmap-test-upd-60c445c8-20b9-4dd2-a0a0-2b21f561f4fd
-STEP: Creating the pod
-STEP: Updating configmap configmap-test-upd-60c445c8-20b9-4dd2-a0a0-2b21f561f4fd
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] ConfigMap
+STEP: Initializing watcher for selector baz=blah,foo=bar
+STEP: Creating stateful set ss in namespace statefulset-3075
+STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-3075
+Oct 26 17:10:47.038: INFO: Found 0 stateful pods, waiting for 1
+Oct 26 17:10:57.059: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod
+Oct 26 17:10:57.076: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-3075 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Oct 26 17:10:57.627: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Oct 26 17:10:57.627: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Oct 26 17:10:57.627: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Oct 26 17:10:57.646: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
+Oct 26 17:11:07.663: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Oct 26 17:11:07.664: INFO: Waiting for statefulset status.replicas updated to 0
+Oct 26 17:11:07.742: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999997941s
+Oct 26 17:11:08.758: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.983275973s
+Oct 26 17:11:09.775: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.966790369s
+Oct 26 17:11:10.791: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.950327136s
+Oct 26 17:11:11.814: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.933476415s
+Oct 26 17:11:12.833: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.910533956s
+Oct 26 17:11:13.862: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.892236235s
+Oct 26 17:11:14.879: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.863306632s
+Oct 26 17:11:15.897: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.845858096s
+Oct 26 17:11:16.914: INFO: Verifying statefulset ss doesn't scale past 1 for another 828.252147ms
+STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-3075
+Oct 26 17:11:17.932: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-3075 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Oct 26 17:11:18.328: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
+Oct 26 17:11:18.328: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
+Oct 26 17:11:18.328: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
+
+Oct 26 17:11:18.346: INFO: Found 1 stateful pods, waiting for 3
+Oct 26 17:11:28.366: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+Oct 26 17:11:28.366: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
+Oct 26 17:11:28.366: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Verifying that stateful set ss was scaled up in order
+STEP: Scale down will halt with unhealthy stateful pod
+Oct 26 17:11:28.391: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-3075 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Oct 26 17:11:28.922: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Oct 26 17:11:28.922: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Oct 26 17:11:28.922: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Oct 26 17:11:28.922: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-3075 ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Oct 26 17:11:29.357: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Oct 26 17:11:29.357: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Oct 26 17:11:29.357: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Oct 26 17:11:29.357: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-3075 ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Oct 26 17:11:29.785: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Oct 26 17:11:29.785: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Oct 26 17:11:29.785: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Oct 26 17:11:29.785: INFO: Waiting for statefulset status.replicas updated to 0
+Oct 26 17:11:29.809: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2
+Oct 26 17:11:39.844: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Oct 26 17:11:39.844: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
+Oct 26 17:11:39.844: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
+Oct 26 17:11:39.900: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999997766s
+Oct 26 17:11:40.918: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.982582553s
+Oct 26 17:11:41.935: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.964307904s
+Oct 26 17:11:42.953: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.94809768s
+Oct 26 17:11:43.973: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.930280865s
+Oct 26 17:11:44.993: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.909634228s
+Oct 26 17:11:46.012: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.889660064s
+Oct 26 17:11:47.029: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.870713442s
+Oct 26 17:11:48.047: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.853426208s
+Oct 26 17:11:49.067: INFO: Verifying statefulset ss doesn't scale past 3 for another 835.62958ms
+STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-3075
+Oct 26 17:11:50.085: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-3075 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Oct 26 17:11:50.660: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
+Oct 26 17:11:50.661: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
+Oct 26 17:11:50.661: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
+
+Oct 26 17:11:50.661: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-3075 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Oct 26 17:11:51.060: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
+Oct 26 17:11:51.060: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
+Oct 26 17:11:51.060: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
+
+Oct 26 17:11:51.060: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=statefulset-3075 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Oct 26 17:11:51.443: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
+Oct 26 17:11:51.443: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
+Oct 26 17:11:51.443: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
+
+Oct 26 17:11:51.443: INFO: Scaling statefulset ss to 0
+STEP: Verifying that stateful set ss was scaled down in reverse order
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
+Oct 26 17:12:01.519: INFO: Deleting all statefulset in ns statefulset-3075
+Oct 26 17:12:01.542: INFO: Scaling statefulset ss to 0
+Oct 26 17:12:01.601: INFO: Waiting for statefulset status.replicas updated to 0
+Oct 26 17:12:01.619: INFO: Deleting statefulset ss
+[AfterEach] [sig-apps] StatefulSet
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:58:09.047: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-4715" for this suite.
+Oct 26 17:12:01.717: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-3075" for this suite.
 
-• [SLOW TEST:5.075 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
-  updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:74.998 seconds]
+[sig-apps] StatefulSet
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+    Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":253,"skipped":4103,"failed":0}
-SSSSSSSSSSS
+{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]","total":280,"completed":259,"skipped":4225,"failed":0}
+SSSSSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+[k8s.io] Docker Containers 
+  should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
+[BeforeEach] [k8s.io] Docker Containers
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:58:09.151: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 17:12:01.766: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename containers
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+[It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:58:09.552: INFO: Waiting up to 5m0s for pod "downwardapi-volume-89baf4ae-5f72-467c-bc12-cf7d87671b67" in namespace "projected-7216" to be "success or failure"
-Sep 21 16:58:09.584: INFO: Pod "downwardapi-volume-89baf4ae-5f72-467c-bc12-cf7d87671b67": Phase="Pending", Reason="", readiness=false. Elapsed: 32.846194ms
-Sep 21 16:58:11.610: INFO: Pod "downwardapi-volume-89baf4ae-5f72-467c-bc12-cf7d87671b67": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.058302104s
+STEP: Creating a pod to test override arguments
+Oct 26 17:12:01.989: INFO: Waiting up to 5m0s for pod "client-containers-a5c04566-28af-4613-a50d-ccdfcdcdc690" in namespace "containers-2871" to be "success or failure"
+Oct 26 17:12:02.012: INFO: Pod "client-containers-a5c04566-28af-4613-a50d-ccdfcdcdc690": Phase="Pending", Reason="", readiness=false. Elapsed: 22.085243ms
+Oct 26 17:12:04.027: INFO: Pod "client-containers-a5c04566-28af-4613-a50d-ccdfcdcdc690": Phase="Pending", Reason="", readiness=false. Elapsed: 2.037720889s
+Oct 26 17:12:06.051: INFO: Pod "client-containers-a5c04566-28af-4613-a50d-ccdfcdcdc690": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.061737228s
 STEP: Saw pod success
-Sep 21 16:58:11.610: INFO: Pod "downwardapi-volume-89baf4ae-5f72-467c-bc12-cf7d87671b67" satisfied condition "success or failure"
-Sep 21 16:58:11.638: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-89baf4ae-5f72-467c-bc12-cf7d87671b67 container client-container: 
+Oct 26 17:12:06.051: INFO: Pod "client-containers-a5c04566-28af-4613-a50d-ccdfcdcdc690" satisfied condition "success or failure"
+Oct 26 17:12:06.068: INFO: Trying to get logs from node 10.72.119.74 pod client-containers-a5c04566-28af-4613-a50d-ccdfcdcdc690 container test-container: 
 STEP: delete the pod
-Sep 21 16:58:11.799: INFO: Waiting for pod downwardapi-volume-89baf4ae-5f72-467c-bc12-cf7d87671b67 to disappear
-Sep 21 16:58:11.827: INFO: Pod downwardapi-volume-89baf4ae-5f72-467c-bc12-cf7d87671b67 no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
+Oct 26 17:12:06.198: INFO: Waiting for pod client-containers-a5c04566-28af-4613-a50d-ccdfcdcdc690 to disappear
+Oct 26 17:12:06.230: INFO: Pod client-containers-a5c04566-28af-4613-a50d-ccdfcdcdc690 no longer exists
+[AfterEach] [k8s.io] Docker Containers
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:58:11.827: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-7216" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":254,"skipped":4114,"failed":0}
+Oct 26 17:12:06.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "containers-2871" for this suite.
+•{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]","total":280,"completed":260,"skipped":4231,"failed":0}
 SSSSSSSS
 ------------------------------
+[sig-network] Networking Granular Checks: Pods 
+  should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Networking
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Oct 26 17:12:06.270: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename pod-network-test
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Performing setup for networking test in namespace pod-network-test-5640
+STEP: creating a selector
+STEP: Creating the service pods in kubernetes
+Oct 26 17:12:06.411: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+STEP: Creating test pods
+Oct 26 17:12:32.933: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.194.117:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-5640 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 17:12:32.933: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 17:12:33.171: INFO: Found all expected endpoints: [netserver-0]
+Oct 26 17:12:33.196: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.103.243:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-5640 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 17:12:33.196: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 17:12:33.422: INFO: Found all expected endpoints: [netserver-1]
+Oct 26 17:12:33.437: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.54.171:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-5640 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 17:12:33.437: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 17:12:33.684: INFO: Found all expected endpoints: [netserver-2]
+[AfterEach] [sig-network] Networking
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:12:33.684: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pod-network-test-5640" for this suite.
+
+• [SLOW TEST:27.454 seconds]
+[sig-network] Networking
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
+  Granular Checks: Pods
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
+    should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":261,"skipped":4239,"failed":0}
 [sig-network] DNS 
-  should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]
+  should provide DNS for ExternalName services [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-network] DNS
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:58:11.910: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:12:33.725: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename dns
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]
+[It] should provide DNS for ExternalName services [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a test headless service
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-2792 A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-2792;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-2792 A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-2792;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-2792.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-2792.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-2792.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-2792.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-2792.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-2792.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-2792.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-2792.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-2792.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-2792.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-2792.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-2792.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-2792.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 91.245.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.245.91_udp@PTR;check="$$(dig +tcp +noall +answer +search 91.245.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.245.91_tcp@PTR;sleep 1; done
+STEP: Creating a test externalName service
+STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-745.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-745.svc.cluster.local; sleep 1; done
 
-STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-2792 A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-2792;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-2792 A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-2792;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-2792.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-2792.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-2792.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-2792.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-2792.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-2792.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-2792.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-2792.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-2792.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-2792.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-2792.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-2792.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-2792.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 91.245.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.245.91_udp@PTR;check="$$(dig +tcp +noall +answer +search 91.245.21.172.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/172.21.245.91_tcp@PTR;sleep 1; done
+STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-745.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-745.svc.cluster.local; sleep 1; done
 
 STEP: creating a pod to probe DNS
 STEP: submitting the pod to kubernetes
 STEP: retrieving the pod
 STEP: looking for the results for each expected name from probers
-Sep 21 16:58:16.753: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:16.803: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:16.850: INFO: Unable to read wheezy_udp@dns-test-service.dns-2792 from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:16.889: INFO: Unable to read wheezy_tcp@dns-test-service.dns-2792 from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:16.934: INFO: Unable to read wheezy_udp@dns-test-service.dns-2792.svc from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:16.974: INFO: Unable to read wheezy_tcp@dns-test-service.dns-2792.svc from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:17.015: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-2792.svc from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:17.051: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-2792.svc from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:17.362: INFO: Unable to read jessie_udp@dns-test-service from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:17.624: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:17.679: INFO: Unable to read jessie_udp@dns-test-service.dns-2792 from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:17.729: INFO: Unable to read jessie_tcp@dns-test-service.dns-2792 from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:17.776: INFO: Unable to read jessie_udp@dns-test-service.dns-2792.svc from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:17.822: INFO: Unable to read jessie_tcp@dns-test-service.dns-2792.svc from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:17.869: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-2792.svc from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:17.914: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-2792.svc from pod dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd: the server could not find the requested resource (get pods dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd)
-Sep 21 16:58:18.205: INFO: Lookups using dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-2792 wheezy_tcp@dns-test-service.dns-2792 wheezy_udp@dns-test-service.dns-2792.svc wheezy_tcp@dns-test-service.dns-2792.svc wheezy_udp@_http._tcp.dns-test-service.dns-2792.svc wheezy_tcp@_http._tcp.dns-test-service.dns-2792.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-2792 jessie_tcp@dns-test-service.dns-2792 jessie_udp@dns-test-service.dns-2792.svc jessie_tcp@dns-test-service.dns-2792.svc jessie_udp@_http._tcp.dns-test-service.dns-2792.svc jessie_tcp@_http._tcp.dns-test-service.dns-2792.svc]
-
-Sep 21 16:58:25.173: INFO: DNS probes using dns-2792/dns-test-4b566a8d-8fae-43bf-8164-f525d847cebd succeeded
+Oct 26 17:12:38.077: INFO: DNS probes using dns-test-9c17de91-871f-43f5-b311-f3a0dcaf267a succeeded
 
 STEP: deleting the pod
-STEP: deleting the test service
-STEP: deleting the test headless service
-[AfterEach] [sig-network] DNS
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:58:26.030: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-2792" for this suite.
+STEP: changing the externalName to bar.example.com
+STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-745.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-745.svc.cluster.local; sleep 1; done
 
-• [SLOW TEST:14.364 seconds]
-[sig-network] DNS
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]","total":280,"completed":255,"skipped":4122,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-apps] Daemon set [Serial] 
-  should run and stop complex daemon [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:58:26.276: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename daemonsets
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
-[It] should run and stop complex daemon [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 16:58:26.788: INFO: Creating daemon "daemon-set" with a node selector
-STEP: Initially, daemon pods should not be running on any nodes.
-Sep 21 16:58:26.836: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:26.836: INFO: Number of running nodes: 0, number of available pods: 0
-STEP: Change node label to blue, check that daemon pod is launched.
-Sep 21 16:58:27.025: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:27.025: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:58:28.056: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:28.056: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:58:29.054: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:29.054: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:58:30.048: INFO: Number of nodes with available pods: 1
-Sep 21 16:58:30.048: INFO: Number of running nodes: 1, number of available pods: 1
-STEP: Update the node label to green, and wait for daemons to be unscheduled
-Sep 21 16:58:30.210: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:30.210: INFO: Number of running nodes: 0, number of available pods: 0
-STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate
-Sep 21 16:58:30.284: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:30.284: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:58:31.307: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:31.307: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:58:32.308: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:32.308: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:58:33.315: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:33.315: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:58:34.310: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:34.310: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:58:35.308: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:35.308: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:58:36.307: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:36.307: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:58:37.513: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:37.513: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:58:38.307: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:38.307: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:58:39.310: INFO: Number of nodes with available pods: 1
-Sep 21 16:58:39.310: INFO: Number of running nodes: 1, number of available pods: 1
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-4098, will wait for the garbage collector to delete the pods
-Sep 21 16:58:39.785: INFO: Deleting DaemonSet.extensions daemon-set took: 280.618517ms
-Sep 21 16:58:40.485: INFO: Terminating DaemonSet.extensions daemon-set pods took: 700.220444ms
-Sep 21 16:58:47.318: INFO: Number of nodes with available pods: 0
-Sep 21 16:58:47.318: INFO: Number of running nodes: 0, number of available pods: 0
-Sep 21 16:58:47.339: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-4098/daemonsets","resourceVersion":"92040"},"items":null}
+STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-745.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-745.svc.cluster.local; sleep 1; done
 
-Sep 21 16:58:47.367: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-4098/pods","resourceVersion":"92040"},"items":null}
+STEP: creating a second pod to probe DNS
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Oct 26 17:12:40.305: INFO: File wheezy_udp@dns-test-service-3.dns-745.svc.cluster.local from pod  dns-745/dns-test-758337fe-e278-4c99-9c7a-2bc96b1742a7 contains 'foo.example.com.
+' instead of 'bar.example.com.'
+Oct 26 17:12:40.342: INFO: File jessie_udp@dns-test-service-3.dns-745.svc.cluster.local from pod  dns-745/dns-test-758337fe-e278-4c99-9c7a-2bc96b1742a7 contains 'foo.example.com.
+' instead of 'bar.example.com.'
+Oct 26 17:12:40.342: INFO: Lookups using dns-745/dns-test-758337fe-e278-4c99-9c7a-2bc96b1742a7 failed for: [wheezy_udp@dns-test-service-3.dns-745.svc.cluster.local jessie_udp@dns-test-service-3.dns-745.svc.cluster.local]
 
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:58:47.578: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "daemonsets-4098" for this suite.
+Oct 26 17:12:45.393: INFO: DNS probes using dns-test-758337fe-e278-4c99-9c7a-2bc96b1742a7 succeeded
 
-• [SLOW TEST:21.377 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should run and stop complex daemon [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance]","total":280,"completed":256,"skipped":4158,"failed":0}
-SSSSS
-------------------------------
-[sig-network] Services 
-  should be able to change the type from NodePort to ExternalName [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:58:47.653: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename services
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should be able to change the type from NodePort to ExternalName [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a service nodeport-service with the type=NodePort in namespace services-632
-STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service
-STEP: creating service externalsvc in namespace services-632
-STEP: creating replication controller externalsvc in namespace services-632
-I0921 16:58:48.167297      24 runners.go:189] Created replication controller with name: externalsvc, namespace: services-632, replica count: 2
-I0921 16:58:51.219224      24 runners.go:189] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-STEP: changing the NodePort service to type=ExternalName
-Sep 21 16:58:51.390: INFO: Creating new exec pod
-Sep 21 16:58:53.624: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-632 execpodnkf7p -- /bin/sh -x -c nslookup nodeport-service'
-Sep 21 16:58:54.213: INFO: stderr: "+ nslookup nodeport-service\n"
-Sep 21 16:58:54.213: INFO: stdout: "Server:\t\t172.21.0.10\nAddress:\t172.21.0.10#53\n\nnodeport-service.services-632.svc.cluster.local\tcanonical name = externalsvc.services-632.svc.cluster.local.\nName:\texternalsvc.services-632.svc.cluster.local\nAddress: 172.21.246.32\n\n"
-STEP: deleting ReplicationController externalsvc in namespace services-632, will wait for the garbage collector to delete the pods
-Sep 21 16:58:54.354: INFO: Deleting ReplicationController externalsvc took: 68.599678ms
-Sep 21 16:58:54.854: INFO: Terminating ReplicationController externalsvc pods took: 500.286602ms
-Sep 21 16:59:09.809: INFO: Cleaning up the NodePort to ExternalName test service
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:59:10.413: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-632" for this suite.
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
+STEP: deleting the pod
+STEP: changing the service to type=ClusterIP
+STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-745.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-745.svc.cluster.local; sleep 1; done
 
-• [SLOW TEST:22.853 seconds]
-[sig-network] Services
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should be able to change the type from NodePort to ExternalName [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance]","total":280,"completed":257,"skipped":4163,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-apps] Daemon set [Serial] 
-  should run and stop simple daemon [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 16:59:10.507: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename daemonsets
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
-[It] should run and stop simple daemon [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating simple DaemonSet "daemon-set"
-STEP: Check that daemon pods launch on every node of the cluster.
-Sep 21 16:59:11.364: INFO: Number of nodes with available pods: 0
-Sep 21 16:59:11.364: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:59:12.427: INFO: Number of nodes with available pods: 0
-Sep 21 16:59:12.427: INFO: Node 10.241.51.147 is running more than one daemon pod
-Sep 21 16:59:13.458: INFO: Number of nodes with available pods: 2
-Sep 21 16:59:13.458: INFO: Node 10.241.51.149 is running more than one daemon pod
-Sep 21 16:59:14.430: INFO: Number of nodes with available pods: 3
-Sep 21 16:59:14.430: INFO: Number of running nodes: 3, number of available pods: 3
-STEP: Stop a daemon pod, check that the daemon pod is revived.
-Sep 21 16:59:14.594: INFO: Number of nodes with available pods: 2
-Sep 21 16:59:14.594: INFO: Node 10.241.51.150 is running more than one daemon pod
-Sep 21 16:59:15.655: INFO: Number of nodes with available pods: 2
-Sep 21 16:59:15.655: INFO: Node 10.241.51.150 is running more than one daemon pod
-Sep 21 16:59:16.672: INFO: Number of nodes with available pods: 2
-Sep 21 16:59:16.672: INFO: Node 10.241.51.150 is running more than one daemon pod
-Sep 21 16:59:17.677: INFO: Number of nodes with available pods: 2
-Sep 21 16:59:17.677: INFO: Node 10.241.51.150 is running more than one daemon pod
-Sep 21 16:59:18.676: INFO: Number of nodes with available pods: 2
-Sep 21 16:59:18.676: INFO: Node 10.241.51.150 is running more than one daemon pod
-Sep 21 16:59:19.647: INFO: Number of nodes with available pods: 2
-Sep 21 16:59:19.647: INFO: Node 10.241.51.150 is running more than one daemon pod
-Sep 21 16:59:20.649: INFO: Number of nodes with available pods: 2
-Sep 21 16:59:20.650: INFO: Node 10.241.51.150 is running more than one daemon pod
-Sep 21 16:59:21.645: INFO: Number of nodes with available pods: 2
-Sep 21 16:59:21.645: INFO: Node 10.241.51.150 is running more than one daemon pod
-Sep 21 16:59:22.639: INFO: Number of nodes with available pods: 3
-Sep 21 16:59:22.639: INFO: Number of running nodes: 3, number of available pods: 3
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-5923, will wait for the garbage collector to delete the pods
-Sep 21 16:59:22.771: INFO: Deleting DaemonSet.extensions daemon-set took: 44.566645ms
-Sep 21 16:59:23.271: INFO: Terminating DaemonSet.extensions daemon-set pods took: 500.248264ms
-Sep 21 16:59:32.296: INFO: Number of nodes with available pods: 0
-Sep 21 16:59:32.297: INFO: Number of running nodes: 0, number of available pods: 0
-Sep 21 16:59:32.317: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-5923/daemonsets","resourceVersion":"92545"},"items":null}
+STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-745.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-745.svc.cluster.local; sleep 1; done
 
-Sep 21 16:59:32.342: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-5923/pods","resourceVersion":"92545"},"items":null}
+STEP: creating a third pod to probe DNS
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Oct 26 17:12:49.662: INFO: DNS probes using dns-test-9ac5c5b8-5259-4cff-9911-a8da55661bb8 succeeded
 
-[AfterEach] [sig-apps] Daemon set [Serial]
+STEP: deleting the pod
+STEP: deleting the test externalName service
+[AfterEach] [sig-network] DNS
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:59:32.443: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "daemonsets-5923" for this suite.
+Oct 26 17:12:49.781: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-745" for this suite.
 
-• [SLOW TEST:21.998 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should run and stop simple daemon [Conformance]
+• [SLOW TEST:16.100 seconds]
+[sig-network] DNS
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should provide DNS for ExternalName services [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance]","total":280,"completed":258,"skipped":4198,"failed":0}
-SSSSSSSS
+{"msg":"PASSED [sig-network] DNS should provide DNS for ExternalName services [Conformance]","total":280,"completed":262,"skipped":4239,"failed":0}
+SSSSSSSSSSS
 ------------------------------
-[sig-node] ConfigMap 
-  should fail to create ConfigMap with empty key [Conformance]
+[sig-storage] Projected configMap 
+  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] ConfigMap
+[BeforeEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:59:32.505: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename configmap
+Oct 26 17:12:49.826: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should fail to create ConfigMap with empty key [Conformance]
+[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap that has name configmap-test-emptyKey-c7498579-d9f9-4533-8e28-5df32da5844a
-[AfterEach] [sig-node] ConfigMap
+STEP: Creating configMap with name projected-configmap-test-volume-c708f6d4-3840-43f8-b80b-d5af514760ff
+STEP: Creating a pod to test consume configMaps
+Oct 26 17:12:50.067: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-82f9d521-257f-497f-a703-a16978880c25" in namespace "projected-4651" to be "success or failure"
+Oct 26 17:12:50.085: INFO: Pod "pod-projected-configmaps-82f9d521-257f-497f-a703-a16978880c25": Phase="Pending", Reason="", readiness=false. Elapsed: 18.165608ms
+Oct 26 17:12:52.104: INFO: Pod "pod-projected-configmaps-82f9d521-257f-497f-a703-a16978880c25": Phase="Pending", Reason="", readiness=false. Elapsed: 2.036699741s
+Oct 26 17:12:54.120: INFO: Pod "pod-projected-configmaps-82f9d521-257f-497f-a703-a16978880c25": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.053065254s
+STEP: Saw pod success
+Oct 26 17:12:54.120: INFO: Pod "pod-projected-configmaps-82f9d521-257f-497f-a703-a16978880c25" satisfied condition "success or failure"
+Oct 26 17:12:54.136: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-configmaps-82f9d521-257f-497f-a703-a16978880c25 container projected-configmap-volume-test: 
+STEP: delete the pod
+Oct 26 17:12:54.222: INFO: Waiting for pod pod-projected-configmaps-82f9d521-257f-497f-a703-a16978880c25 to disappear
+Oct 26 17:12:54.237: INFO: Pod pod-projected-configmaps-82f9d521-257f-497f-a703-a16978880c25 no longer exists
+[AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:59:32.727: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-2309" for this suite.
-•{"msg":"PASSED [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance]","total":280,"completed":259,"skipped":4206,"failed":0}
-SS
+Oct 26 17:12:54.237: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-4651" for this suite.
+•{"msg":"PASSED [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":280,"completed":263,"skipped":4250,"failed":0}
+SSSSS
 ------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume [NodeConformance] [Conformance]
+[sig-storage] Projected secret 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
+[BeforeEach] [sig-storage] Projected secret
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:59:32.821: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename secrets
+Oct 26 17:12:54.290: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-5698a62c-aa5d-4da1-920e-bdca3790bc93
+STEP: Creating projection with secret that has name projected-secret-test-map-11ca83c6-0fe9-4f17-8044-ccc2d9a8ac48
 STEP: Creating a pod to test consume secrets
-Sep 21 16:59:33.227: INFO: Waiting up to 5m0s for pod "pod-secrets-c69c0404-992a-476e-abd3-976e05333f0b" in namespace "secrets-843" to be "success or failure"
-Sep 21 16:59:33.257: INFO: Pod "pod-secrets-c69c0404-992a-476e-abd3-976e05333f0b": Phase="Pending", Reason="", readiness=false. Elapsed: 29.305429ms
-Sep 21 16:59:35.286: INFO: Pod "pod-secrets-c69c0404-992a-476e-abd3-976e05333f0b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.058983449s
-Sep 21 16:59:37.312: INFO: Pod "pod-secrets-c69c0404-992a-476e-abd3-976e05333f0b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.084603305s
+Oct 26 17:12:54.528: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-9d861947-9a9c-437c-b3a6-6db055ef796a" in namespace "projected-4880" to be "success or failure"
+Oct 26 17:12:54.546: INFO: Pod "pod-projected-secrets-9d861947-9a9c-437c-b3a6-6db055ef796a": Phase="Pending", Reason="", readiness=false. Elapsed: 17.638695ms
+Oct 26 17:12:56.565: INFO: Pod "pod-projected-secrets-9d861947-9a9c-437c-b3a6-6db055ef796a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.037081874s
 STEP: Saw pod success
-Sep 21 16:59:37.312: INFO: Pod "pod-secrets-c69c0404-992a-476e-abd3-976e05333f0b" satisfied condition "success or failure"
-Sep 21 16:59:37.334: INFO: Trying to get logs from node 10.241.51.147 pod pod-secrets-c69c0404-992a-476e-abd3-976e05333f0b container secret-volume-test: 
+Oct 26 17:12:56.566: INFO: Pod "pod-projected-secrets-9d861947-9a9c-437c-b3a6-6db055ef796a" satisfied condition "success or failure"
+Oct 26 17:12:56.589: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-secrets-9d861947-9a9c-437c-b3a6-6db055ef796a container projected-secret-volume-test: 
 STEP: delete the pod
-Sep 21 16:59:37.479: INFO: Waiting for pod pod-secrets-c69c0404-992a-476e-abd3-976e05333f0b to disappear
-Sep 21 16:59:37.503: INFO: Pod pod-secrets-c69c0404-992a-476e-abd3-976e05333f0b no longer exists
-[AfterEach] [sig-storage] Secrets
+Oct 26 17:12:56.692: INFO: Waiting for pod pod-projected-secrets-9d861947-9a9c-437c-b3a6-6db055ef796a to disappear
+Oct 26 17:12:56.709: INFO: Pod pod-projected-secrets-9d861947-9a9c-437c-b3a6-6db055ef796a no longer exists
+[AfterEach] [sig-storage] Projected secret
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:59:37.503: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-843" for this suite.
-•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":260,"skipped":4208,"failed":0}
-SSSSSSSSSS
+Oct 26 17:12:56.709: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-4880" for this suite.
+•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":264,"skipped":4255,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
-  should have an terminated reason [NodeConformance] [Conformance]
+[sig-cli] Kubectl client Kubectl run --rm job 
+  should create a job from an image, then delete the job  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Kubelet
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:59:37.565: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename kubelet-test
+Oct 26 17:12:56.764: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubectl
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
-[BeforeEach] when scheduling a busybox command that always fails in a pod
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
-[It] should have an terminated reason [NodeConformance] [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
+[It] should create a job from an image, then delete the job  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [k8s.io] Kubelet
+STEP: executing a command with run --rm and attach with stdin
+Oct 26 17:12:56.925: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 --namespace=kubectl-5042 run e2e-test-rm-busybox-job --image=docker.io/library/busybox:1.29 --rm=true --generator=job/v1 --restart=OnFailure --attach=true --stdin -- sh -c cat && echo 'stdin closed''
+Oct 26 17:12:59.650: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\nIf you don't see a command prompt, try pressing enter.\n"
+Oct 26 17:12:59.650: INFO: stdout: "abcd1234stdin closed\njob.batch \"e2e-test-rm-busybox-job\" deleted\n"
+STEP: verifying the job e2e-test-rm-busybox-job was deleted
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:59:41.919: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubelet-test-9582" for this suite.
-•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance]","total":280,"completed":261,"skipped":4218,"failed":0}
-SSS
+Oct 26 17:13:01.689: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-5042" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run --rm job should create a job from an image, then delete the job  [Conformance]","total":280,"completed":265,"skipped":4312,"failed":0}
+SSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide container's memory limit [NodeConformance] [Conformance]
+[sig-storage] EmptyDir volumes 
+  should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:59:41.967: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename downward-api
+Oct 26 17:13:01.733: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should provide container's memory limit [NodeConformance] [Conformance]
+[It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Sep 21 16:59:42.296: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fc42e488-b525-4c24-9226-56b73fe4a634" in namespace "downward-api-6976" to be "success or failure"
-Sep 21 16:59:42.323: INFO: Pod "downwardapi-volume-fc42e488-b525-4c24-9226-56b73fe4a634": Phase="Pending", Reason="", readiness=false. Elapsed: 27.461047ms
-Sep 21 16:59:44.344: INFO: Pod "downwardapi-volume-fc42e488-b525-4c24-9226-56b73fe4a634": Phase="Pending", Reason="", readiness=false. Elapsed: 2.04843915s
-Sep 21 16:59:46.366: INFO: Pod "downwardapi-volume-fc42e488-b525-4c24-9226-56b73fe4a634": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.069957886s
+STEP: Creating a pod to test emptydir 0644 on tmpfs
+Oct 26 17:13:01.940: INFO: Waiting up to 5m0s for pod "pod-6cc01310-8ec9-4193-ba03-9211c677f7a5" in namespace "emptydir-8362" to be "success or failure"
+Oct 26 17:13:01.956: INFO: Pod "pod-6cc01310-8ec9-4193-ba03-9211c677f7a5": Phase="Pending", Reason="", readiness=false. Elapsed: 15.627912ms
+Oct 26 17:13:03.974: INFO: Pod "pod-6cc01310-8ec9-4193-ba03-9211c677f7a5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033876509s
+Oct 26 17:13:05.995: INFO: Pod "pod-6cc01310-8ec9-4193-ba03-9211c677f7a5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.055011763s
 STEP: Saw pod success
-Sep 21 16:59:46.366: INFO: Pod "downwardapi-volume-fc42e488-b525-4c24-9226-56b73fe4a634" satisfied condition "success or failure"
-Sep 21 16:59:46.387: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-fc42e488-b525-4c24-9226-56b73fe4a634 container client-container: 
+Oct 26 17:13:05.996: INFO: Pod "pod-6cc01310-8ec9-4193-ba03-9211c677f7a5" satisfied condition "success or failure"
+Oct 26 17:13:06.011: INFO: Trying to get logs from node 10.72.119.74 pod pod-6cc01310-8ec9-4193-ba03-9211c677f7a5 container test-container: 
 STEP: delete the pod
-Sep 21 16:59:46.510: INFO: Waiting for pod downwardapi-volume-fc42e488-b525-4c24-9226-56b73fe4a634 to disappear
-Sep 21 16:59:46.533: INFO: Pod downwardapi-volume-fc42e488-b525-4c24-9226-56b73fe4a634 no longer exists
-[AfterEach] [sig-storage] Downward API volume
+Oct 26 17:13:06.096: INFO: Waiting for pod pod-6cc01310-8ec9-4193-ba03-9211c677f7a5 to disappear
+Oct 26 17:13:06.111: INFO: Pod pod-6cc01310-8ec9-4193-ba03-9211c677f7a5 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:59:46.533: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-6976" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance]","total":280,"completed":262,"skipped":4221,"failed":0}
-SSSSSSSSSSSSSS
+Oct 26 17:13:06.111: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-8362" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":266,"skipped":4330,"failed":0}
+
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should unconditionally reject operations on fail closed webhook [Conformance]
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+[BeforeEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:59:46.589: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename webhook
+Oct 26 17:13:06.152: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Sep 21 16:59:47.848: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Sep 21 16:59:50.932: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should unconditionally reject operations on fail closed webhook [Conformance]
+[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering a webhook that server cannot talk to, with fail closed policy, via the AdmissionRegistration API
-STEP: create a namespace for the webhook
-STEP: create a configmap should be unconditionally rejected by the webhook
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+STEP: Creating configMap with name projected-configmap-test-volume-10e66cba-09a7-4038-af76-3dfa916244b5
+STEP: Creating a pod to test consume configMaps
+Oct 26 17:13:06.380: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-5097a052-18ca-4306-a8d3-a822821dbe95" in namespace "projected-6049" to be "success or failure"
+Oct 26 17:13:06.397: INFO: Pod "pod-projected-configmaps-5097a052-18ca-4306-a8d3-a822821dbe95": Phase="Pending", Reason="", readiness=false. Elapsed: 16.645909ms
+Oct 26 17:13:08.413: INFO: Pod "pod-projected-configmaps-5097a052-18ca-4306-a8d3-a822821dbe95": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.032955713s
+STEP: Saw pod success
+Oct 26 17:13:08.414: INFO: Pod "pod-projected-configmaps-5097a052-18ca-4306-a8d3-a822821dbe95" satisfied condition "success or failure"
+Oct 26 17:13:08.430: INFO: Trying to get logs from node 10.72.119.74 pod pod-projected-configmaps-5097a052-18ca-4306-a8d3-a822821dbe95 container projected-configmap-volume-test: 
+STEP: delete the pod
+Oct 26 17:13:08.518: INFO: Waiting for pod pod-projected-configmaps-5097a052-18ca-4306-a8d3-a822821dbe95 to disappear
+Oct 26 17:13:08.532: INFO: Pod pod-projected-configmaps-5097a052-18ca-4306-a8d3-a822821dbe95 no longer exists
+[AfterEach] [sig-storage] Projected configMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 16:59:51.235: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-9084" for this suite.
-STEP: Destroying namespace "webhook-9084-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
-
-• [SLOW TEST:5.074 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should unconditionally reject operations on fail closed webhook [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]","total":280,"completed":263,"skipped":4235,"failed":0}
-SSSSSSS
+Oct 26 17:13:08.532: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-6049" for this suite.
+•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":267,"skipped":4330,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-network] Networking Granular Checks: Pods 
-  should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+  should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-network] Networking
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 16:59:51.665: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:13:08.574: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename pod-network-test
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+[It] should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Performing setup for networking test in namespace pod-network-test-5419
+STEP: Performing setup for networking test in namespace pod-network-test-5563
 STEP: creating a selector
 STEP: Creating the service pods in kubernetes
-Sep 21 16:59:51.880: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+Oct 26 17:13:08.732: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
 STEP: Creating test pods
-Sep 21 17:00:14.706: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.237.100:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-5419 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 17:00:14.706: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 17:00:14.981: INFO: Found all expected endpoints: [netserver-0]
-Sep 21 17:00:15.008: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.139.45:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-5419 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 17:00:15.008: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 17:00:15.240: INFO: Found all expected endpoints: [netserver-1]
-Sep 21 17:00:15.271: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.30.81.59:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-5419 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Sep 21 17:00:15.271: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 17:00:15.561: INFO: Found all expected endpoints: [netserver-2]
+Oct 26 17:13:27.191: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.194.126:8080/dial?request=hostname&protocol=udp&host=172.30.194.122&port=8081&tries=1'] Namespace:pod-network-test-5563 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 17:13:27.192: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 17:13:27.435: INFO: Waiting for responses: map[]
+Oct 26 17:13:27.470: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.194.126:8080/dial?request=hostname&protocol=udp&host=172.30.103.248&port=8081&tries=1'] Namespace:pod-network-test-5563 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 17:13:27.470: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 17:13:27.721: INFO: Waiting for responses: map[]
+Oct 26 17:13:27.736: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.30.194.126:8080/dial?request=hostname&protocol=udp&host=172.30.54.181&port=8081&tries=1'] Namespace:pod-network-test-5563 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Oct 26 17:13:27.737: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+Oct 26 17:13:27.974: INFO: Waiting for responses: map[]
 [AfterEach] [sig-network] Networking
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:00:15.561: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pod-network-test-5419" for this suite.
+Oct 26 17:13:27.974: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pod-network-test-5563" for this suite.
 
-• [SLOW TEST:23.962 seconds]
+• [SLOW TEST:19.435 seconds]
 [sig-network] Networking
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
   Granular Checks: Pods
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
-    should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+    should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
     /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":264,"skipped":4242,"failed":0}
-SSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-auth] ServiceAccounts 
-  should mount an API token into pods  [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-auth] ServiceAccounts
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 17:00:15.627: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename svcaccounts
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should mount an API token into pods  [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: getting the auto-created API token
-STEP: reading a file in the container
-Sep 21 17:00:18.602: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-3900 pod-service-account-ca64ab13-5617-46c1-ac42-2bc1d7f96ceb -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token'
-STEP: reading a file in the container
-Sep 21 17:00:18.956: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-3900 pod-service-account-ca64ab13-5617-46c1-ac42-2bc1d7f96ceb -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt'
-STEP: reading a file in the container
-Sep 21 17:00:19.361: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-3900 pod-service-account-ca64ab13-5617-46c1-ac42-2bc1d7f96ceb -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace'
-[AfterEach] [sig-auth] ServiceAccounts
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:00:19.731: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "svcaccounts-3900" for this suite.
-•{"msg":"PASSED [sig-auth] ServiceAccounts should mount an API token into pods  [Conformance]","total":280,"completed":265,"skipped":4261,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":268,"skipped":4403,"failed":0}
+S
 ------------------------------
-[sig-api-machinery] Secrets 
-  should be consumable from pods in env vars [NodeConformance] [Conformance]
+[sig-node] Downward API 
+  should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Secrets
+[BeforeEach] [sig-node] Downward API
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 17:00:19.800: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename secrets
+Oct 26 17:13:28.010: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename downward-api
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in env vars [NodeConformance] [Conformance]
+[It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-3d409e33-2969-4583-86a1-56da9684bb1f
-STEP: Creating a pod to test consume secrets
-Sep 21 17:00:20.311: INFO: Waiting up to 5m0s for pod "pod-secrets-efef628d-2c27-46fc-8b8c-e17ade1d44d3" in namespace "secrets-6215" to be "success or failure"
-Sep 21 17:00:20.336: INFO: Pod "pod-secrets-efef628d-2c27-46fc-8b8c-e17ade1d44d3": Phase="Pending", Reason="", readiness=false. Elapsed: 24.936948ms
-Sep 21 17:00:22.372: INFO: Pod "pod-secrets-efef628d-2c27-46fc-8b8c-e17ade1d44d3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.060526022s
-Sep 21 17:00:24.396: INFO: Pod "pod-secrets-efef628d-2c27-46fc-8b8c-e17ade1d44d3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.084471348s
+STEP: Creating a pod to test downward api env vars
+Oct 26 17:13:29.243: INFO: Waiting up to 5m0s for pod "downward-api-7a3965a4-97bb-41e3-95c5-95e802b65545" in namespace "downward-api-7700" to be "success or failure"
+Oct 26 17:13:29.265: INFO: Pod "downward-api-7a3965a4-97bb-41e3-95c5-95e802b65545": Phase="Pending", Reason="", readiness=false. Elapsed: 22.346797ms
+Oct 26 17:13:31.282: INFO: Pod "downward-api-7a3965a4-97bb-41e3-95c5-95e802b65545": Phase="Pending", Reason="", readiness=false. Elapsed: 2.039245071s
+Oct 26 17:13:33.303: INFO: Pod "downward-api-7a3965a4-97bb-41e3-95c5-95e802b65545": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.060450028s
 STEP: Saw pod success
-Sep 21 17:00:24.396: INFO: Pod "pod-secrets-efef628d-2c27-46fc-8b8c-e17ade1d44d3" satisfied condition "success or failure"
-Sep 21 17:00:24.417: INFO: Trying to get logs from node 10.241.51.147 pod pod-secrets-efef628d-2c27-46fc-8b8c-e17ade1d44d3 container secret-env-test: 
+Oct 26 17:13:33.303: INFO: Pod "downward-api-7a3965a4-97bb-41e3-95c5-95e802b65545" satisfied condition "success or failure"
+Oct 26 17:13:33.320: INFO: Trying to get logs from node 10.72.119.74 pod downward-api-7a3965a4-97bb-41e3-95c5-95e802b65545 container dapi-container: 
 STEP: delete the pod
-Sep 21 17:00:24.536: INFO: Waiting for pod pod-secrets-efef628d-2c27-46fc-8b8c-e17ade1d44d3 to disappear
-Sep 21 17:00:24.556: INFO: Pod pod-secrets-efef628d-2c27-46fc-8b8c-e17ade1d44d3 no longer exists
-[AfterEach] [sig-api-machinery] Secrets
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:00:24.557: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-6215" for this suite.
-•{"msg":"PASSED [sig-api-machinery] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance]","total":280,"completed":266,"skipped":4300,"failed":0}
-SSS
-------------------------------
-[k8s.io] Pods 
-  should support retrieving logs from the container over websockets [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 17:00:24.609: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename pods
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
-[It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 17:00:24.881: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: creating the pod
-STEP: submitting the pod to kubernetes
-[AfterEach] [k8s.io] Pods
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:00:27.168: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-555" for this suite.
-•{"msg":"PASSED [k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]","total":280,"completed":267,"skipped":4303,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  works for multiple CRDs of same group but different versions [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 17:00:27.228: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename crd-publish-openapi
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] works for multiple CRDs of same group but different versions [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: CRs in the same group but different versions (one multiversion CRD) show up in OpenAPI documentation
-Sep 21 17:00:27.468: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: CRs in the same group but different versions (two CRDs) show up in OpenAPI documentation
-Sep 21 17:00:55.429: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-Sep 21 17:01:03.932: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+Oct 26 17:13:33.406: INFO: Waiting for pod downward-api-7a3965a4-97bb-41e3-95c5-95e802b65545 to disappear
+Oct 26 17:13:33.420: INFO: Pod downward-api-7a3965a4-97bb-41e3-95c5-95e802b65545 no longer exists
+[AfterEach] [sig-node] Downward API
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:01:35.405: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-1790" for this suite.
+Oct 26 17:13:33.420: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-7700" for this suite.
 
-• [SLOW TEST:68.233 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for multiple CRDs of same group but different versions [Conformance]
+• [SLOW TEST:5.448 seconds]
+[sig-node] Downward API
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:33
+  should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance]","total":280,"completed":268,"skipped":4331,"failed":0}
-SSSSSSSSS
-------------------------------
+{"msg":"PASSED [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]","total":280,"completed":269,"skipped":4404,"failed":0}
 [sig-network] Services 
-  should be able to change the type from ExternalName to NodePort [Conformance]
+  should be able to create a functioning NodePort service [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 17:01:35.461: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
+Oct 26 17:13:33.458: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
 STEP: Building a namespace api object, basename services
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should be able to change the type from ExternalName to NodePort [Conformance]
+[It] should be able to create a functioning NodePort service [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a service externalname-service with the type=ExternalName in namespace services-4685
-STEP: changing the ExternalName service to type=NodePort
-STEP: creating replication controller externalname-service in namespace services-4685
-I0921 17:01:35.948222      24 runners.go:189] Created replication controller with name: externalname-service, namespace: services-4685, replica count: 2
-Sep 21 17:01:38.998: INFO: Creating new exec pod
-I0921 17:01:38.998713      24 runners.go:189] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-Sep 21 17:01:42.320: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-4685 execpodqj7s9 -- /bin/sh -x -c nc -zv -t -w 2 externalname-service 80'
-Sep 21 17:01:42.679: INFO: stderr: "+ nc -zv -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n"
-Sep 21 17:01:42.679: INFO: stdout: ""
-Sep 21 17:01:42.680: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-4685 execpodqj7s9 -- /bin/sh -x -c nc -zv -t -w 2 172.21.243.140 80'
-Sep 21 17:01:43.013: INFO: stderr: "+ nc -zv -t -w 2 172.21.243.140 80\nConnection to 172.21.243.140 80 port [tcp/http] succeeded!\n"
-Sep 21 17:01:43.013: INFO: stdout: ""
-Sep 21 17:01:43.014: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-4685 execpodqj7s9 -- /bin/sh -x -c nc -zv -t -w 2 10.241.51.150 31569'
-Sep 21 17:01:43.354: INFO: stderr: "+ nc -zv -t -w 2 10.241.51.150 31569\nConnection to 10.241.51.150 31569 port [tcp/31569] succeeded!\n"
-Sep 21 17:01:43.354: INFO: stdout: ""
-Sep 21 17:01:43.354: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-4685 execpodqj7s9 -- /bin/sh -x -c nc -zv -t -w 2 10.241.51.149 31569'
-Sep 21 17:01:43.726: INFO: stderr: "+ nc -zv -t -w 2 10.241.51.149 31569\nConnection to 10.241.51.149 31569 port [tcp/31569] succeeded!\n"
-Sep 21 17:01:43.726: INFO: stdout: ""
-Sep 21 17:01:43.726: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-4685 execpodqj7s9 -- /bin/sh -x -c nc -zv -t -w 2 169.48.196.37 31569'
-Sep 21 17:01:44.074: INFO: stderr: "+ nc -zv -t -w 2 169.48.196.37 31569\nConnection to 169.48.196.37 31569 port [tcp/31569] succeeded!\n"
-Sep 21 17:01:44.074: INFO: stdout: ""
-Sep 21 17:01:44.074: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 exec --namespace=services-4685 execpodqj7s9 -- /bin/sh -x -c nc -zv -t -w 2 169.48.196.46 31569'
-Sep 21 17:01:44.429: INFO: stderr: "+ nc -zv -t -w 2 169.48.196.46 31569\nConnection to 169.48.196.46 31569 port [tcp/31569] succeeded!\n"
-Sep 21 17:01:44.429: INFO: stdout: ""
-Sep 21 17:01:44.429: INFO: Cleaning up the ExternalName to NodePort test service
+STEP: creating service nodeport-test with type=NodePort in namespace services-8313
+STEP: creating replication controller nodeport-test in namespace services-8313
+I1026 17:13:33.680902      26 runners.go:189] Created replication controller with name: nodeport-test, namespace: services-8313, replica count: 2
+I1026 17:13:36.733724      26 runners.go:189] nodeport-test Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+Oct 26 17:13:36.733: INFO: Creating new exec pod
+Oct 26 17:13:39.832: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-8313 execpodrmbjr -- /bin/sh -x -c nc -zv -t -w 2 nodeport-test 80'
+Oct 26 17:13:40.227: INFO: stderr: "+ nc -zv -t -w 2 nodeport-test 80\nConnection to nodeport-test 80 port [tcp/http] succeeded!\n"
+Oct 26 17:13:40.227: INFO: stdout: ""
+Oct 26 17:13:40.227: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-8313 execpodrmbjr -- /bin/sh -x -c nc -zv -t -w 2 172.21.92.169 80'
+Oct 26 17:13:40.580: INFO: stderr: "+ nc -zv -t -w 2 172.21.92.169 80\nConnection to 172.21.92.169 80 port [tcp/http] succeeded!\n"
+Oct 26 17:13:40.580: INFO: stdout: ""
+Oct 26 17:13:40.580: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-8313 execpodrmbjr -- /bin/sh -x -c nc -zv -t -w 2 10.72.119.76 32423'
+Oct 26 17:13:40.990: INFO: stderr: "+ nc -zv -t -w 2 10.72.119.76 32423\nConnection to 10.72.119.76 32423 port [tcp/32423] succeeded!\n"
+Oct 26 17:13:40.990: INFO: stdout: ""
+Oct 26 17:13:40.990: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-8313 execpodrmbjr -- /bin/sh -x -c nc -zv -t -w 2 10.72.119.74 32423'
+Oct 26 17:13:41.371: INFO: stderr: "+ nc -zv -t -w 2 10.72.119.74 32423\nConnection to 10.72.119.74 32423 port [tcp/32423] succeeded!\n"
+Oct 26 17:13:41.371: INFO: stdout: ""
+Oct 26 17:13:41.371: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-8313 execpodrmbjr -- /bin/sh -x -c nc -zv -t -w 2 158.176.65.138 32423'
+Oct 26 17:13:41.734: INFO: stderr: "+ nc -zv -t -w 2 158.176.65.138 32423\nConnection to 158.176.65.138 32423 port [tcp/32423] succeeded!\n"
+Oct 26 17:13:41.734: INFO: stdout: ""
+Oct 26 17:13:41.734: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 exec --namespace=services-8313 execpodrmbjr -- /bin/sh -x -c nc -zv -t -w 2 158.176.65.134 32423'
+Oct 26 17:13:42.106: INFO: stderr: "+ nc -zv -t -w 2 158.176.65.134 32423\nConnection to 158.176.65.134 32423 port [tcp/32423] succeeded!\n"
+Oct 26 17:13:42.106: INFO: stdout: ""
 [AfterEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:01:44.583: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-4685" for this suite.
+Oct 26 17:13:42.106: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-8313" for this suite.
 [AfterEach] [sig-network] Services
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
 
-• [SLOW TEST:9.193 seconds]
+• [SLOW TEST:8.684 seconds]
 [sig-network] Services
 /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should be able to change the type from ExternalName to NodePort [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance]","total":280,"completed":269,"skipped":4340,"failed":0}
-SSSSSSSSSSSSSSS
-------------------------------
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  works for CRD with validation schema [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Sep 21 17:01:44.654: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename crd-publish-openapi
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] works for CRD with validation schema [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 17:01:44.895: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: client-side validation (kubectl create and apply) allows request with known and required properties
-Sep 21 17:01:53.497: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-402 create -f -'
-Sep 21 17:01:54.253: INFO: stderr: ""
-Sep 21 17:01:54.253: INFO: stdout: "e2e-test-crd-publish-openapi-1626-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n"
-Sep 21 17:01:54.253: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-402 delete e2e-test-crd-publish-openapi-1626-crds test-foo'
-Sep 21 17:01:54.417: INFO: stderr: ""
-Sep 21 17:01:54.417: INFO: stdout: "e2e-test-crd-publish-openapi-1626-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n"
-Sep 21 17:01:54.417: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-402 apply -f -'
-Sep 21 17:01:55.106: INFO: stderr: ""
-Sep 21 17:01:55.106: INFO: stdout: "e2e-test-crd-publish-openapi-1626-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n"
-Sep 21 17:01:55.106: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-402 delete e2e-test-crd-publish-openapi-1626-crds test-foo'
-Sep 21 17:01:55.279: INFO: stderr: ""
-Sep 21 17:01:55.279: INFO: stdout: "e2e-test-crd-publish-openapi-1626-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n"
-STEP: client-side validation (kubectl create and apply) rejects request with unknown properties when disallowed by the schema
-Sep 21 17:01:55.279: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-402 create -f -'
-Sep 21 17:01:55.732: INFO: rc: 1
-Sep 21 17:01:55.732: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-402 apply -f -'
-Sep 21 17:01:56.317: INFO: rc: 1
-STEP: client-side validation (kubectl create and apply) rejects request without required properties
-Sep 21 17:01:56.317: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-402 create -f -'
-Sep 21 17:01:56.853: INFO: rc: 1
-Sep 21 17:01:56.853: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 --namespace=crd-publish-openapi-402 apply -f -'
-Sep 21 17:01:57.115: INFO: rc: 1
-STEP: kubectl explain works to explain CR properties
-Sep 21 17:01:57.116: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 explain e2e-test-crd-publish-openapi-1626-crds'
-Sep 21 17:01:57.669: INFO: stderr: ""
-Sep 21 17:01:57.669: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-1626-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nDESCRIPTION:\n     Foo CRD for Testing\n\nFIELDS:\n   apiVersion\t\n     APIVersion defines the versioned schema of this representation of an\n     object. Servers should convert recognized schemas to the latest internal\n     value, and may reject unrecognized values. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n   kind\t\n     Kind is a string value representing the REST resource this object\n     represents. Servers may infer this from the endpoint the client submits\n     requests to. Cannot be updated. In CamelCase. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n   metadata\t\n     Standard object's metadata. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   spec\t\n     Specification of Foo\n\n   status\t\n     Status of Foo\n\n"
-STEP: kubectl explain works to explain CR properties recursively
-Sep 21 17:01:57.669: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 explain e2e-test-crd-publish-openapi-1626-crds.metadata'
-Sep 21 17:01:58.122: INFO: stderr: ""
-Sep 21 17:01:58.122: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-1626-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: metadata \n\nDESCRIPTION:\n     Standard object's metadata. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n     ObjectMeta is metadata that all persisted resources must have, which\n     includes all objects users must create.\n\nFIELDS:\n   annotations\t\n     Annotations is an unstructured key value map stored with a resource that\n     may be set by external tools to store and retrieve arbitrary metadata. They\n     are not queryable and should be preserved when modifying objects. More\n     info: http://kubernetes.io/docs/user-guide/annotations\n\n   clusterName\t\n     The name of the cluster which the object belongs to. This is used to\n     distinguish resources with same name and namespace in different clusters.\n     This field is not set anywhere right now and apiserver is going to ignore\n     it if set in create or update request.\n\n   creationTimestamp\t\n     CreationTimestamp is a timestamp representing the server time when this\n     object was created. It is not guaranteed to be set in happens-before order\n     across separate operations. Clients may not set this value. It is\n     represented in RFC3339 form and is in UTC. Populated by the system.\n     Read-only. Null for lists. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   deletionGracePeriodSeconds\t\n     Number of seconds allowed for this object to gracefully terminate before it\n     will be removed from the system. Only set when deletionTimestamp is also\n     set. May only be shortened. Read-only.\n\n   deletionTimestamp\t\n     DeletionTimestamp is RFC 3339 date and time at which this resource will be\n     deleted. This field is set by the server when a graceful deletion is\n     requested by the user, and is not directly settable by a client. The\n     resource is expected to be deleted (no longer visible from resource lists,\n     and not reachable by name) after the time in this field, once the\n     finalizers list is empty. As long as the finalizers list contains items,\n     deletion is blocked. Once the deletionTimestamp is set, this value may not\n     be unset or be set further into the future, although it may be shortened or\n     the resource may be deleted prior to this time. For example, a user may\n     request that a pod is deleted in 30 seconds. The Kubelet will react by\n     sending a graceful termination signal to the containers in the pod. After\n     that 30 seconds, the Kubelet will send a hard termination signal (SIGKILL)\n     to the container and after cleanup, remove the pod from the API. In the\n     presence of network partitions, this object may still exist after this\n     timestamp, until an administrator or automated process can determine the\n     resource is fully terminated. If not set, graceful deletion of the object\n     has not been requested. Populated by the system when a graceful deletion is\n     requested. Read-only. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   finalizers\t<[]string>\n     Must be empty before the object is deleted from the registry. Each entry is\n     an identifier for the responsible component that will remove the entry from\n     the list. If the deletionTimestamp of the object is non-nil, entries in\n     this list can only be removed. Finalizers may be processed and removed in\n     any order. Order is NOT enforced because it introduces significant risk of\n     stuck finalizers. finalizers is a shared field, any actor with permission\n     can reorder it. If the finalizer list is processed in order, then this can\n     lead to a situation in which the component responsible for the first\n     finalizer in the list is waiting for a signal (field value, external\n     system, or other) produced by a component responsible for a finalizer later\n     in the list, resulting in a deadlock. Without enforced ordering finalizers\n     are free to order amongst themselves and are not vulnerable to ordering\n     changes in the list.\n\n   generateName\t\n     GenerateName is an optional prefix, used by the server, to generate a\n     unique name ONLY IF the Name field has not been provided. If this field is\n     used, the name returned to the client will be different than the name\n     passed. This value will also be combined with a unique suffix. The provided\n     value has the same validation rules as the Name field, and may be truncated\n     by the length of the suffix required to make the value unique on the\n     server. If this field is specified and the generated name exists, the\n     server will NOT return a 409 - instead, it will either return 201 Created\n     or 500 with Reason ServerTimeout indicating a unique name could not be\n     found in the time allotted, and the client should retry (optionally after\n     the time indicated in the Retry-After header). Applied only if Name is not\n     specified. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency\n\n   generation\t\n     A sequence number representing a specific generation of the desired state.\n     Populated by the system. Read-only.\n\n   labels\t\n     Map of string keys and values that can be used to organize and categorize\n     (scope and select) objects. May match selectors of replication controllers\n     and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n   managedFields\t<[]Object>\n     ManagedFields maps workflow-id and version to the set of fields that are\n     managed by that workflow. This is mostly for internal housekeeping, and\n     users typically shouldn't need to set or understand this field. A workflow\n     can be the user's name, a controller's name, or the name of a specific\n     apply path like \"ci-cd\". The set of fields is always in the version that\n     the workflow used when modifying the object.\n\n   name\t\n     Name must be unique within a namespace. Is required when creating\n     resources, although some resources may allow a client to request the\n     generation of an appropriate name automatically. Name is primarily intended\n     for creation idempotence and configuration definition. Cannot be updated.\n     More info: http://kubernetes.io/docs/user-guide/identifiers#names\n\n   namespace\t\n     Namespace defines the space within each name must be unique. An empty\n     namespace is equivalent to the \"default\" namespace, but \"default\" is the\n     canonical representation. Not all objects are required to be scoped to a\n     namespace - the value of this field for those objects will be empty. Must\n     be a DNS_LABEL. Cannot be updated. More info:\n     http://kubernetes.io/docs/user-guide/namespaces\n\n   ownerReferences\t<[]Object>\n     List of objects depended by this object. If ALL objects in the list have\n     been deleted, this object will be garbage collected. If this object is\n     managed by a controller, then an entry in this list will point to this\n     controller, with the controller field set to true. There cannot be more\n     than one managing controller.\n\n   resourceVersion\t\n     An opaque value that represents the internal version of this object that\n     can be used by clients to determine when objects have changed. May be used\n     for optimistic concurrency, change detection, and the watch operation on a\n     resource or set of resources. Clients must treat these values as opaque and\n     passed unmodified back to the server. They may only be valid for a\n     particular resource or set of resources. Populated by the system.\n     Read-only. Value must be treated as opaque by clients and . More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency\n\n   selfLink\t\n     SelfLink is a URL representing this object. Populated by the system.\n     Read-only. DEPRECATED Kubernetes will stop propagating this field in 1.20\n     release and the field is planned to be removed in 1.21 release.\n\n   uid\t\n     UID is the unique in time and space value for this object. It is typically\n     generated by the server on successful creation of a resource and is not\n     allowed to change on PUT operations. Populated by the system. Read-only.\n     More info: http://kubernetes.io/docs/user-guide/identifiers#uids\n\n"
-Sep 21 17:01:58.123: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 explain e2e-test-crd-publish-openapi-1626-crds.spec'
-Sep 21 17:01:58.421: INFO: stderr: ""
-Sep 21 17:01:58.421: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-1626-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: spec \n\nDESCRIPTION:\n     Specification of Foo\n\nFIELDS:\n   bars\t<[]Object>\n     List of Bars and their specs.\n\n"
-Sep 21 17:01:58.421: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 explain e2e-test-crd-publish-openapi-1626-crds.spec.bars'
-Sep 21 17:01:58.973: INFO: stderr: ""
-Sep 21 17:01:58.973: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-1626-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: bars <[]Object>\n\nDESCRIPTION:\n     List of Bars and their specs.\n\nFIELDS:\n   age\t\n     Age of Bar.\n\n   bazs\t<[]string>\n     List of Bazs.\n\n   name\t -required-\n     Name of Bar.\n\n"
-STEP: kubectl explain works to return error when explain is called on property that doesn't exist
-Sep 21 17:01:58.973: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-468631458 explain e2e-test-crd-publish-openapi-1626-crds.spec.bars2'
-Sep 21 17:01:59.426: INFO: rc: 1
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:02:08.398: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-402" for this suite.
-
-• [SLOW TEST:23.811 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for CRD with validation schema [Conformance]
+  should be able to create a functioning NodePort service [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance]","total":280,"completed":270,"skipped":4355,"failed":0}
-SSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-network] Services should be able to create a functioning NodePort service [Conformance]","total":280,"completed":270,"skipped":4404,"failed":0}
+SSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Namespaces [Serial] 
-  should ensure that all services are removed when a namespace is deleted [Conformance]
+[sig-network] Proxy version v1 
+  should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Namespaces [Serial]
+[BeforeEach] version v1
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 17:02:08.466: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename namespaces
+Oct 26 17:13:42.143: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename proxy
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should ensure that all services are removed when a namespace is deleted [Conformance]
+[It] should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a test namespace
-STEP: Waiting for a default service account to be provisioned in namespace
-STEP: Creating a service in the namespace
-STEP: Deleting the namespace
-STEP: Waiting for the namespace to be removed.
-STEP: Recreating the namespace
-STEP: Verifying there is no service in the namespace
-[AfterEach] [sig-api-machinery] Namespaces [Serial]
+Oct 26 17:13:42.340: INFO: (0) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 32.427706ms)
+Oct 26 17:13:42.359: INFO: (1) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 19.380377ms)
+Oct 26 17:13:42.379: INFO: (2) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 20.290701ms)
+Oct 26 17:13:42.403: INFO: (3) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 23.731716ms)
+Oct 26 17:13:42.423: INFO: (4) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 20.12661ms)
+Oct 26 17:13:42.448: INFO: (5) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 24.443145ms)
+Oct 26 17:13:42.470: INFO: (6) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 21.967779ms)
+Oct 26 17:13:42.506: INFO: (7) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 36.232752ms)
+Oct 26 17:13:42.529: INFO: (8) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 23.319799ms)
+Oct 26 17:13:42.552: INFO: (9) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 22.958174ms)
+Oct 26 17:13:42.572: INFO: (10) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 19.46257ms)
+Oct 26 17:13:42.592: INFO: (11) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 20.219542ms)
+Oct 26 17:13:42.612: INFO: (12) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 20.126489ms)
+Oct 26 17:13:42.631: INFO: (13) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 19.092201ms)
+Oct 26 17:13:42.652: INFO: (14) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 20.317559ms)
+Oct 26 17:13:42.672: INFO: (15) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 19.720222ms)
+Oct 26 17:13:42.692: INFO: (16) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 20.820721ms)
+Oct 26 17:13:42.715: INFO: (17) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 22.29034ms)
+Oct 26 17:13:42.736: INFO: (18) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 20.845668ms)
+Oct 26 17:13:42.756: INFO: (19) /api/v1/nodes/10.72.119.74:10250/proxy/logs/: 
+anaconda/
+at/
+at-no-rotat... (200; 20.626283ms)
+[AfterEach] version v1
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:02:17.469: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "namespaces-154" for this suite.
-STEP: Destroying namespace "nsdeletetest-2434" for this suite.
-Sep 21 17:02:17.586: INFO: Namespace nsdeletetest-2434 was already deleted
-STEP: Destroying namespace "nsdeletetest-6004" for this suite.
-
-• [SLOW TEST:9.161 seconds]
-[sig-api-machinery] Namespaces [Serial]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should ensure that all services are removed when a namespace is deleted [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance]","total":280,"completed":271,"skipped":4371,"failed":0}
-SSSSSSSSSSSSSSSS
+Oct 26 17:13:42.757: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "proxy-952" for this suite.
+•{"msg":"PASSED [sig-network] Proxy version v1 should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]","total":280,"completed":271,"skipped":4417,"failed":0}
+SSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Probing container 
-  should have monotonically increasing restart count [NodeConformance] [Conformance]
+[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition 
+  creating/deleting custom resource definition objects works  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Probing container
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 17:02:17.627: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename container-probe
+Oct 26 17:13:42.796: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename custom-resource-definition
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] should have monotonically increasing restart count [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod liveness-ed2f19d8-144a-465f-bb83-241e39b93770 in namespace container-probe-4647
-Sep 21 17:02:20.012: INFO: Started pod liveness-ed2f19d8-144a-465f-bb83-241e39b93770 in namespace container-probe-4647
-STEP: checking the pod's current state and verifying that restartCount is present
-Sep 21 17:02:20.034: INFO: Initial restart count of pod liveness-ed2f19d8-144a-465f-bb83-241e39b93770 is 0
-Sep 21 17:02:34.239: INFO: Restart count of pod container-probe-4647/liveness-ed2f19d8-144a-465f-bb83-241e39b93770 is now 1 (14.205066187s elapsed)
-Sep 21 17:02:52.446: INFO: Restart count of pod container-probe-4647/liveness-ed2f19d8-144a-465f-bb83-241e39b93770 is now 2 (32.41247442s elapsed)
-Sep 21 17:03:12.700: INFO: Restart count of pod container-probe-4647/liveness-ed2f19d8-144a-465f-bb83-241e39b93770 is now 3 (52.665743594s elapsed)
-Sep 21 17:03:32.956: INFO: Restart count of pod container-probe-4647/liveness-ed2f19d8-144a-465f-bb83-241e39b93770 is now 4 (1m12.922233979s elapsed)
-Sep 21 17:04:40.200: INFO: Restart count of pod container-probe-4647/liveness-ed2f19d8-144a-465f-bb83-241e39b93770 is now 5 (2m20.166709487s elapsed)
-STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:04:40.282: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-4647" for this suite.
-
-• [SLOW TEST:142.718 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should have monotonically increasing restart count [NodeConformance] [Conformance]
+[It] creating/deleting custom resource definition objects works  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Oct 26 17:13:42.921: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Oct 26 17:13:44.019: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "custom-resource-definition-6172" for this suite.
+•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works  [Conformance]","total":280,"completed":272,"skipped":4430,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-{"msg":"PASSED [k8s.io] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance]","total":280,"completed":272,"skipped":4387,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-network] DNS 
-  should provide DNS for pods for Subdomain [Conformance]
+[sig-cli] Kubectl client Update Demo 
+  should do a rolling update of a replication controller  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] DNS
+[BeforeEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 17:04:40.345: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename dns
+Oct 26 17:13:44.061: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename kubectl
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide DNS for pods for Subdomain [Conformance]
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:278
+[BeforeEach] Update Demo
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:330
+[It] should do a rolling update of a replication controller  [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a test headless service
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-6105.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service-2.dns-6105.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-6105.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service-2.dns-6105.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6105.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
+STEP: creating the initial replication controller
+Oct 26 17:13:44.205: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 create -f - --namespace=kubectl-9089'
+Oct 26 17:13:44.770: INFO: stderr: ""
+Oct 26 17:13:44.770: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Oct 26 17:13:44.770: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9089'
+Oct 26 17:13:44.930: INFO: stderr: ""
+Oct 26 17:13:44.930: INFO: stdout: "update-demo-nautilus-c8grz update-demo-nautilus-glsf6 "
+Oct 26 17:13:44.930: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-c8grz -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9089'
+Oct 26 17:13:45.070: INFO: stderr: ""
+Oct 26 17:13:45.070: INFO: stdout: ""
+Oct 26 17:13:45.070: INFO: update-demo-nautilus-c8grz is created but not running
+Oct 26 17:13:50.070: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9089'
+Oct 26 17:13:50.211: INFO: stderr: ""
+Oct 26 17:13:50.211: INFO: stdout: "update-demo-nautilus-c8grz update-demo-nautilus-glsf6 "
+Oct 26 17:13:50.211: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-c8grz -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9089'
+Oct 26 17:13:50.360: INFO: stderr: ""
+Oct 26 17:13:50.360: INFO: stdout: "true"
+Oct 26 17:13:50.360: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-c8grz -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9089'
+Oct 26 17:13:50.496: INFO: stderr: ""
+Oct 26 17:13:50.496: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Oct 26 17:13:50.496: INFO: validating pod update-demo-nautilus-c8grz
+Oct 26 17:13:50.532: INFO: got data: {
+  "image": "nautilus.jpg"
+}
 
-STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-6105.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service-2.dns-6105.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-6105.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service-2.dns-6105.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6105.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+Oct 26 17:13:50.532: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Oct 26 17:13:50.532: INFO: update-demo-nautilus-c8grz is verified up and running
+Oct 26 17:13:50.532: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-glsf6 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9089'
+Oct 26 17:13:50.691: INFO: stderr: ""
+Oct 26 17:13:50.691: INFO: stdout: "true"
+Oct 26 17:13:50.691: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-nautilus-glsf6 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9089'
+Oct 26 17:13:50.828: INFO: stderr: ""
+Oct 26 17:13:50.828: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Oct 26 17:13:50.828: INFO: validating pod update-demo-nautilus-glsf6
+Oct 26 17:13:50.860: INFO: got data: {
+  "image": "nautilus.jpg"
+}
 
-STEP: creating a pod to probe DNS
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Sep 21 17:04:44.861: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local from pod dns-6105/dns-test-ffb48184-bc2f-4746-b523-3d551a89693b: the server could not find the requested resource (get pods dns-test-ffb48184-bc2f-4746-b523-3d551a89693b)
-Sep 21 17:04:44.911: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local from pod dns-6105/dns-test-ffb48184-bc2f-4746-b523-3d551a89693b: the server could not find the requested resource (get pods dns-test-ffb48184-bc2f-4746-b523-3d551a89693b)
-Sep 21 17:04:45.198: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local from pod dns-6105/dns-test-ffb48184-bc2f-4746-b523-3d551a89693b: the server could not find the requested resource (get pods dns-test-ffb48184-bc2f-4746-b523-3d551a89693b)
-Sep 21 17:04:45.319: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-6105.svc.cluster.local from pod dns-6105/dns-test-ffb48184-bc2f-4746-b523-3d551a89693b: the server could not find the requested resource (get pods dns-test-ffb48184-bc2f-4746-b523-3d551a89693b)
-Sep 21 17:04:45.414: INFO: Lookups using dns-6105/dns-test-ffb48184-bc2f-4746-b523-3d551a89693b failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-6105.svc.cluster.local jessie_tcp@dns-test-service-2.dns-6105.svc.cluster.local]
+Oct 26 17:13:50.860: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Oct 26 17:13:50.860: INFO: update-demo-nautilus-glsf6 is verified up and running
+STEP: rolling-update to new replication controller
+Oct 26 17:13:50.866: INFO: scanned /root for discovery docs: 
+Oct 26 17:13:50.866: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 rolling-update update-demo-nautilus --update-period=1s -f - --namespace=kubectl-9089'
+Oct 26 17:14:15.291: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n"
+Oct 26 17:14:15.291: INFO: stdout: "Created update-demo-kitten\nScaling up update-demo-kitten from 0 to 2, scaling down update-demo-nautilus from 2 to 0 (keep 2 pods available, don't exceed 3 pods)\nScaling update-demo-kitten up to 1\nScaling update-demo-nautilus down to 1\nScaling update-demo-kitten up to 2\nScaling update-demo-nautilus down to 0\nUpdate succeeded. Deleting old controller: update-demo-nautilus\nRenaming update-demo-kitten to update-demo-nautilus\nreplicationcontroller/update-demo-nautilus rolling updated\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Oct 26 17:14:15.291: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9089'
+Oct 26 17:14:15.460: INFO: stderr: ""
+Oct 26 17:14:15.460: INFO: stdout: "update-demo-kitten-htfsm update-demo-kitten-vsp59 update-demo-nautilus-glsf6 "
+STEP: Replicas for name=update-demo: expected=2 actual=3
+Oct 26 17:14:20.460: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9089'
+Oct 26 17:14:20.601: INFO: stderr: ""
+Oct 26 17:14:20.601: INFO: stdout: "update-demo-kitten-htfsm update-demo-kitten-vsp59 "
+Oct 26 17:14:20.601: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-kitten-htfsm -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9089'
+Oct 26 17:14:20.747: INFO: stderr: ""
+Oct 26 17:14:20.747: INFO: stdout: "true"
+Oct 26 17:14:20.747: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-kitten-htfsm -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9089'
+Oct 26 17:14:20.878: INFO: stderr: ""
+Oct 26 17:14:20.878: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
+Oct 26 17:14:20.878: INFO: validating pod update-demo-kitten-htfsm
+Oct 26 17:14:20.911: INFO: got data: {
+  "image": "kitten.jpg"
+}
 
-Sep 21 17:04:51.230: INFO: DNS probes using dns-6105/dns-test-ffb48184-bc2f-4746-b523-3d551a89693b succeeded
+Oct 26 17:14:20.911: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
+Oct 26 17:14:20.911: INFO: update-demo-kitten-htfsm is verified up and running
+Oct 26 17:14:20.911: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-kitten-vsp59 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9089'
+Oct 26 17:14:21.043: INFO: stderr: ""
+Oct 26 17:14:21.043: INFO: stdout: "true"
+Oct 26 17:14:21.043: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-871182394 get pods update-demo-kitten-vsp59 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9089'
+Oct 26 17:14:21.192: INFO: stderr: ""
+Oct 26 17:14:21.192: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
+Oct 26 17:14:21.192: INFO: validating pod update-demo-kitten-vsp59
+Oct 26 17:14:21.225: INFO: got data: {
+  "image": "kitten.jpg"
+}
 
-STEP: deleting the pod
-STEP: deleting the test headless service
-[AfterEach] [sig-network] DNS
+Oct 26 17:14:21.225: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
+Oct 26 17:14:21.225: INFO: update-demo-kitten-vsp59 is verified up and running
+[AfterEach] [sig-cli] Kubectl client
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:04:51.740: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-6105" for this suite.
+Oct 26 17:14:21.225: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-9089" for this suite.
 
-• [SLOW TEST:11.464 seconds]
-[sig-network] DNS
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should provide DNS for pods for Subdomain [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:37.221 seconds]
+[sig-cli] Kubectl client
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Update Demo
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:328
+    should do a rolling update of a replication controller  [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Subdomain [Conformance]","total":280,"completed":273,"skipped":4418,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-cli] Kubectl client Update Demo should do a rolling update of a replication controller  [Conformance]","total":280,"completed":273,"skipped":4456,"failed":0}
+SSSSS
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should mutate custom resource with pruning [Conformance]
+[sig-storage] Projected downwardAPI 
+  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+[BeforeEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 17:04:51.810: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename webhook
+Oct 26 17:14:21.282: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename projected
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Sep 21 17:04:53.155: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736304693, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736304693, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736304693, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736304693, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Sep 21 17:04:55.266: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736304693, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736304693, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63736304693, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63736304693, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Sep 21 17:04:58.305: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should mutate custom resource with pruning [Conformance]
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 17:04:58.336: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Registering the mutating webhook for custom resource e2e-test-webhook-3518-crds.webhook.example.com via the AdmissionRegistration API
-STEP: Creating a custom resource that should be mutated by the webhook
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+STEP: Creating a pod to test downward API volume plugin
+Oct 26 17:14:21.486: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9eb6bdb8-df15-4aaa-ade8-ee3e4eb9a804" in namespace "projected-6291" to be "success or failure"
+Oct 26 17:14:21.503: INFO: Pod "downwardapi-volume-9eb6bdb8-df15-4aaa-ade8-ee3e4eb9a804": Phase="Pending", Reason="", readiness=false. Elapsed: 17.099695ms
+Oct 26 17:14:23.519: INFO: Pod "downwardapi-volume-9eb6bdb8-df15-4aaa-ade8-ee3e4eb9a804": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033308579s
+Oct 26 17:14:25.536: INFO: Pod "downwardapi-volume-9eb6bdb8-df15-4aaa-ade8-ee3e4eb9a804": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.050791456s
+STEP: Saw pod success
+Oct 26 17:14:25.536: INFO: Pod "downwardapi-volume-9eb6bdb8-df15-4aaa-ade8-ee3e4eb9a804" satisfied condition "success or failure"
+Oct 26 17:14:25.550: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-9eb6bdb8-df15-4aaa-ade8-ee3e4eb9a804 container client-container: 
+STEP: delete the pod
+Oct 26 17:14:25.631: INFO: Waiting for pod downwardapi-volume-9eb6bdb8-df15-4aaa-ade8-ee3e4eb9a804 to disappear
+Oct 26 17:14:25.647: INFO: Pod downwardapi-volume-9eb6bdb8-df15-4aaa-ade8-ee3e4eb9a804 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:04:59.616: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-8643" for this suite.
-STEP: Destroying namespace "webhook-8643-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
-
-• [SLOW TEST:8.353 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should mutate custom resource with pruning [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance]","total":280,"completed":274,"skipped":4465,"failed":0}
+Oct 26 17:14:25.647: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-6291" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":274,"skipped":4461,"failed":0}
 SSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir wrapper volumes 
-  should not conflict [Conformance]
+[sig-storage] Downward API volume 
+  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir wrapper volumes
+[BeforeEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 17:05:00.163: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename emptydir-wrapper
+Oct 26 17:14:25.679: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename downward-api
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not conflict [Conformance]
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Cleaning up the secret
-STEP: Cleaning up the configmap
-STEP: Cleaning up the pod
-[AfterEach] [sig-storage] EmptyDir wrapper volumes
+STEP: Creating a pod to test downward API volume plugin
+Oct 26 17:14:25.893: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9e18154c-9387-4453-9ac2-7d883f0c004a" in namespace "downward-api-9415" to be "success or failure"
+Oct 26 17:14:25.908: INFO: Pod "downwardapi-volume-9e18154c-9387-4453-9ac2-7d883f0c004a": Phase="Pending", Reason="", readiness=false. Elapsed: 14.361343ms
+Oct 26 17:14:27.926: INFO: Pod "downwardapi-volume-9e18154c-9387-4453-9ac2-7d883f0c004a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.032791847s
+STEP: Saw pod success
+Oct 26 17:14:27.926: INFO: Pod "downwardapi-volume-9e18154c-9387-4453-9ac2-7d883f0c004a" satisfied condition "success or failure"
+Oct 26 17:14:27.943: INFO: Trying to get logs from node 10.72.119.74 pod downwardapi-volume-9e18154c-9387-4453-9ac2-7d883f0c004a container client-container: 
+STEP: delete the pod
+Oct 26 17:14:28.039: INFO: Waiting for pod downwardapi-volume-9e18154c-9387-4453-9ac2-7d883f0c004a to disappear
+Oct 26 17:14:28.054: INFO: Pod downwardapi-volume-9e18154c-9387-4453-9ac2-7d883f0c004a no longer exists
+[AfterEach] [sig-storage] Downward API volume
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:05:04.586: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-wrapper-8634" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance]","total":280,"completed":275,"skipped":4474,"failed":0}
-SSSSSSSSSSSSSSS
+Oct 26 17:14:28.055: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-9415" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":275,"skipped":4470,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  Should recreate evicted statefulset [Conformance]
+[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] 
+  should be able to convert from CR v1 to CR v2 [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] StatefulSet
+[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 17:05:04.653: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename statefulset
+Oct 26 17:14:28.101: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename crd-webhook
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
-STEP: Creating service test in namespace statefulset-9760
-[It] Should recreate evicted statefulset [Conformance]
+[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:125
+STEP: Setting up server cert
+STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication
+STEP: Deploying the custom resource conversion webhook pod
+STEP: Wait for the deployment to be ready
+Oct 26 17:14:29.240: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set
+Oct 26 17:14:31.292: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329269, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329269, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329269, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329269, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-78dcf5dd84\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Oct 26 17:14:34.346: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1
+[It] should be able to convert from CR v1 to CR v2 [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Looking for a node to schedule stateful set and pod
-STEP: Creating pod with conflicting port in namespace statefulset-9760
-STEP: Creating statefulset with conflicting port in namespace statefulset-9760
-STEP: Waiting until pod test-pod will start running in namespace statefulset-9760
-STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-9760
-Sep 21 17:05:09.518: INFO: Observed stateful pod in namespace: statefulset-9760, name: ss-0, uid: 0c258eaf-8b4a-49f0-91fd-580ecb164431, status phase: Pending. Waiting for statefulset controller to delete.
-Sep 21 17:05:09.518: INFO: Observed stateful pod in namespace: statefulset-9760, name: ss-0, uid: 0c258eaf-8b4a-49f0-91fd-580ecb164431, status phase: Failed. Waiting for statefulset controller to delete.
-Sep 21 17:05:09.658: INFO: Observed stateful pod in namespace: statefulset-9760, name: ss-0, uid: 0c258eaf-8b4a-49f0-91fd-580ecb164431, status phase: Failed. Waiting for statefulset controller to delete.
-Sep 21 17:05:09.658: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-9760
-STEP: Removing pod with conflicting port in namespace statefulset-9760
-STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-9760 and will be in running state
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
-Sep 21 17:05:13.860: INFO: Deleting all statefulset in ns statefulset-9760
-Sep 21 17:05:13.883: INFO: Scaling statefulset ss to 0
-Sep 21 17:05:23.985: INFO: Waiting for statefulset status.replicas updated to 0
-Sep 21 17:05:24.007: INFO: Deleting statefulset ss
-[AfterEach] [sig-apps] StatefulSet
+Oct 26 17:14:34.363: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Creating a v1 custom resource
+STEP: v2 custom resource should be converted
+[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:05:24.341: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-9760" for this suite.
+Oct 26 17:14:35.739: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-webhook-807" for this suite.
+[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:136
 
-• [SLOW TEST:19.824 seconds]
-[sig-apps] StatefulSet
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-    Should recreate evicted statefulset [Conformance]
-    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:7.898 seconds]
+[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should be able to convert from CR v1 to CR v2 [Conformance]
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance]","total":280,"completed":276,"skipped":4489,"failed":0}
-SSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance]","total":280,"completed":276,"skipped":4492,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] Job 
-  should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Job
+[BeforeEach] [sig-storage] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 17:05:24.477: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename job
+Oct 26 17:14:35.999: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename configmap
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a job
-STEP: Ensuring job reaches completions
-[AfterEach] [sig-apps] Job
+STEP: Creating configMap with name configmap-test-volume-6b982c07-32cf-43d6-96be-082b3e207be8
+STEP: Creating a pod to test consume configMaps
+Oct 26 17:14:36.240: INFO: Waiting up to 5m0s for pod "pod-configmaps-2f637ae6-9980-46a9-bec0-4298292fe1db" in namespace "configmap-1823" to be "success or failure"
+Oct 26 17:14:36.257: INFO: Pod "pod-configmaps-2f637ae6-9980-46a9-bec0-4298292fe1db": Phase="Pending", Reason="", readiness=false. Elapsed: 16.365202ms
+Oct 26 17:14:38.271: INFO: Pod "pod-configmaps-2f637ae6-9980-46a9-bec0-4298292fe1db": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030913406s
+Oct 26 17:14:40.288: INFO: Pod "pod-configmaps-2f637ae6-9980-46a9-bec0-4298292fe1db": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.047649929s
+STEP: Saw pod success
+Oct 26 17:14:40.288: INFO: Pod "pod-configmaps-2f637ae6-9980-46a9-bec0-4298292fe1db" satisfied condition "success or failure"
+Oct 26 17:14:40.304: INFO: Trying to get logs from node 10.72.119.74 pod pod-configmaps-2f637ae6-9980-46a9-bec0-4298292fe1db container configmap-volume-test: 
+STEP: delete the pod
+Oct 26 17:14:40.389: INFO: Waiting for pod pod-configmaps-2f637ae6-9980-46a9-bec0-4298292fe1db to disappear
+Oct 26 17:14:40.403: INFO: Pod pod-configmaps-2f637ae6-9980-46a9-bec0-4298292fe1db no longer exists
+[AfterEach] [sig-storage] ConfigMap
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:05:34.814: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "job-7545" for this suite.
-
-• [SLOW TEST:10.404 seconds]
-[sig-apps] Job
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]","total":280,"completed":277,"skipped":4502,"failed":0}
-SSSSS
+Oct 26 17:14:40.403: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-1823" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":277,"skipped":4527,"failed":0}
+SSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  updates the published spec when one version gets renamed [Conformance]
+[sig-apps] Deployment 
+  RollingUpdateDeployment should delete old pods and create new ones [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+[BeforeEach] [sig-apps] Deployment
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 17:05:34.883: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename crd-publish-openapi
+Oct 26 17:14:40.455: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename deployment
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] updates the published spec when one version gets renamed [Conformance]
+[BeforeEach] [sig-apps] Deployment
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
+[It] RollingUpdateDeployment should delete old pods and create new ones [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: set up a multi version CRD
-Sep 21 17:05:35.257: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: rename a version
-STEP: check the new version name is served
-STEP: check the old version name is removed
-STEP: check the other version is not changed
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+Oct 26 17:14:40.590: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted)
+Oct 26 17:14:40.642: INFO: Pod name sample-pod: Found 0 pods out of 1
+Oct 26 17:14:45.659: INFO: Pod name sample-pod: Found 1 pods out of 1
+STEP: ensuring each pod is running
+Oct 26 17:14:45.659: INFO: Creating deployment "test-rolling-update-deployment"
+Oct 26 17:14:45.680: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has
+Oct 26 17:14:45.712: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created
+Oct 26 17:14:47.758: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected
+Oct 26 17:14:47.774: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329285, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329285, loc:(*time.Location)(0x7db5bc0)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329285, loc:(*time.Location)(0x7db5bc0)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63739329285, loc:(*time.Location)(0x7db5bc0)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-67cf4f6444\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Oct 26 17:14:49.794: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted)
+[AfterEach] [sig-apps] Deployment
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
+Oct 26 17:14:49.852: INFO: Deployment "test-rolling-update-deployment":
+&Deployment{ObjectMeta:{test-rolling-update-deployment  deployment-3618 /apis/apps/v1/namespaces/deployment-3618/deployments/test-rolling-update-deployment 0eca3701-0c6d-477a-8683-8b1fcae66424 93423 1 2020-10-26 17:14:45 +0000 UTC   map[name:sample-pod] map[deployment.kubernetes.io/revision:3546343826724305833] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc004875bb8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-10-26 17:14:45 +0000 UTC,LastTransitionTime:2020-10-26 17:14:45 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rolling-update-deployment-67cf4f6444" has successfully progressed.,LastUpdateTime:2020-10-26 17:14:47 +0000 UTC,LastTransitionTime:2020-10-26 17:14:45 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},}
+
+Oct 26 17:14:49.869: INFO: New ReplicaSet "test-rolling-update-deployment-67cf4f6444" of Deployment "test-rolling-update-deployment":
+&ReplicaSet{ObjectMeta:{test-rolling-update-deployment-67cf4f6444  deployment-3618 /apis/apps/v1/namespaces/deployment-3618/replicasets/test-rolling-update-deployment-67cf4f6444 daee5097-41a5-4390-832c-8c119368a46c 93413 1 2020-10-26 17:14:45 +0000 UTC   map[name:sample-pod pod-template-hash:67cf4f6444] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305833] [{apps/v1 Deployment test-rolling-update-deployment 0eca3701-0c6d-477a-8683-8b1fcae66424 0xc0048b6077 0xc0048b6078}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 67cf4f6444,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod pod-template-hash:67cf4f6444] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0048b60e8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},}
+Oct 26 17:14:49.869: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment":
+Oct 26 17:14:49.869: INFO: &ReplicaSet{ObjectMeta:{test-rolling-update-controller  deployment-3618 /apis/apps/v1/namespaces/deployment-3618/replicasets/test-rolling-update-controller 25cb4b94-667c-429e-be33-4f8d41a0d159 93422 2 2020-10-26 17:14:40 +0000 UTC   map[name:sample-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305832] [{apps/v1 Deployment test-rolling-update-deployment 0eca3701-0c6d-477a-8683-8b1fcae66424 0xc004875fa7 0xc004875fa8}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod pod:httpd] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc0048b6008  ClusterFirst map[]     false false false  PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
+Oct 26 17:14:49.884: INFO: Pod "test-rolling-update-deployment-67cf4f6444-x79jg" is available:
+&Pod{ObjectMeta:{test-rolling-update-deployment-67cf4f6444-x79jg test-rolling-update-deployment-67cf4f6444- deployment-3618 /api/v1/namespaces/deployment-3618/pods/test-rolling-update-deployment-67cf4f6444-x79jg a3d84222-7c13-4a56-aa8f-1f0bc4ac3a29 93412 0 2020-10-26 17:14:45 +0000 UTC   map[name:sample-pod pod-template-hash:67cf4f6444] map[cni.projectcalico.org/podIP:172.30.194.87/32 cni.projectcalico.org/podIPs:172.30.194.87/32 k8s.v1.cni.cncf.io/networks-status:[{
+    "name": "k8s-pod-network",
+    "ips": [
+        "172.30.194.87"
+    ],
+    "dns": {}
+}] openshift.io/scc:privileged] [{apps/v1 ReplicaSet test-rolling-update-deployment-67cf4f6444 daee5097-41a5-4390-832c-8c119368a46c 0xc0048b65b7 0xc0048b65b8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-qc8p9,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-qc8p9,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-qc8p9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:10.72.119.74,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{LocalObjectReference{Name:default-dockercfg-sr4b2,},},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:14:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:14:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-10-26 17:14:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:10.72.119.74,PodIP:172.30.194.87,StartTime:2020-10-26 17:14:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-10-26 17:14:47 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:gcr.io/kubernetes-e2e-test-images/agnhost@sha256:02b9e10ea1f7439a03d9bc3ee16d984172ca7eac1818792ad4d721c5c8f72ff4,ContainerID:cri-o://a7c64f68e8112d0b37ab60eab888f9304910a800ed96ba26da97a65e098284ae,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.30.194.87,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+[AfterEach] [sig-apps] Deployment
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:06:18.217: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-5582" for this suite.
+Oct 26 17:14:49.884: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-3618" for this suite.
 
-• [SLOW TEST:43.391 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  updates the published spec when one version gets renamed [Conformance]
+• [SLOW TEST:9.474 seconds]
+[sig-apps] Deployment
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  RollingUpdateDeployment should delete old pods and create new ones [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance]","total":280,"completed":278,"skipped":4507,"failed":0}
-SSSSSSSSS
+{"msg":"PASSED [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance]","total":280,"completed":278,"skipped":4540,"failed":0}
+SSSSS
 ------------------------------
-[sig-storage] Projected configMap 
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute prestop http hook properly [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
+[BeforeEach] [k8s.io] Container Lifecycle Hook
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 17:06:18.273: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 17:14:49.931: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename container-lifecycle-hook
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute prestop http hook properly [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Sep 21 17:06:18.549: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
-STEP: Creating configMap with name cm-test-opt-del-c373fbb0-6859-4fe2-af03-09cfe4cde909
-STEP: Creating configMap with name cm-test-opt-upd-93e14696-0573-42ee-91ef-419ef2ba8fad
-STEP: Creating the pod
-STEP: Deleting configmap cm-test-opt-del-c373fbb0-6859-4fe2-af03-09cfe4cde909
-STEP: Updating configmap cm-test-opt-upd-93e14696-0573-42ee-91ef-419ef2ba8fad
-STEP: Creating configMap with name cm-test-opt-create-c08ffd79-aae2-4994-8191-c151a4fed353
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] Projected configMap
+STEP: create the pod with lifecycle hook
+STEP: delete the pod with lifecycle hook
+Oct 26 17:14:56.333: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Oct 26 17:14:56.348: INFO: Pod pod-with-prestop-http-hook still exists
+Oct 26 17:14:58.348: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Oct 26 17:14:58.363: INFO: Pod pod-with-prestop-http-hook still exists
+Oct 26 17:15:00.348: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Oct 26 17:15:00.365: INFO: Pod pod-with-prestop-http-hook still exists
+Oct 26 17:15:02.348: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Oct 26 17:15:02.363: INFO: Pod pod-with-prestop-http-hook still exists
+Oct 26 17:15:04.348: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Oct 26 17:15:04.365: INFO: Pod pod-with-prestop-http-hook still exists
+Oct 26 17:15:06.348: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Oct 26 17:15:06.365: INFO: Pod pod-with-prestop-http-hook still exists
+Oct 26 17:15:08.348: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Oct 26 17:15:08.364: INFO: Pod pod-with-prestop-http-hook no longer exists
+STEP: check prestop hook
+[AfterEach] [k8s.io] Container Lifecycle Hook
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:07:54.406: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-6788" for this suite.
+Oct 26 17:15:08.395: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-2566" for this suite.
 
-• [SLOW TEST:96.209 seconds]
-[sig-storage] Projected configMap
-/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:18.498 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  when create a pod with lifecycle hook
+  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
+    should execute prestop http hook properly [NodeConformance] [Conformance]
+    /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":279,"skipped":4516,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance]","total":280,"completed":279,"skipped":4545,"failed":0}
+SSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
+[BeforeEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Sep 21 17:07:54.485: INFO: >>> kubeConfig: /tmp/kubeconfig-468631458
-STEP: Building a namespace api object, basename projected
+Oct 26 17:15:08.429: INFO: >>> kubeConfig: /tmp/kubeconfig-871182394
+STEP: Building a namespace api object, basename emptydir
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
+[It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Sep 21 17:07:54.905: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7ad6d0d2-413f-4c79-9c8c-6aca1f4757bf" in namespace "projected-1338" to be "success or failure"
-Sep 21 17:07:54.926: INFO: Pod "downwardapi-volume-7ad6d0d2-413f-4c79-9c8c-6aca1f4757bf": Phase="Pending", Reason="", readiness=false. Elapsed: 21.092209ms
-Sep 21 17:07:56.956: INFO: Pod "downwardapi-volume-7ad6d0d2-413f-4c79-9c8c-6aca1f4757bf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.050765919s
+STEP: Creating a pod to test emptydir 0777 on tmpfs
+Oct 26 17:15:08.625: INFO: Waiting up to 5m0s for pod "pod-2a127306-d1d6-46eb-910e-db1514336b68" in namespace "emptydir-1976" to be "success or failure"
+Oct 26 17:15:08.642: INFO: Pod "pod-2a127306-d1d6-46eb-910e-db1514336b68": Phase="Pending", Reason="", readiness=false. Elapsed: 17.294604ms
+Oct 26 17:15:10.658: INFO: Pod "pod-2a127306-d1d6-46eb-910e-db1514336b68": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033457208s
+Oct 26 17:15:12.674: INFO: Pod "pod-2a127306-d1d6-46eb-910e-db1514336b68": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.048872522s
 STEP: Saw pod success
-Sep 21 17:07:56.956: INFO: Pod "downwardapi-volume-7ad6d0d2-413f-4c79-9c8c-6aca1f4757bf" satisfied condition "success or failure"
-Sep 21 17:07:56.981: INFO: Trying to get logs from node 10.241.51.147 pod downwardapi-volume-7ad6d0d2-413f-4c79-9c8c-6aca1f4757bf container client-container: 
+Oct 26 17:15:12.674: INFO: Pod "pod-2a127306-d1d6-46eb-910e-db1514336b68" satisfied condition "success or failure"
+Oct 26 17:15:12.688: INFO: Trying to get logs from node 10.72.119.74 pod pod-2a127306-d1d6-46eb-910e-db1514336b68 container test-container: 
 STEP: delete the pod
-Sep 21 17:07:57.096: INFO: Waiting for pod downwardapi-volume-7ad6d0d2-413f-4c79-9c8c-6aca1f4757bf to disappear
-Sep 21 17:07:57.116: INFO: Pod downwardapi-volume-7ad6d0d2-413f-4c79-9c8c-6aca1f4757bf no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
+Oct 26 17:15:12.769: INFO: Waiting for pod pod-2a127306-d1d6-46eb-910e-db1514336b68 to disappear
+Oct 26 17:15:12.784: INFO: Pod pod-2a127306-d1d6-46eb-910e-db1514336b68 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
   /workspace/anago-v1.17.1-beta.0.42+d224476cd0730b/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Sep 21 17:07:57.116: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-1338" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":280,"skipped":4553,"failed":0}
-SSSSSSSSSSSep 21 17:07:57.181: INFO: Running AfterSuite actions on all nodes
-Sep 21 17:07:57.181: INFO: Running AfterSuite actions on node 1
-Sep 21 17:07:57.181: INFO: Skipping dumping logs from cluster
+Oct 26 17:15:12.784: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-1976" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":280,"skipped":4562,"failed":0}
+SOct 26 17:15:12.820: INFO: Running AfterSuite actions on all nodes
+Oct 26 17:15:12.820: INFO: Running AfterSuite actions on node 1
+Oct 26 17:15:12.820: INFO: Skipping dumping logs from cluster
 {"msg":"Test Suite completed","total":280,"completed":280,"skipped":4563,"failed":0}
 
-Ran 280 of 4843 Specs in 5258.350 seconds
+Ran 280 of 4843 Specs in 4474.585 seconds
 SUCCESS! -- 280 Passed | 0 Failed | 0 Pending | 4563 Skipped
 PASS
 
-Ginkgo ran 1 suite in 1h27m39.963181709s
+Ginkgo ran 1 suite in 1h14m36.514576846s
 Test Suite Passed
diff --git a/v1.17/ibm-openshift/junit_01.xml b/v1.17/ibm-openshift/junit_01.xml
index e88a1d2967..c681c885d7 100644
--- a/v1.17/ibm-openshift/junit_01.xml
+++ b/v1.17/ibm-openshift/junit_01.xml
@@ -1,13972 +1,13972 @@
 
-  
-      
+  
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
   
\ No newline at end of file