From a9b506123c756b99d70b20444f68d95de1f14b35 Mon Sep 17 00:00:00 2001 From: Christoph Mewes Date: Sat, 13 Feb 2021 00:41:28 +0100 Subject: [PATCH] Conformance results for v1.17/kubermatic (#1321) Signed-off-by: Christoph Mewes --- v1.17/kubermatic/PRODUCT.yaml | 2 +- v1.17/kubermatic/e2e.log | 22096 ++++++++++++++++---------------- v1.17/kubermatic/junit_01.xml | 9693 +++++++------- 3 files changed, 15935 insertions(+), 15856 deletions(-) diff --git a/v1.17/kubermatic/PRODUCT.yaml b/v1.17/kubermatic/PRODUCT.yaml index 34f51ed27c..f9770ce187 100644 --- a/v1.17/kubermatic/PRODUCT.yaml +++ b/v1.17/kubermatic/PRODUCT.yaml @@ -1,7 +1,7 @@ vendor: Kubermatic name: Kubermatic Kubernetes Platform description: 'Kubermatic Kubernetes Platform is in an open source project to centrally manage the global automation of thousands of Kubernetes clusters across multicloud, on-prem and edge.' -version: v2.14 +version: v2.16 website_url: https://kubermatic.com documentation_url: https://docs.kubermatic.com/kubermatic/ repo_url: https://github.com/kubermatic/kubermatic diff --git a/v1.17/kubermatic/e2e.log b/v1.17/kubermatic/e2e.log index 215f70f613..95fb43aca3 100644 --- a/v1.17/kubermatic/e2e.log +++ b/v1.17/kubermatic/e2e.log @@ -1,8926 +1,8488 @@ -I0830 16:29:41.536470 23 test_context.go:406] Using a temporary kubeconfig file from in-cluster config : /tmp/kubeconfig-962893509 -I0830 16:29:41.536489 23 test_context.go:419] Tolerating taints "node-role.kubernetes.io/master" when considering if nodes are ready -I0830 16:29:41.536584 23 e2e.go:109] Starting e2e run "eda2017a-4ba1-4abb-bd48-e853704860b3" on Ginkgo node 1 +I0212 09:49:09.205147 21 test_context.go:406] Using a temporary kubeconfig file from in-cluster config : /tmp/kubeconfig-008079477 +I0212 09:49:09.205167 21 test_context.go:419] Tolerating taints "node-role.kubernetes.io/master" when considering if nodes are ready +I0212 09:49:09.205282 21 e2e.go:109] Starting e2e run "290da14e-be04-4851-88fa-116a06471991" on Ginkgo node 1 {"msg":"Test Suite starting","total":280,"completed":0,"skipped":0,"failed":0} Running Suite: Kubernetes e2e suite =================================== -Random Seed: 1598804980 - Will randomize all specs -Will run 280 of 4843 specs - -Aug 30 16:29:41.544: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -Aug 30 16:29:41.546: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable -Aug 30 16:29:41.573: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready -Aug 30 16:29:41.609: INFO: 13 / 13 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) -Aug 30 16:29:41.609: INFO: expected 3 pod replicas in namespace 'kube-system', 3 are Running and Ready. -Aug 30 16:29:41.610: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start -Aug 30 16:29:41.620: INFO: 2 / 2 pods ready in namespace 'kube-system' in daemonset 'canal' (0 seconds elapsed) -Aug 30 16:29:41.620: INFO: 2 / 2 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) -Aug 30 16:29:41.620: INFO: 2 / 2 pods ready in namespace 'kube-system' in daemonset 'logrotate' (0 seconds elapsed) -Aug 30 16:29:41.620: INFO: 2 / 2 pods ready in namespace 'kube-system' in daemonset 'node-local-dns' (0 seconds elapsed) -Aug 30 16:29:41.620: INFO: 2 / 2 pods ready in namespace 'kube-system' in daemonset 'user-ssh-keys-agent' (0 seconds elapsed) -Aug 30 16:29:41.620: INFO: e2e test version: v1.17.9 -Aug 30 16:29:41.623: INFO: kube-apiserver version: v1.17.9 -Aug 30 16:29:41.623: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -Aug 30 16:29:41.632: INFO: Cluster IP family: ipv4 -SSSSSSSSSSS +Random Seed: 1613123348 - Will randomize all specs +Will run 280 of 4846 specs + +Feb 12 09:49:09.228: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +Feb 12 09:49:09.230: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable +E0212 09:49:09.230707 21 progress.go:119] Failed to post progress update to http://localhost:8099/progress: Post http://localhost:8099/progress: dial tcp 127.0.0.1:8099: connect: connection refused +Feb 12 09:49:09.269: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready +Feb 12 09:49:09.316: INFO: 18 / 18 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) +Feb 12 09:49:09.316: INFO: expected 3 pod replicas in namespace 'kube-system', 3 are Running and Ready. +Feb 12 09:49:09.316: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start +Feb 12 09:49:09.328: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'canal' (0 seconds elapsed) +Feb 12 09:49:09.328: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) +Feb 12 09:49:09.328: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'logrotate' (0 seconds elapsed) +Feb 12 09:49:09.328: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'node-local-dns' (0 seconds elapsed) +Feb 12 09:49:09.328: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'user-ssh-keys-agent' (0 seconds elapsed) +Feb 12 09:49:09.328: INFO: e2e test version: v1.17.16 +Feb 12 09:49:09.335: INFO: kube-apiserver version: v1.17.16 +Feb 12 09:49:09.335: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +Feb 12 09:49:09.342: INFO: Cluster IP family: ipv4 +SSSSSSSSSSSS ------------------------------ -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should be able to deny pod and configmap creation [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +[sig-api-machinery] Garbage collector + should delete pods created by rc when not orphaning [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-api-machinery] Garbage collector + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:29:41.632: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename webhook -Aug 30 16:29:41.694: INFO: Found PodSecurityPolicies; assuming PodSecurityPolicy is enabled. -Aug 30 16:29:41.716: INFO: Found ClusterRoles; assuming RBAC is enabled. -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-2721 +Feb 12 09:49:09.343: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename gc +Feb 12 09:49:09.405: INFO: Found PodSecurityPolicies; assuming PodSecurityPolicy is enabled. +Feb 12 09:49:09.426: INFO: Found ClusterRoles; assuming RBAC is enabled. +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-8364 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 -STEP: Setting up server cert -STEP: Create role binding to let webhook read extension-apiserver-authentication -STEP: Deploying the webhook pod -STEP: Wait for the deployment to be ready -Aug 30 16:29:42.097: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -Aug 30 16:29:44.113: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734401782, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734401782, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734401782, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734401782, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)} -Aug 30 16:29:46.119: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734401782, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734401782, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734401782, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734401782, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)} -STEP: Deploying the webhook service -STEP: Verifying the service has paired with the endpoint -Aug 30 16:29:49.133: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should be able to deny pod and configmap creation [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Registering the webhook via the AdmissionRegistration API -STEP: create a pod that should be denied by the webhook -STEP: create a pod that causes the webhook to hang -STEP: create a configmap that should be denied by the webhook -STEP: create a configmap that should be admitted by the webhook -STEP: update (PUT) the admitted configmap to a non-compliant one should be rejected by the webhook -STEP: update (PATCH) the admitted configmap to a non-compliant one should be rejected by the webhook -STEP: create a namespace that bypass the webhook -STEP: create a configmap that violates the webhook policy but is in a whitelisted namespace -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:30:00.086: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-2721" for this suite. -STEP: Destroying namespace "webhook-2721-markers" for this suite. -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 +[It] should delete pods created by rc when not orphaning [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: create the rc +STEP: delete the rc +STEP: wait for all pods to be garbage collected +STEP: Gathering metrics +Feb 12 09:49:19.628: INFO: For apiserver_request_total: +For apiserver_request_latency_seconds: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: -• [SLOW TEST:18.548 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - should be able to deny pod and configmap creation [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance]","total":280,"completed":1,"skipped":11,"failed":0} -SSSSSSS +W0212 09:49:19.628426 21 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. +[AfterEach] [sig-api-machinery] Garbage collector + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:49:19.628: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-8364" for this suite. + +• [SLOW TEST:10.301 seconds] +[sig-api-machinery] Garbage collector +/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 + should delete pods created by rc when not orphaning [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - patching/updating a mutating webhook should work [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Aug 30 16:30:00.182: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename webhook -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-667 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 -STEP: Setting up server cert -STEP: Create role binding to let webhook read extension-apiserver-authentication -STEP: Deploying the webhook pod -STEP: Wait for the deployment to be ready -Aug 30 16:30:00.770: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -STEP: Deploying the webhook service -STEP: Verifying the service has paired with the endpoint -Aug 30 16:30:03.798: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] patching/updating a mutating webhook should work [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a mutating webhook configuration -STEP: Updating a mutating webhook configuration's rules to not include the create operation -STEP: Creating a configMap that should not be mutated -STEP: Patching a mutating webhook configuration's rules to include the create operation -STEP: Creating a configMap that should be mutated -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:30:04.101: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-667" for this suite. -STEP: Destroying namespace "webhook-667-markers" for this suite. -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 -•{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance]","total":280,"completed":2,"skipped":18,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance]","total":280,"completed":1,"skipped":12,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Servers with support for Table transformation - should return a 406 for a backend which does not implement metadata [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] Servers with support for Table transformation - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +[sig-storage] Projected configMap + should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-storage] Projected configMap + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:30:04.210: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename tables -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in tables-8053 +Feb 12 09:49:19.644: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8446 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-api-machinery] Servers with support for Table transformation - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/table_conversion.go:46 -[It] should return a 406 for a backend which does not implement metadata [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[AfterEach] [sig-api-machinery] Servers with support for Table transformation - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:30:04.377: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "tables-8053" for this suite. -•{"msg":"PASSED [sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance]","total":280,"completed":3,"skipped":42,"failed":0} -SSS +[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating configMap with name projected-configmap-test-volume-map-552f0a1e-09f5-451f-a0e1-67868c216ab0 +STEP: Creating a pod to test consume configMaps +Feb 12 09:49:19.821: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-f13ad958-4324-487e-a87e-9db22202d678" in namespace "projected-8446" to be "success or failure" +Feb 12 09:49:19.828: INFO: Pod "pod-projected-configmaps-f13ad958-4324-487e-a87e-9db22202d678": Phase="Pending", Reason="", readiness=false. Elapsed: 6.150425ms +Feb 12 09:49:21.841: INFO: Pod "pod-projected-configmaps-f13ad958-4324-487e-a87e-9db22202d678": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01954719s +Feb 12 09:49:23.882: INFO: Pod "pod-projected-configmaps-f13ad958-4324-487e-a87e-9db22202d678": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.06036365s +STEP: Saw pod success +Feb 12 09:49:23.882: INFO: Pod "pod-projected-configmaps-f13ad958-4324-487e-a87e-9db22202d678" satisfied condition "success or failure" +Feb 12 09:49:23.890: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-projected-configmaps-f13ad958-4324-487e-a87e-9db22202d678 container projected-configmap-volume-test: +STEP: delete the pod +Feb 12 09:49:23.933: INFO: Waiting for pod pod-projected-configmaps-f13ad958-4324-487e-a87e-9db22202d678 to disappear +Feb 12 09:49:23.948: INFO: Pod pod-projected-configmaps-f13ad958-4324-487e-a87e-9db22202d678 no longer exists +[AfterEach] [sig-storage] Projected configMap + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:49:23.948: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-8446" for this suite. +•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","total":280,"completed":2,"skipped":35,"failed":0} +SSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] Deployment - RecreateDeployment should delete old pods and create new ones [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +[sig-storage] HostPath + should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-storage] HostPath + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:30:04.402: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename deployment -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-5519 +Feb 12 09:49:23.970: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename hostpath +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in hostpath-4218 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [sig-apps] Deployment - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69 -[It] RecreateDeployment should delete old pods and create new ones [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Aug 30 16:30:04.612: INFO: Creating deployment "test-recreate-deployment" -Aug 30 16:30:04.619: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 -Aug 30 16:30:04.639: INFO: deployment "test-recreate-deployment" doesn't have the required revision set -Aug 30 16:30:06.687: INFO: Waiting deployment "test-recreate-deployment" to complete -Aug 30 16:30:06.709: INFO: Triggering a new rollout for deployment "test-recreate-deployment" -Aug 30 16:30:06.721: INFO: Updating deployment test-recreate-deployment -Aug 30 16:30:06.721: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63 -Aug 30 16:30:07.022: INFO: Deployment "test-recreate-deployment": -&Deployment{ObjectMeta:{test-recreate-deployment deployment-5519 /apis/apps/v1/namespaces/deployment-5519/deployments/test-recreate-deployment 8dbf4b47-805e-4a54-b8f3-82fbc3010540 2100 2 2020-08-30 16:30:04 +0000 UTC map[name:sample-pod-3] map[deployment.kubernetes.io/revision:2] [] [] []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc002826d08 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} [] nil default-scheduler [] [] nil [] map[] []}},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2020-08-30 16:30:06 +0000 UTC,LastTransitionTime:2020-08-30 16:30:06 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "test-recreate-deployment-5f94c574ff" is progressing.,LastUpdateTime:2020-08-30 16:30:06 +0000 UTC,LastTransitionTime:2020-08-30 16:30:04 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},} - -Aug 30 16:30:07.031: INFO: New ReplicaSet "test-recreate-deployment-5f94c574ff" of Deployment "test-recreate-deployment": -&ReplicaSet{ObjectMeta:{test-recreate-deployment-5f94c574ff deployment-5519 /apis/apps/v1/namespaces/deployment-5519/replicasets/test-recreate-deployment-5f94c574ff 2c30a2a2-d8b5-467f-b9a3-1746b909c66a 2099 1 2020-08-30 16:30:06 +0000 UTC map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-recreate-deployment 8dbf4b47-805e-4a54-b8f3-82fbc3010540 0xc002827337 0xc002827338}] [] []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 5f94c574ff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[] [] [] []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc002827398 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} -Aug 30 16:30:07.031: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": -Aug 30 16:30:07.031: INFO: &ReplicaSet{ObjectMeta:{test-recreate-deployment-799c574856 deployment-5519 /apis/apps/v1/namespaces/deployment-5519/replicasets/test-recreate-deployment-799c574856 dcb78500-58b3-4fc5-9748-4b422e8bb25c 2087 2 2020-08-30 16:30:04 +0000 UTC map[name:sample-pod-3 pod-template-hash:799c574856] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-recreate-deployment 8dbf4b47-805e-4a54-b8f3-82fbc3010540 0xc0028273f7 0xc0028273f8}] [] []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 799c574856,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:799c574856] map[] [] [] []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc002827468 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} [] nil default-scheduler [] [] nil [] map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} -Aug 30 16:30:07.057: INFO: Pod "test-recreate-deployment-5f94c574ff-ldg5l" is not available: -&Pod{ObjectMeta:{test-recreate-deployment-5f94c574ff-ldg5l test-recreate-deployment-5f94c574ff- deployment-5519 /api/v1/namespaces/deployment-5519/pods/test-recreate-deployment-5f94c574ff-ldg5l 12cd748b-0b2e-4706-8ef5-327df40a3a4d 2098 0 2020-08-30 16:30:06 +0000 UTC map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[] [{apps/v1 ReplicaSet test-recreate-deployment-5f94c574ff 2c30a2a2-d8b5-467f-b9a3-1746b909c66a 0xc0028278c7 0xc0028278c8}] [] []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-df8nz,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-df8nz,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-df8nz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:30:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:30:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:30:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:30:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 16:30:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -[AfterEach] [sig-apps] Deployment - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:30:07.057: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "deployment-5519" for this suite. -•{"msg":"PASSED [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance]","total":280,"completed":4,"skipped":45,"failed":0} -SS +[BeforeEach] [sig-storage] HostPath + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37 +[It] should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating a pod to test hostPath mode +Feb 12 09:49:24.174: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "hostpath-4218" to be "success or failure" +Feb 12 09:49:24.181: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 7.377847ms +Feb 12 09:49:26.188: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014331152s +Feb 12 09:49:28.194: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020673321s +STEP: Saw pod success +Feb 12 09:49:28.194: INFO: Pod "pod-host-path-test" satisfied condition "success or failure" +Feb 12 09:49:28.201: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-host-path-test container test-container-1: +STEP: delete the pod +Feb 12 09:49:28.234: INFO: Waiting for pod pod-host-path-test to disappear +Feb 12 09:49:28.241: INFO: Pod pod-host-path-test no longer exists +[AfterEach] [sig-storage] HostPath + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:49:28.241: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "hostpath-4218" for this suite. +•{"msg":"PASSED [sig-storage] HostPath should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":3,"skipped":51,"failed":0} +SSSSSSSSSS ------------------------------ -[sig-node] ConfigMap - should be consumable via environment variable [NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-node] ConfigMap - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +[sig-scheduling] SchedulerPredicates [Serial] + validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:30:07.079: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename configmap -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-7845 +Feb 12 09:49:28.265: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename sched-pred +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-2461 STEP: Waiting for a default service account to be provisioned in namespace -[It] should be consumable via environment variable [NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating configMap configmap-7845/configmap-test-665aab27-47b5-482c-81e3-f6ea3fb39ce2 -STEP: Creating a pod to test consume configMaps -Aug 30 16:30:07.309: INFO: Waiting up to 5m0s for pod "pod-configmaps-3d7dd31f-c195-4eee-93f4-a4395d094f3d" in namespace "configmap-7845" to be "success or failure" -Aug 30 16:30:07.316: INFO: Pod "pod-configmaps-3d7dd31f-c195-4eee-93f4-a4395d094f3d": Phase="Pending", Reason="", readiness=false. Elapsed: 7.456934ms -Aug 30 16:30:09.329: INFO: Pod "pod-configmaps-3d7dd31f-c195-4eee-93f4-a4395d094f3d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019920308s -Aug 30 16:30:11.338: INFO: Pod "pod-configmaps-3d7dd31f-c195-4eee-93f4-a4395d094f3d": Phase="Pending", Reason="", readiness=false. Elapsed: 4.029670313s -Aug 30 16:30:13.344: INFO: Pod "pod-configmaps-3d7dd31f-c195-4eee-93f4-a4395d094f3d": Phase="Pending", Reason="", readiness=false. Elapsed: 6.034850169s -Aug 30 16:30:15.351: INFO: Pod "pod-configmaps-3d7dd31f-c195-4eee-93f4-a4395d094f3d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.041734266s -STEP: Saw pod success -Aug 30 16:30:15.351: INFO: Pod "pod-configmaps-3d7dd31f-c195-4eee-93f4-a4395d094f3d" satisfied condition "success or failure" -Aug 30 16:30:15.355: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-configmaps-3d7dd31f-c195-4eee-93f4-a4395d094f3d container env-test: -STEP: delete the pod -Aug 30 16:30:15.429: INFO: Waiting for pod pod-configmaps-3d7dd31f-c195-4eee-93f4-a4395d094f3d to disappear -Aug 30 16:30:15.433: INFO: Pod pod-configmaps-3d7dd31f-c195-4eee-93f4-a4395d094f3d no longer exists -[AfterEach] [sig-node] ConfigMap - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:30:15.433: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-7845" for this suite. +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86 +Feb 12 09:49:28.440: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready +Feb 12 09:49:28.458: INFO: Waiting for terminating namespaces to be deleted... +Feb 12 09:49:28.467: INFO: +Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-dbhjj before test +Feb 12 09:49:28.495: INFO: kube-proxy-c88c9 from kube-system started at 2021-02-12 09:48:16 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.496: INFO: Container kube-proxy ready: true, restart count 0 +Feb 12 09:49:28.496: INFO: canal-k48gh from kube-system started at 2021-02-12 09:48:16 +0000 UTC (2 container statuses recorded) +Feb 12 09:49:28.496: INFO: Container calico-node ready: true, restart count 0 +Feb 12 09:49:28.496: INFO: Container kube-flannel ready: true, restart count 0 +Feb 12 09:49:28.496: INFO: node-local-dns-k2khb from kube-system started at 2021-02-12 09:48:36 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.496: INFO: Container node-cache ready: true, restart count 0 +Feb 12 09:49:28.496: INFO: logrotate-4f9dq from kube-system started at 2021-02-12 09:48:36 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.496: INFO: Container logrotate ready: true, restart count 0 +Feb 12 09:49:28.496: INFO: sonobuoy-e2e-job-690c418daa0f4403 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded) +Feb 12 09:49:28.496: INFO: Container e2e ready: true, restart count 0 +Feb 12 09:49:28.497: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 12 09:49:28.497: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-vkql5 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded) +Feb 12 09:49:28.497: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 12 09:49:28.497: INFO: Container systemd-logs ready: true, restart count 0 +Feb 12 09:49:28.497: INFO: user-ssh-keys-agent-8lrpd from kube-system started at 2021-02-12 09:48:16 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.497: INFO: Container user-ssh-keys-agent ready: true, restart count 0 +Feb 12 09:49:28.497: INFO: +Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-r6snm before test +Feb 12 09:49:28.586: INFO: canal-b9q6f from kube-system started at 2021-02-12 09:48:02 +0000 UTC (2 container statuses recorded) +Feb 12 09:49:28.586: INFO: Container calico-node ready: true, restart count 0 +Feb 12 09:49:28.586: INFO: Container kube-flannel ready: true, restart count 0 +Feb 12 09:49:28.586: INFO: sonobuoy from sonobuoy started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.586: INFO: Container kube-sonobuoy ready: true, restart count 0 +Feb 12 09:49:28.586: INFO: coredns-6449d494cc-psw4s from kube-system started at 2021-02-12 09:48:21 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.586: INFO: Container coredns ready: true, restart count 0 +Feb 12 09:49:28.586: INFO: kube-proxy-gfdtg from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.586: INFO: Container kube-proxy ready: true, restart count 0 +Feb 12 09:49:28.586: INFO: user-ssh-keys-agent-vxhth from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.587: INFO: Container user-ssh-keys-agent ready: true, restart count 0 +Feb 12 09:49:28.587: INFO: dashboard-metrics-scraper-59bfc65dc9-rhjf7 from kubernetes-dashboard started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.587: INFO: Container dashboard-metrics-scraper ready: true, restart count 0 +Feb 12 09:49:28.587: INFO: dashboard-metrics-scraper-59bfc65dc9-d8l4b from kubernetes-dashboard started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.587: INFO: Container dashboard-metrics-scraper ready: true, restart count 0 +Feb 12 09:49:28.587: INFO: node-local-dns-fkrhg from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.587: INFO: Container node-cache ready: true, restart count 0 +Feb 12 09:49:28.587: INFO: logrotate-7vwrm from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.587: INFO: Container logrotate ready: true, restart count 0 +Feb 12 09:49:28.587: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-zp8g7 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded) +Feb 12 09:49:28.587: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 12 09:49:28.588: INFO: Container systemd-logs ready: true, restart count 0 +Feb 12 09:49:28.588: INFO: +Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-xg74t before test +Feb 12 09:49:28.621: INFO: logrotate-m2dmm from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.621: INFO: Container logrotate ready: true, restart count 0 +Feb 12 09:49:28.621: INFO: node-local-dns-mxsff from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.621: INFO: Container node-cache ready: true, restart count 0 +Feb 12 09:49:28.621: INFO: coredns-6449d494cc-plbcg from kube-system started at 2021-02-12 09:48:21 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.621: INFO: Container coredns ready: true, restart count 0 +Feb 12 09:49:28.621: INFO: openvpn-client-78d595f58b-xgrgj from kube-system started at 2021-02-12 09:48:21 +0000 UTC (2 container statuses recorded) +Feb 12 09:49:28.621: INFO: Container dnat-controller ready: true, restart count 0 +Feb 12 09:49:28.621: INFO: Container openvpn-client ready: true, restart count 0 +Feb 12 09:49:28.621: INFO: user-ssh-keys-agent-kc6zq from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.622: INFO: Container user-ssh-keys-agent ready: true, restart count 0 +Feb 12 09:49:28.622: INFO: canal-ww2cz from kube-system started at 2021-02-12 09:48:02 +0000 UTC (2 container statuses recorded) +Feb 12 09:49:28.622: INFO: Container calico-node ready: true, restart count 0 +Feb 12 09:49:28.622: INFO: Container kube-flannel ready: true, restart count 0 +Feb 12 09:49:28.622: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-wbbpd from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded) +Feb 12 09:49:28.622: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 12 09:49:28.622: INFO: Container systemd-logs ready: true, restart count 0 +Feb 12 09:49:28.622: INFO: kube-proxy-4zfp6 from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded) +Feb 12 09:49:28.622: INFO: Container kube-proxy ready: true, restart count 0 +[It] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Trying to launch a pod without a label to get a node which can launch it. +STEP: Explicitly delete pod here to free the resource it takes. +STEP: Trying to apply a random label on the found node. +STEP: verifying the node has the label kubernetes.io/e2e-c9000567-f4d3-4700-91c9-66508a40661a 95 +STEP: Trying to create a pod(pod4) with hostport 54322 and hostIP 0.0.0.0(empty string here) and expect scheduled +STEP: Trying to create another pod(pod5) with hostport 54322 but hostIP 127.0.0.1 on the node which pod4 resides and expect not scheduled +STEP: removing the label kubernetes.io/e2e-c9000567-f4d3-4700-91c9-66508a40661a off the node nostalgic-fermat-cdc5d8777-xg74t +STEP: verifying the node doesn't have the label kubernetes.io/e2e-c9000567-f4d3-4700-91c9-66508a40661a +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:54:32.837: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-pred-2461" for this suite. +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77 -• [SLOW TEST:8.368 seconds] -[sig-node] ConfigMap -/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:31 - should be consumable via environment variable [NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +• [SLOW TEST:304.599 seconds] +[sig-scheduling] SchedulerPredicates [Serial] +/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40 + validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance]","total":280,"completed":5,"skipped":47,"failed":0} -SSSSSSSSSSSS +{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]","total":280,"completed":4,"skipped":61,"failed":0} +SSSSS ------------------------------ -[sig-cli] Kubectl client Kubectl logs - should be able to retrieve and filter logs [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[sig-cli] Kubectl client Kubectl api-versions + should check if v1 is in available api versions [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 [BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:30:15.447: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 +Feb 12 09:54:32.866: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 STEP: Building a namespace api object, basename kubectl -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3537 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-885 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-cli] Kubectl client - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272 -[BeforeEach] Kubectl logs - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1357 -STEP: creating an pod -Aug 30 16:30:15.617: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 run logs-generator --generator=run-pod/v1 --image=gcr.io/kubernetes-e2e-test-images/agnhost:2.8 --namespace=kubectl-3537 -- logs-generator --log-lines-total 100 --run-duration 20s' -Aug 30 16:30:15.825: INFO: stderr: "" -Aug 30 16:30:15.825: INFO: stdout: "pod/logs-generator created\n" -[It] should be able to retrieve and filter logs [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Waiting for log generator to start. -Aug 30 16:30:15.825: INFO: Waiting up to 5m0s for 1 pods to be running and ready, or succeeded: [logs-generator] -Aug 30 16:30:15.825: INFO: Waiting up to 5m0s for pod "logs-generator" in namespace "kubectl-3537" to be "running and ready, or succeeded" -Aug 30 16:30:15.835: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 10.107724ms -Aug 30 16:30:17.841: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015708335s -Aug 30 16:30:19.846: INFO: Pod "logs-generator": Phase="Running", Reason="", readiness=true. Elapsed: 4.020860013s -Aug 30 16:30:19.846: INFO: Pod "logs-generator" satisfied condition "running and ready, or succeeded" -Aug 30 16:30:19.846: INFO: Wanted all 1 pods to be running and ready, or succeeded. Result: true. Pods: [logs-generator] -STEP: checking for a matching strings -Aug 30 16:30:19.846: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 logs logs-generator logs-generator --namespace=kubectl-3537' -Aug 30 16:30:19.942: INFO: stderr: "" -Aug 30 16:30:19.942: INFO: stdout: "I0830 16:30:17.059896 1 logs_generator.go:76] 0 PUT /api/v1/namespaces/kube-system/pods/j4g6 334\nI0830 16:30:17.260067 1 logs_generator.go:76] 1 PUT /api/v1/namespaces/kube-system/pods/rhsr 202\nI0830 16:30:17.460060 1 logs_generator.go:76] 2 PUT /api/v1/namespaces/ns/pods/lhrc 456\nI0830 16:30:17.660104 1 logs_generator.go:76] 3 GET /api/v1/namespaces/default/pods/pw8z 405\nI0830 16:30:17.860045 1 logs_generator.go:76] 4 POST /api/v1/namespaces/kube-system/pods/kkgn 394\nI0830 16:30:18.060078 1 logs_generator.go:76] 5 PUT /api/v1/namespaces/kube-system/pods/g97 308\nI0830 16:30:18.260067 1 logs_generator.go:76] 6 PUT /api/v1/namespaces/kube-system/pods/x2j 488\nI0830 16:30:18.460012 1 logs_generator.go:76] 7 PUT /api/v1/namespaces/default/pods/qlcn 545\nI0830 16:30:18.660076 1 logs_generator.go:76] 8 GET /api/v1/namespaces/default/pods/2wxn 209\nI0830 16:30:18.860075 1 logs_generator.go:76] 9 GET /api/v1/namespaces/ns/pods/plg 432\nI0830 16:30:19.060078 1 logs_generator.go:76] 10 POST /api/v1/namespaces/default/pods/ftjf 276\nI0830 16:30:19.260176 1 logs_generator.go:76] 11 PUT /api/v1/namespaces/ns/pods/lzh4 499\nI0830 16:30:19.460198 1 logs_generator.go:76] 12 PUT /api/v1/namespaces/kube-system/pods/qnmh 257\nI0830 16:30:19.660149 1 logs_generator.go:76] 13 POST /api/v1/namespaces/ns/pods/hvbp 503\nI0830 16:30:19.860090 1 logs_generator.go:76] 14 GET /api/v1/namespaces/ns/pods/pstk 435\n" -STEP: limiting log lines -Aug 30 16:30:19.942: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 logs logs-generator logs-generator --namespace=kubectl-3537 --tail=1' -Aug 30 16:30:20.157: INFO: stderr: "" -Aug 30 16:30:20.157: INFO: stdout: "I0830 16:30:20.060091 1 logs_generator.go:76] 15 PUT /api/v1/namespaces/default/pods/gf4 349\n" -Aug 30 16:30:20.157: INFO: got output "I0830 16:30:20.060091 1 logs_generator.go:76] 15 PUT /api/v1/namespaces/default/pods/gf4 349\n" -STEP: limiting log bytes -Aug 30 16:30:20.157: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 logs logs-generator logs-generator --namespace=kubectl-3537 --limit-bytes=1' -Aug 30 16:30:20.241: INFO: stderr: "" -Aug 30 16:30:20.241: INFO: stdout: "I" -Aug 30 16:30:20.241: INFO: got output "I" -STEP: exposing timestamps -Aug 30 16:30:20.241: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 logs logs-generator logs-generator --namespace=kubectl-3537 --tail=1 --timestamps' -Aug 30 16:30:20.333: INFO: stderr: "" -Aug 30 16:30:20.333: INFO: stdout: "2020-08-30T16:30:20.260116997Z I0830 16:30:20.260000 1 logs_generator.go:76] 16 GET /api/v1/namespaces/default/pods/kfg6 210\n" -Aug 30 16:30:20.333: INFO: got output "2020-08-30T16:30:20.260116997Z I0830 16:30:20.260000 1 logs_generator.go:76] 16 GET /api/v1/namespaces/default/pods/kfg6 210\n" -STEP: restricting to a time range -Aug 30 16:30:22.835: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 logs logs-generator logs-generator --namespace=kubectl-3537 --since=1s' -Aug 30 16:30:22.919: INFO: stderr: "" -Aug 30 16:30:22.919: INFO: stdout: "I0830 16:30:22.060133 1 logs_generator.go:76] 25 POST /api/v1/namespaces/default/pods/d7k4 517\nI0830 16:30:22.259997 1 logs_generator.go:76] 26 GET /api/v1/namespaces/default/pods/msm 375\nI0830 16:30:22.460144 1 logs_generator.go:76] 27 GET /api/v1/namespaces/default/pods/v8zq 384\nI0830 16:30:22.660081 1 logs_generator.go:76] 28 PUT /api/v1/namespaces/ns/pods/9n6 331\nI0830 16:30:22.860156 1 logs_generator.go:76] 29 GET /api/v1/namespaces/kube-system/pods/g6g 284\n" -Aug 30 16:30:22.919: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 logs logs-generator logs-generator --namespace=kubectl-3537 --since=24h' -Aug 30 16:30:23.003: INFO: stderr: "" -Aug 30 16:30:23.003: INFO: stdout: "I0830 16:30:17.059896 1 logs_generator.go:76] 0 PUT /api/v1/namespaces/kube-system/pods/j4g6 334\nI0830 16:30:17.260067 1 logs_generator.go:76] 1 PUT /api/v1/namespaces/kube-system/pods/rhsr 202\nI0830 16:30:17.460060 1 logs_generator.go:76] 2 PUT /api/v1/namespaces/ns/pods/lhrc 456\nI0830 16:30:17.660104 1 logs_generator.go:76] 3 GET /api/v1/namespaces/default/pods/pw8z 405\nI0830 16:30:17.860045 1 logs_generator.go:76] 4 POST /api/v1/namespaces/kube-system/pods/kkgn 394\nI0830 16:30:18.060078 1 logs_generator.go:76] 5 PUT /api/v1/namespaces/kube-system/pods/g97 308\nI0830 16:30:18.260067 1 logs_generator.go:76] 6 PUT /api/v1/namespaces/kube-system/pods/x2j 488\nI0830 16:30:18.460012 1 logs_generator.go:76] 7 PUT /api/v1/namespaces/default/pods/qlcn 545\nI0830 16:30:18.660076 1 logs_generator.go:76] 8 GET /api/v1/namespaces/default/pods/2wxn 209\nI0830 16:30:18.860075 1 logs_generator.go:76] 9 GET /api/v1/namespaces/ns/pods/plg 432\nI0830 16:30:19.060078 1 logs_generator.go:76] 10 POST /api/v1/namespaces/default/pods/ftjf 276\nI0830 16:30:19.260176 1 logs_generator.go:76] 11 PUT /api/v1/namespaces/ns/pods/lzh4 499\nI0830 16:30:19.460198 1 logs_generator.go:76] 12 PUT /api/v1/namespaces/kube-system/pods/qnmh 257\nI0830 16:30:19.660149 1 logs_generator.go:76] 13 POST /api/v1/namespaces/ns/pods/hvbp 503\nI0830 16:30:19.860090 1 logs_generator.go:76] 14 GET /api/v1/namespaces/ns/pods/pstk 435\nI0830 16:30:20.060091 1 logs_generator.go:76] 15 PUT /api/v1/namespaces/default/pods/gf4 349\nI0830 16:30:20.260000 1 logs_generator.go:76] 16 GET /api/v1/namespaces/default/pods/kfg6 210\nI0830 16:30:20.460022 1 logs_generator.go:76] 17 PUT /api/v1/namespaces/kube-system/pods/t7h 363\nI0830 16:30:20.660114 1 logs_generator.go:76] 18 GET /api/v1/namespaces/default/pods/c6l 526\nI0830 16:30:20.860084 1 logs_generator.go:76] 19 GET /api/v1/namespaces/ns/pods/7sd 506\nI0830 16:30:21.060022 1 logs_generator.go:76] 20 GET /api/v1/namespaces/ns/pods/qqt 453\nI0830 16:30:21.260008 1 logs_generator.go:76] 21 GET /api/v1/namespaces/kube-system/pods/5rqp 293\nI0830 16:30:21.460017 1 logs_generator.go:76] 22 PUT /api/v1/namespaces/ns/pods/pvzn 361\nI0830 16:30:21.660070 1 logs_generator.go:76] 23 PUT /api/v1/namespaces/kube-system/pods/ckpg 491\nI0830 16:30:21.860075 1 logs_generator.go:76] 24 POST /api/v1/namespaces/ns/pods/rkf 366\nI0830 16:30:22.060133 1 logs_generator.go:76] 25 POST /api/v1/namespaces/default/pods/d7k4 517\nI0830 16:30:22.259997 1 logs_generator.go:76] 26 GET /api/v1/namespaces/default/pods/msm 375\nI0830 16:30:22.460144 1 logs_generator.go:76] 27 GET /api/v1/namespaces/default/pods/v8zq 384\nI0830 16:30:22.660081 1 logs_generator.go:76] 28 PUT /api/v1/namespaces/ns/pods/9n6 331\nI0830 16:30:22.860156 1 logs_generator.go:76] 29 GET /api/v1/namespaces/kube-system/pods/g6g 284\n" -[AfterEach] Kubectl logs - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1363 -Aug 30 16:30:23.004: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete pod logs-generator --namespace=kubectl-3537' -Aug 30 16:30:25.036: INFO: stderr: "" -Aug 30 16:30:25.036: INFO: stdout: "pod \"logs-generator\" deleted\n" + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273 +[It] should check if v1 is in available api versions [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: validating api versions +Feb 12 09:54:33.029: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 api-versions' +Feb 12 09:54:33.106: INFO: stderr: "" +Feb 12 09:54:33.106: INFO: stdout: "admissionregistration.k8s.io/v1\nadmissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps/v1\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\ncertificates.k8s.io/v1beta1\ncluster.k8s.io/v1alpha1\ncoordination.k8s.io/v1\ncoordination.k8s.io/v1beta1\ncrd.projectcalico.org/v1\ndiscovery.k8s.io/v1beta1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nmetrics.k8s.io/v1beta1\nnetworking.k8s.io/v1\nnetworking.k8s.io/v1beta1\nnode.k8s.io/v1beta1\npolicy/v1beta1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1\nscheduling.k8s.io/v1beta1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n" [AfterEach] [sig-cli] Kubectl client - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:30:25.036: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-3537" for this suite. - -• [SLOW TEST:9.605 seconds] -[sig-cli] Kubectl client -/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23 - Kubectl logs - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1353 - should be able to retrieve and filter logs [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:54:33.106: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-885" for this suite. +•{"msg":"PASSED [sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions [Conformance]","total":280,"completed":5,"skipped":66,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -{"msg":"PASSED [sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs [Conformance]","total":280,"completed":6,"skipped":59,"failed":0} -[sig-api-machinery] ResourceQuota - should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] ResourceQuota - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +[sig-auth] ServiceAccounts + should allow opting out of API token automount [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-auth] ServiceAccounts + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:30:25.052: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename resourcequota -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-4080 +Feb 12 09:54:33.124: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename svcaccounts +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svcaccounts-1468 STEP: Waiting for a default service account to be provisioned in namespace -[It] should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Counting existing ResourceQuota -STEP: Creating a ResourceQuota -STEP: Ensuring resource quota status is calculated -[AfterEach] [sig-api-machinery] ResourceQuota - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:30:32.242: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-4080" for this suite. - -• [SLOW TEST:7.202 seconds] -[sig-api-machinery] ResourceQuota -/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[It] should allow opting out of API token automount [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: getting the auto-created API token +Feb 12 09:54:33.826: INFO: created pod pod-service-account-defaultsa +Feb 12 09:54:33.826: INFO: pod pod-service-account-defaultsa service account token volume mount: true +Feb 12 09:54:33.837: INFO: created pod pod-service-account-mountsa +Feb 12 09:54:33.837: INFO: pod pod-service-account-mountsa service account token volume mount: true +Feb 12 09:54:33.848: INFO: created pod pod-service-account-nomountsa +Feb 12 09:54:33.848: INFO: pod pod-service-account-nomountsa service account token volume mount: false +Feb 12 09:54:33.865: INFO: created pod pod-service-account-defaultsa-mountspec +Feb 12 09:54:33.865: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true +Feb 12 09:54:33.876: INFO: created pod pod-service-account-mountsa-mountspec +Feb 12 09:54:33.876: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true +Feb 12 09:54:33.890: INFO: created pod pod-service-account-nomountsa-mountspec +Feb 12 09:54:33.890: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true +Feb 12 09:54:33.917: INFO: created pod pod-service-account-defaultsa-nomountspec +Feb 12 09:54:33.917: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false +Feb 12 09:54:33.929: INFO: created pod pod-service-account-mountsa-nomountspec +Feb 12 09:54:33.929: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false +Feb 12 09:54:33.945: INFO: created pod pod-service-account-nomountsa-nomountspec +Feb 12 09:54:33.945: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false +[AfterEach] [sig-auth] ServiceAccounts + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:54:33.945: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svcaccounts-1468" for this suite. +•{"msg":"PASSED [sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance]","total":280,"completed":6,"skipped":91,"failed":0} +SSSSSSS ------------------------------ -{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]","total":280,"completed":7,"skipped":59,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +[sig-api-machinery] Watchers + should be able to start watching from a specific resource version [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-api-machinery] Watchers + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Feb 12 09:54:34.043: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename watch +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-8946 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be able to start watching from a specific resource version [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: creating a new configmap +STEP: modifying the configmap once +STEP: modifying the configmap a second time +STEP: deleting the configmap +STEP: creating a watch on configmaps from the resource version returned by the first update +STEP: Expecting to observe notifications for all changes to the configmap after the first update +Feb 12 09:54:34.309: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-8946 /api/v1/namespaces/watch-8946/configmaps/e2e-watch-test-resource-version 65292306-b9fa-4aab-b53e-c2af2d5b0daa 3819 0 2021-02-12 09:54:34 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +Feb 12 09:54:34.309: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-8946 /api/v1/namespaces/watch-8946/configmaps/e2e-watch-test-resource-version 65292306-b9fa-4aab-b53e-c2af2d5b0daa 3820 0 2021-02-12 09:54:34 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},} +[AfterEach] [sig-api-machinery] Watchers + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:54:34.309: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "watch-8946" for this suite. +•{"msg":"PASSED [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance]","total":280,"completed":7,"skipped":98,"failed":0} +SSSSS ------------------------------ -[k8s.io] InitContainer [NodeConformance] - should invoke init containers on a RestartNever pod [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +[k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class + should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [k8s.io] [sig-node] Pods Extended + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:30:32.254: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename init-container -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-2085 +Feb 12 09:54:34.338: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename pods +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-6397 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153 -[It] should invoke init containers on a RestartNever pod [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [k8s.io] Pods Set QOS Class + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:185 +[It] should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 STEP: creating the pod -Aug 30 16:30:32.430: INFO: PodSpec: initContainers in spec.initContainers -[AfterEach] [k8s.io] InitContainer [NodeConformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:30:36.420: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "init-container-2085" for this suite. -•{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance]","total":280,"completed":8,"skipped":114,"failed":0} -SSSSSS +STEP: submitting the pod to kubernetes +STEP: verifying QOS class is set on the pod +[AfterEach] [k8s.io] [sig-node] Pods Extended + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:54:34.540: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-6397" for this suite. +•{"msg":"PASSED [k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]","total":280,"completed":8,"skipped":103,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] ResourceQuota - should create a ResourceQuota and capture the life of a service. [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-api-machinery] ResourceQuota - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +[sig-storage] Projected secret + should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-storage] Projected secret + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:30:36.437: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename resourcequota -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-7369 +Feb 12 09:54:34.573: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename projected +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3853 STEP: Waiting for a default service account to be provisioned in namespace -[It] should create a ResourceQuota and capture the life of a service. [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Counting existing ResourceQuota -STEP: Creating a ResourceQuota -STEP: Ensuring resource quota status is calculated -STEP: Creating a Service -STEP: Ensuring resource quota status captures service creation -STEP: Deleting a Service -STEP: Ensuring resource quota status released usage -[AfterEach] [sig-api-machinery] ResourceQuota - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:30:47.705: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-7369" for this suite. +[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating secret with name projected-secret-test-e2df1cc1-c40e-4da4-a3af-19df4f5b2fa3 +STEP: Creating a pod to test consume secrets +Feb 12 09:54:34.790: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-03a6a1f1-41fe-4a85-985f-36553e4a1ff3" in namespace "projected-3853" to be "success or failure" +Feb 12 09:54:34.794: INFO: Pod "pod-projected-secrets-03a6a1f1-41fe-4a85-985f-36553e4a1ff3": Phase="Pending", Reason="", readiness=false. Elapsed: 4.666765ms +Feb 12 09:54:36.800: INFO: Pod "pod-projected-secrets-03a6a1f1-41fe-4a85-985f-36553e4a1ff3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010597847s +Feb 12 09:54:38.806: INFO: Pod "pod-projected-secrets-03a6a1f1-41fe-4a85-985f-36553e4a1ff3": Phase="Pending", Reason="", readiness=false. Elapsed: 4.016553091s +Feb 12 09:54:40.812: INFO: Pod "pod-projected-secrets-03a6a1f1-41fe-4a85-985f-36553e4a1ff3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.022586334s +STEP: Saw pod success +Feb 12 09:54:40.812: INFO: Pod "pod-projected-secrets-03a6a1f1-41fe-4a85-985f-36553e4a1ff3" satisfied condition "success or failure" +Feb 12 09:54:40.818: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-dbhjj pod pod-projected-secrets-03a6a1f1-41fe-4a85-985f-36553e4a1ff3 container secret-volume-test: +STEP: delete the pod +Feb 12 09:54:40.888: INFO: Waiting for pod pod-projected-secrets-03a6a1f1-41fe-4a85-985f-36553e4a1ff3 to disappear +Feb 12 09:54:40.893: INFO: Pod pod-projected-secrets-03a6a1f1-41fe-4a85-985f-36553e4a1ff3 no longer exists +[AfterEach] [sig-storage] Projected secret + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:54:40.893: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-3853" for this suite. -• [SLOW TEST:11.283 seconds] -[sig-api-machinery] ResourceQuota -/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 - should create a ResourceQuota and capture the life of a service. [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +• [SLOW TEST:6.335 seconds] +[sig-storage] Projected secret +/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_secret.go:34 + should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance]","total":280,"completed":9,"skipped":120,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":280,"completed":9,"skipped":127,"failed":0} +SSSSSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir wrapper volumes - should not conflict [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] EmptyDir wrapper volumes - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Aug 30 16:30:47.723: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename emptydir-wrapper -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-wrapper-2046 +[sig-api-machinery] Namespaces [Serial] + should ensure that all services are removed when a namespace is deleted [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-api-machinery] Namespaces [Serial] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Feb 12 09:54:40.908: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename namespaces +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in namespaces-1980 STEP: Waiting for a default service account to be provisioned in namespace -[It] should not conflict [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Cleaning up the secret -STEP: Cleaning up the configmap -STEP: Cleaning up the pod -[AfterEach] [sig-storage] EmptyDir wrapper volumes - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:30:51.970: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-wrapper-2046" for this suite. -•{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance]","total":280,"completed":10,"skipped":163,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSSSSSSS +[It] should ensure that all services are removed when a namespace is deleted [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating a test namespace +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-4465 +STEP: Waiting for a default service account to be provisioned in namespace +STEP: Creating a service in the namespace +STEP: Deleting the namespace +STEP: Waiting for the namespace to be removed. +STEP: Recreating the namespace +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-6696 +STEP: Verifying there is no service in the namespace +[AfterEach] [sig-api-machinery] Namespaces [Serial] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:54:47.449: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "namespaces-1980" for this suite. +STEP: Destroying namespace "nsdeletetest-4465" for this suite. +Feb 12 09:54:47.484: INFO: Namespace nsdeletetest-4465 was already deleted +STEP: Destroying namespace "nsdeletetest-6696" for this suite. + +• [SLOW TEST:6.587 seconds] +[sig-api-machinery] Namespaces [Serial] +/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 + should ensure that all services are removed when a namespace is deleted [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -[sig-storage] Projected combined - should project all components that make up the projection API [Projection][NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [sig-storage] Projected combined - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance]","total":280,"completed":10,"skipped":140,"failed":0} +SSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-scheduling] SchedulerPredicates [Serial] + validates that NodeSelector is respected if not matching [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:30:51.987: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename projected -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-4887 +Feb 12 09:54:47.496: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename sched-pred +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-3883 STEP: Waiting for a default service account to be provisioned in namespace -[It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating configMap with name configmap-projected-all-test-volume-db12fb4d-5722-41ae-82cb-9d52ea8f8579 -STEP: Creating secret with name secret-projected-all-test-volume-b5a5de62-7b54-4f91-b3e1-ac2da66c0996 -STEP: Creating a pod to test Check all projections for projected volume plugin -Aug 30 16:30:52.199: INFO: Waiting up to 5m0s for pod "projected-volume-7226a450-af9a-4914-b623-0e7157923755" in namespace "projected-4887" to be "success or failure" -Aug 30 16:30:52.205: INFO: Pod "projected-volume-7226a450-af9a-4914-b623-0e7157923755": Phase="Pending", Reason="", readiness=false. Elapsed: 5.386609ms -Aug 30 16:30:54.210: INFO: Pod "projected-volume-7226a450-af9a-4914-b623-0e7157923755": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010488812s +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86 +Feb 12 09:54:47.661: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready +Feb 12 09:54:47.680: INFO: Waiting for terminating namespaces to be deleted... +Feb 12 09:54:47.685: INFO: +Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-dbhjj before test +Feb 12 09:54:47.702: INFO: node-local-dns-k2khb from kube-system started at 2021-02-12 09:48:36 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.702: INFO: Container node-cache ready: true, restart count 0 +Feb 12 09:54:47.702: INFO: logrotate-4f9dq from kube-system started at 2021-02-12 09:48:36 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.703: INFO: Container logrotate ready: true, restart count 0 +Feb 12 09:54:47.703: INFO: user-ssh-keys-agent-8lrpd from kube-system started at 2021-02-12 09:48:16 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.703: INFO: Container user-ssh-keys-agent ready: true, restart count 0 +Feb 12 09:54:47.703: INFO: kube-proxy-c88c9 from kube-system started at 2021-02-12 09:48:16 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.703: INFO: Container kube-proxy ready: true, restart count 0 +Feb 12 09:54:47.703: INFO: canal-k48gh from kube-system started at 2021-02-12 09:48:16 +0000 UTC (2 container statuses recorded) +Feb 12 09:54:47.703: INFO: Container calico-node ready: true, restart count 0 +Feb 12 09:54:47.703: INFO: Container kube-flannel ready: true, restart count 0 +Feb 12 09:54:47.703: INFO: sonobuoy-e2e-job-690c418daa0f4403 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded) +Feb 12 09:54:47.703: INFO: Container e2e ready: true, restart count 0 +Feb 12 09:54:47.703: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 12 09:54:47.703: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-vkql5 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded) +Feb 12 09:54:47.703: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 12 09:54:47.703: INFO: Container systemd-logs ready: true, restart count 0 +Feb 12 09:54:47.703: INFO: +Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-r6snm before test +Feb 12 09:54:47.789: INFO: user-ssh-keys-agent-vxhth from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.789: INFO: Container user-ssh-keys-agent ready: true, restart count 0 +Feb 12 09:54:47.789: INFO: dashboard-metrics-scraper-59bfc65dc9-rhjf7 from kubernetes-dashboard started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.789: INFO: Container dashboard-metrics-scraper ready: true, restart count 0 +Feb 12 09:54:47.789: INFO: dashboard-metrics-scraper-59bfc65dc9-d8l4b from kubernetes-dashboard started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.789: INFO: Container dashboard-metrics-scraper ready: true, restart count 0 +Feb 12 09:54:47.789: INFO: node-local-dns-fkrhg from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.789: INFO: Container node-cache ready: true, restart count 0 +Feb 12 09:54:47.789: INFO: logrotate-7vwrm from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.789: INFO: Container logrotate ready: true, restart count 0 +Feb 12 09:54:47.790: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-zp8g7 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded) +Feb 12 09:54:47.790: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 12 09:54:47.790: INFO: Container systemd-logs ready: true, restart count 0 +Feb 12 09:54:47.790: INFO: kube-proxy-gfdtg from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.790: INFO: Container kube-proxy ready: true, restart count 0 +Feb 12 09:54:47.790: INFO: sonobuoy from sonobuoy started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.790: INFO: Container kube-sonobuoy ready: true, restart count 0 +Feb 12 09:54:47.790: INFO: coredns-6449d494cc-psw4s from kube-system started at 2021-02-12 09:48:21 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.790: INFO: Container coredns ready: true, restart count 0 +Feb 12 09:54:47.790: INFO: pod-qos-class-8a814817-efcb-4c4c-b648-efc101abf3fb from pods-6397 started at 2021-02-12 09:54:34 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.790: INFO: Container nginx ready: true, restart count 0 +Feb 12 09:54:47.790: INFO: canal-b9q6f from kube-system started at 2021-02-12 09:48:02 +0000 UTC (2 container statuses recorded) +Feb 12 09:54:47.790: INFO: Container calico-node ready: true, restart count 0 +Feb 12 09:54:47.790: INFO: Container kube-flannel ready: true, restart count 0 +Feb 12 09:54:47.790: INFO: +Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-xg74t before test +Feb 12 09:54:47.861: INFO: node-local-dns-mxsff from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.861: INFO: Container node-cache ready: true, restart count 0 +Feb 12 09:54:47.861: INFO: coredns-6449d494cc-plbcg from kube-system started at 2021-02-12 09:48:21 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.861: INFO: Container coredns ready: true, restart count 0 +Feb 12 09:54:47.861: INFO: openvpn-client-78d595f58b-xgrgj from kube-system started at 2021-02-12 09:48:21 +0000 UTC (2 container statuses recorded) +Feb 12 09:54:47.861: INFO: Container dnat-controller ready: true, restart count 0 +Feb 12 09:54:47.861: INFO: Container openvpn-client ready: true, restart count 0 +Feb 12 09:54:47.861: INFO: user-ssh-keys-agent-kc6zq from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.861: INFO: Container user-ssh-keys-agent ready: true, restart count 0 +Feb 12 09:54:47.862: INFO: logrotate-m2dmm from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.862: INFO: Container logrotate ready: true, restart count 0 +Feb 12 09:54:47.862: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-wbbpd from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded) +Feb 12 09:54:47.862: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 12 09:54:47.862: INFO: Container systemd-logs ready: true, restart count 0 +Feb 12 09:54:47.862: INFO: kube-proxy-4zfp6 from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded) +Feb 12 09:54:47.862: INFO: Container kube-proxy ready: true, restart count 0 +Feb 12 09:54:47.862: INFO: canal-ww2cz from kube-system started at 2021-02-12 09:48:02 +0000 UTC (2 container statuses recorded) +Feb 12 09:54:47.862: INFO: Container calico-node ready: true, restart count 0 +Feb 12 09:54:47.862: INFO: Container kube-flannel ready: true, restart count 0 +[It] validates that NodeSelector is respected if not matching [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Trying to schedule Pod with nonempty NodeSelector. +STEP: Considering event: +Type = [Warning], Name = [restricted-pod.1662f7728475058f], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match node selector.] +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:54:48.900: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-pred-3883" for this suite. +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77 +•{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching [Conformance]","total":280,"completed":11,"skipped":159,"failed":0} +SSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] Secrets + should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-storage] Secrets + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +STEP: Creating a kubernetes client +Feb 12 09:54:48.916: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename secrets +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-5269 +STEP: Waiting for a default service account to be provisioned in namespace +[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating secret with name secret-test-f71a2c04-39cc-4745-adb0-8e59b9d78c04 +STEP: Creating a pod to test consume secrets +Feb 12 09:54:49.101: INFO: Waiting up to 5m0s for pod "pod-secrets-026c74e4-799a-477c-8e54-5c66627eee34" in namespace "secrets-5269" to be "success or failure" +Feb 12 09:54:49.106: INFO: Pod "pod-secrets-026c74e4-799a-477c-8e54-5c66627eee34": Phase="Pending", Reason="", readiness=false. Elapsed: 4.442359ms +Feb 12 09:54:51.112: INFO: Pod "pod-secrets-026c74e4-799a-477c-8e54-5c66627eee34": Phase="Running", Reason="", readiness=true. Elapsed: 2.011059296s +Feb 12 09:54:53.118: INFO: Pod "pod-secrets-026c74e4-799a-477c-8e54-5c66627eee34": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016971589s STEP: Saw pod success -Aug 30 16:30:54.210: INFO: Pod "projected-volume-7226a450-af9a-4914-b623-0e7157923755" satisfied condition "success or failure" -Aug 30 16:30:54.214: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod projected-volume-7226a450-af9a-4914-b623-0e7157923755 container projected-all-volume-test: +Feb 12 09:54:53.118: INFO: Pod "pod-secrets-026c74e4-799a-477c-8e54-5c66627eee34" satisfied condition "success or failure" +Feb 12 09:54:53.124: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-secrets-026c74e4-799a-477c-8e54-5c66627eee34 container secret-volume-test: STEP: delete the pod -Aug 30 16:30:54.280: INFO: Waiting for pod projected-volume-7226a450-af9a-4914-b623-0e7157923755 to disappear -Aug 30 16:30:54.285: INFO: Pod projected-volume-7226a450-af9a-4914-b623-0e7157923755 no longer exists -[AfterEach] [sig-storage] Projected combined - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:30:54.285: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-4887" for this suite. -•{"msg":"PASSED [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance]","total":280,"completed":11,"skipped":191,"failed":0} -SSSSS +Feb 12 09:54:53.181: INFO: Waiting for pod pod-secrets-026c74e4-799a-477c-8e54-5c66627eee34 to disappear +Feb 12 09:54:53.186: INFO: Pod pod-secrets-026c74e4-799a-477c-8e54-5c66627eee34 no longer exists +[AfterEach] [sig-storage] Secrets + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:54:53.186: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-5269" for this suite. +•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":12,"skipped":177,"failed":0} +SS ------------------------------ -[k8s.io] Variable Expansion - should allow substituting values in a container's command [NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Variable Expansion - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +[sig-storage] Downward API volume + should provide container's memory request [NodeConformance] [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-storage] Downward API volume + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:30:54.300: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename var-expansion -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-6793 +Feb 12 09:54:53.200: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename downward-api +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-4777 STEP: Waiting for a default service account to be provisioned in namespace -[It] should allow substituting values in a container's command [NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating a pod to test substitution in container's command -Aug 30 16:30:54.494: INFO: Waiting up to 5m0s for pod "var-expansion-96457c9d-db75-4b2f-8c7a-d9550f80356f" in namespace "var-expansion-6793" to be "success or failure" -Aug 30 16:30:54.502: INFO: Pod "var-expansion-96457c9d-db75-4b2f-8c7a-d9550f80356f": Phase="Pending", Reason="", readiness=false. Elapsed: 8.243971ms -Aug 30 16:30:56.508: INFO: Pod "var-expansion-96457c9d-db75-4b2f-8c7a-d9550f80356f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014528319s +[BeforeEach] [sig-storage] Downward API volume + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40 +[It] should provide container's memory request [NodeConformance] [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +STEP: Creating a pod to test downward API volume plugin +Feb 12 09:54:53.392: INFO: Waiting up to 5m0s for pod "downwardapi-volume-06aabe5c-6a19-48d6-9aa9-4232dcb1b9a2" in namespace "downward-api-4777" to be "success or failure" +Feb 12 09:54:53.398: INFO: Pod "downwardapi-volume-06aabe5c-6a19-48d6-9aa9-4232dcb1b9a2": Phase="Pending", Reason="", readiness=false. Elapsed: 5.048771ms +Feb 12 09:54:55.405: INFO: Pod "downwardapi-volume-06aabe5c-6a19-48d6-9aa9-4232dcb1b9a2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011765476s +Feb 12 09:54:57.413: INFO: Pod "downwardapi-volume-06aabe5c-6a19-48d6-9aa9-4232dcb1b9a2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.019472141s STEP: Saw pod success -Aug 30 16:30:56.509: INFO: Pod "var-expansion-96457c9d-db75-4b2f-8c7a-d9550f80356f" satisfied condition "success or failure" -Aug 30 16:30:56.515: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod var-expansion-96457c9d-db75-4b2f-8c7a-d9550f80356f container dapi-container: +Feb 12 09:54:57.413: INFO: Pod "downwardapi-volume-06aabe5c-6a19-48d6-9aa9-4232dcb1b9a2" satisfied condition "success or failure" +Feb 12 09:54:57.419: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-06aabe5c-6a19-48d6-9aa9-4232dcb1b9a2 container client-container: STEP: delete the pod -Aug 30 16:30:56.589: INFO: Waiting for pod var-expansion-96457c9d-db75-4b2f-8c7a-d9550f80356f to disappear -Aug 30 16:30:56.595: INFO: Pod var-expansion-96457c9d-db75-4b2f-8c7a-d9550f80356f no longer exists -[AfterEach] [k8s.io] Variable Expansion - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:30:56.595: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-6793" for this suite. -•{"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance]","total":280,"completed":12,"skipped":196,"failed":0} -S +Feb 12 09:54:57.478: INFO: Waiting for pod downwardapi-volume-06aabe5c-6a19-48d6-9aa9-4232dcb1b9a2 to disappear +Feb 12 09:54:57.483: INFO: Pod downwardapi-volume-06aabe5c-6a19-48d6-9aa9-4232dcb1b9a2 no longer exists +[AfterEach] [sig-storage] Downward API volume + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:54:57.483: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-4777" for this suite. +•{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance]","total":280,"completed":13,"skipped":179,"failed":0} +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] - should run and stop complex daemon [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 + should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 [BeforeEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:30:56.609: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 +Feb 12 09:54:57.502: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 STEP: Building a namespace api object, basename daemonsets -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-7433 +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-4286 STEP: Waiting for a default service account to be provisioned in namespace [BeforeEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133 -[It] should run and stop complex daemon [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Aug 30 16:30:56.813: INFO: Creating daemon "daemon-set" with a node selector -STEP: Initially, daemon pods should not be running on any nodes. -Aug 30 16:30:56.827: INFO: Number of nodes with available pods: 0 -Aug 30 16:30:56.827: INFO: Number of running nodes: 0, number of available pods: 0 -STEP: Change node label to blue, check that daemon pod is launched. -Aug 30 16:30:56.862: INFO: Number of nodes with available pods: 0 -Aug 30 16:30:56.862: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:30:57.869: INFO: Number of nodes with available pods: 0 -Aug 30 16:30:57.869: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:30:58.868: INFO: Number of nodes with available pods: 0 -Aug 30 16:30:58.868: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:30:59.873: INFO: Number of nodes with available pods: 0 -Aug 30 16:30:59.873: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:31:00.868: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:00.868: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:31:01.867: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:01.867: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:31:02.868: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:02.868: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:31:03.868: INFO: Number of nodes with available pods: 1 -Aug 30 16:31:03.868: INFO: Number of running nodes: 1, number of available pods: 1 -STEP: Update the node label to green, and wait for daemons to be unscheduled -Aug 30 16:31:03.902: INFO: Number of nodes with available pods: 1 -Aug 30 16:31:03.902: INFO: Number of running nodes: 0, number of available pods: 1 -Aug 30 16:31:04.909: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:04.909: INFO: Number of running nodes: 0, number of available pods: 0 -STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate -Aug 30 16:31:04.923: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:04.924: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:31:05.930: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:05.930: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:31:06.929: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:06.929: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:31:07.930: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:07.930: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:31:08.930: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:08.931: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:31:09.930: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:09.930: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:31:10.931: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:10.931: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:31:11.930: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:11.930: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod -Aug 30 16:31:12.929: INFO: Number of nodes with available pods: 1 -Aug 30 16:31:12.929: INFO: Number of running nodes: 1, number of available pods: 1 + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133 +[It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +Feb 12 09:54:57.735: INFO: Creating simple daemon set daemon-set +STEP: Check that daemon pods launch on every node of the cluster. +Feb 12 09:54:57.769: INFO: Number of nodes with available pods: 0 +Feb 12 09:54:57.769: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod +Feb 12 09:54:58.787: INFO: Number of nodes with available pods: 0 +Feb 12 09:54:58.787: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod +Feb 12 09:54:59.781: INFO: Number of nodes with available pods: 0 +Feb 12 09:54:59.781: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod +Feb 12 09:55:00.783: INFO: Number of nodes with available pods: 0 +Feb 12 09:55:00.783: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod +Feb 12 09:55:01.782: INFO: Number of nodes with available pods: 0 +Feb 12 09:55:01.782: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod +Feb 12 09:55:02.784: INFO: Number of nodes with available pods: 0 +Feb 12 09:55:02.784: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod +Feb 12 09:55:03.781: INFO: Number of nodes with available pods: 0 +Feb 12 09:55:03.781: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod +Feb 12 09:55:04.783: INFO: Number of nodes with available pods: 0 +Feb 12 09:55:04.783: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod +Feb 12 09:55:05.786: INFO: Number of nodes with available pods: 3 +Feb 12 09:55:05.786: INFO: Number of running nodes: 3, number of available pods: 3 +STEP: Update daemon pods image. +STEP: Check that daemon pods images are updated. +Feb 12 09:55:05.831: INFO: Wrong image for pod: daemon-set-j4hrv. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:05.831: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:05.831: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:06.863: INFO: Wrong image for pod: daemon-set-j4hrv. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:06.863: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:06.863: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:07.857: INFO: Wrong image for pod: daemon-set-j4hrv. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:07.857: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:07.857: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:08.864: INFO: Wrong image for pod: daemon-set-j4hrv. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:08.864: INFO: Pod daemon-set-j4hrv is not available +Feb 12 09:55:08.864: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:08.864: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:09.858: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:09.858: INFO: Pod daemon-set-vztm4 is not available +Feb 12 09:55:09.858: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:10.861: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:10.861: INFO: Pod daemon-set-vztm4 is not available +Feb 12 09:55:10.861: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:11.857: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:11.858: INFO: Pod daemon-set-vztm4 is not available +Feb 12 09:55:11.858: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:12.860: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:12.861: INFO: Pod daemon-set-vztm4 is not available +Feb 12 09:55:12.861: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:13.859: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:13.859: INFO: Pod daemon-set-vztm4 is not available +Feb 12 09:55:13.859: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:14.865: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:14.865: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:15.864: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:15.864: INFO: Pod daemon-set-prmd5 is not available +Feb 12 09:55:15.864: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:16.860: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:16.860: INFO: Pod daemon-set-prmd5 is not available +Feb 12 09:55:16.860: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:17.856: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:17.856: INFO: Pod daemon-set-prmd5 is not available +Feb 12 09:55:17.856: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:18.859: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:18.859: INFO: Pod daemon-set-prmd5 is not available +Feb 12 09:55:18.860: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:19.857: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:19.857: INFO: Pod daemon-set-prmd5 is not available +Feb 12 09:55:19.857: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:20.857: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:20.857: INFO: Pod daemon-set-prmd5 is not available +Feb 12 09:55:20.857: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:21.857: INFO: Wrong image for pod: daemon-set-prmd5. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:21.857: INFO: Pod daemon-set-prmd5 is not available +Feb 12 09:55:21.857: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:22.872: INFO: Pod daemon-set-7gc2b is not available +Feb 12 09:55:22.872: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:23.868: INFO: Pod daemon-set-7gc2b is not available +Feb 12 09:55:23.868: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:24.858: INFO: Pod daemon-set-7gc2b is not available +Feb 12 09:55:24.858: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:25.860: INFO: Pod daemon-set-7gc2b is not available +Feb 12 09:55:25.860: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:26.867: INFO: Pod daemon-set-7gc2b is not available +Feb 12 09:55:26.867: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:27.857: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:28.866: INFO: Wrong image for pod: daemon-set-xrbs4. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine. +Feb 12 09:55:28.866: INFO: Pod daemon-set-xrbs4 is not available +Feb 12 09:55:29.858: INFO: Pod daemon-set-jbfmg is not available +STEP: Check that daemon pods are still running on every node of the cluster. +Feb 12 09:55:29.883: INFO: Number of nodes with available pods: 2 +Feb 12 09:55:29.883: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod +Feb 12 09:55:30.900: INFO: Number of nodes with available pods: 2 +Feb 12 09:55:30.901: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod +Feb 12 09:55:31.897: INFO: Number of nodes with available pods: 2 +Feb 12 09:55:31.897: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod +Feb 12 09:55:32.898: INFO: Number of nodes with available pods: 2 +Feb 12 09:55:32.898: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod +Feb 12 09:55:33.897: INFO: Number of nodes with available pods: 3 +Feb 12 09:55:33.897: INFO: Number of running nodes: 3, number of available pods: 3 [AfterEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99 + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99 STEP: Deleting DaemonSet "daemon-set" -STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-7433, will wait for the garbage collector to delete the pods -Aug 30 16:31:13.007: INFO: Deleting DaemonSet.extensions daemon-set took: 14.360742ms -Aug 30 16:31:13.507: INFO: Terminating DaemonSet.extensions daemon-set pods took: 500.179115ms -Aug 30 16:31:16.113: INFO: Number of nodes with available pods: 0 -Aug 30 16:31:16.113: INFO: Number of running nodes: 0, number of available pods: 0 -Aug 30 16:31:16.119: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-7433/daemonsets","resourceVersion":"2742"},"items":null} +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-4286, will wait for the garbage collector to delete the pods +Feb 12 09:55:34.005: INFO: Deleting DaemonSet.extensions daemon-set took: 13.052615ms +Feb 12 09:55:34.505: INFO: Terminating DaemonSet.extensions daemon-set pods took: 500.246055ms +Feb 12 09:55:42.411: INFO: Number of nodes with available pods: 0 +Feb 12 09:55:42.411: INFO: Number of running nodes: 0, number of available pods: 0 +Feb 12 09:55:42.416: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-4286/daemonsets","resourceVersion":"4583"},"items":null} -Aug 30 16:31:16.125: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-7433/pods","resourceVersion":"2742"},"items":null} +Feb 12 09:55:42.421: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-4286/pods","resourceVersion":"4583"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:31:16.152: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "daemonsets-7433" for this suite. + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:55:42.470: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-4286" for this suite. -• [SLOW TEST:19.557 seconds] +• [SLOW TEST:44.983 seconds] [sig-apps] Daemon set [Serial] -/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 - should run and stop complex daemon [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------- -{"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance]","total":280,"completed":13,"skipped":197,"failed":0} -SSSSSSSSSSSSSSSS +/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23 + should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -[k8s.io] Probing container - should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Probing container - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 -STEP: Creating a kubernetes client -Aug 30 16:31:16.166: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename container-probe -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-1373 -STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Probing container - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51 -[It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: Creating pod test-webserver-9bf8d6d9-21dd-487a-ae9a-7eca3f01ded2 in namespace container-probe-1373 -Aug 30 16:31:20.362: INFO: Started pod test-webserver-9bf8d6d9-21dd-487a-ae9a-7eca3f01ded2 in namespace container-probe-1373 -STEP: checking the pod's current state and verifying that restartCount is present -Aug 30 16:31:20.367: INFO: Initial restart count of pod test-webserver-9bf8d6d9-21dd-487a-ae9a-7eca3f01ded2 is 0 -STEP: deleting the pod -[AfterEach] [k8s.io] Probing container - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:35:22.143: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-1373" for this suite. - -• [SLOW TEST:245.992 seconds] -[k8s.io] Probing container -/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +{"msg":"PASSED [sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]","total":280,"completed":14,"skipped":230,"failed":0} +SSSSSSSSSSSSSS ------------------------------ -{"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":280,"completed":14,"skipped":213,"failed":0} -[k8s.io] Pods - should be updated [NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + should mutate custom resource [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:35:22.165: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename pods -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-9381 +Feb 12 09:55:42.487: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename webhook +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-1283 STEP: Waiting for a default service account to be provisioned in namespace -[BeforeEach] [k8s.io] Pods - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177 -[It] should be updated [NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -STEP: creating the pod -STEP: submitting the pod to kubernetes -STEP: verifying the pod is in kubernetes -STEP: updating the pod -Aug 30 16:35:28.880: INFO: Successfully updated pod "pod-update-21b6066a-33bf-407b-86b1-7ebe3b5e5df0" -STEP: verifying the updated pod is in kubernetes -Aug 30 16:35:28.888: INFO: Pod update OK -[AfterEach] [k8s.io] Pods - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 -Aug 30 16:35:28.888: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-9381" for this suite. +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87 +STEP: Setting up server cert +STEP: Create role binding to let webhook read extension-apiserver-authentication +STEP: Deploying the webhook pod +STEP: Wait for the deployment to be ready +Feb 12 09:55:43.118: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +Feb 12 09:55:45.140: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720543, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720543, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720543, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720542, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service +STEP: Verifying the service has paired with the endpoint +Feb 12 09:55:48.161: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should mutate custom resource [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +Feb 12 09:55:48.168: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Registering the mutating webhook for custom resource e2e-test-webhook-6516-crds.webhook.example.com via the AdmissionRegistration API +STEP: Creating a custom resource that should be mutated by the webhook +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152 +Feb 12 09:55:49.725: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-1283" for this suite. +STEP: Destroying namespace "webhook-1283-markers" for this suite. +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102 -• [SLOW TEST:6.737 seconds] -[k8s.io] Pods -/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716 - should be updated [NodeConformance] [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +• [SLOW TEST:7.342 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23 + should mutate custom resource [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 ------------------------------ -{"msg":"PASSED [k8s.io] Pods should be updated [NodeConformance] [Conformance]","total":280,"completed":15,"skipped":213,"failed":0} -SSSSSSSSSSSSSSSSSSSSSSS +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance]","total":280,"completed":15,"skipped":244,"failed":0} +SSS ------------------------------ -[sig-network] Proxy version v1 - should proxy logs on node using proxy subresource [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -[BeforeEach] version v1 - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 +[k8s.io] Security Context When creating a container with runAsUser + should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 +[BeforeEach] [k8s.io] Security Context + /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151 STEP: Creating a kubernetes client -Aug 30 16:35:28.902: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509 -STEP: Building a namespace api object, basename proxy -STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-2480 +Feb 12 09:55:49.832: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477 +STEP: Building a namespace api object, basename security-context-test +STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in security-context-test-2792 STEP: Waiting for a default service account to be provisioned in namespace -[It] should proxy logs on node using proxy subresource [Conformance] - /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721 -Aug 30 16:35:29.114: INFO: (0) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/:
-alternatives.log
-apt/
-... (200; 47.011998ms)
-Aug 30 16:35:29.162: INFO: (1) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 47.607247ms)
-Aug 30 16:35:29.209: INFO: (2) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 47.436353ms)
-Aug 30 16:35:29.218: INFO: (3) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 9.281031ms)
-Aug 30 16:35:29.251: INFO: (4) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 32.713884ms)
-Aug 30 16:35:29.323: INFO: (5) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 71.603236ms)
-Aug 30 16:35:29.333: INFO: (6) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 9.81989ms)
-Aug 30 16:35:29.341: INFO: (7) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 8.36592ms)
-Aug 30 16:35:29.351: INFO: (8) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 10.066191ms)
-Aug 30 16:35:29.361: INFO: (9) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 9.186453ms)
-Aug 30 16:35:29.442: INFO: (10) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 81.678026ms)
-Aug 30 16:35:29.457: INFO: (11) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 15.113422ms)
-Aug 30 16:35:29.507: INFO: (12) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 49.204417ms)
-Aug 30 16:35:29.515: INFO: (13) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 8.32563ms)
-Aug 30 16:35:29.562: INFO: (14) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 46.991215ms)
-Aug 30 16:35:29.571: INFO: (15) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 9.204257ms)
-Aug 30 16:35:29.619: INFO: (16) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 47.767621ms)
-Aug 30 16:35:29.627: INFO: (17) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 8.286886ms)
-Aug 30 16:35:29.640: INFO: (18) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 12.918541ms)
-Aug 30 16:35:29.654: INFO: (19) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 13.179856ms)
-[AfterEach] version v1
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:35:29.654: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "proxy-2480" for this suite.
-•{"msg":"PASSED [sig-network] Proxy version v1 should proxy logs on node using proxy subresource  [Conformance]","total":280,"completed":16,"skipped":236,"failed":0}
-SSSSSS
+[BeforeEach] [k8s.io] Security Context
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39
+[It] should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 09:55:50.027: INFO: Waiting up to 5m0s for pod "busybox-user-65534-789b0fd3-5c59-4e48-b655-94ab1ed0f364" in namespace "security-context-test-2792" to be "success or failure"
+Feb 12 09:55:50.032: INFO: Pod "busybox-user-65534-789b0fd3-5c59-4e48-b655-94ab1ed0f364": Phase="Pending", Reason="", readiness=false. Elapsed: 4.856805ms
+Feb 12 09:55:52.044: INFO: Pod "busybox-user-65534-789b0fd3-5c59-4e48-b655-94ab1ed0f364": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016856323s
+Feb 12 09:55:54.070: INFO: Pod "busybox-user-65534-789b0fd3-5c59-4e48-b655-94ab1ed0f364": Phase="Pending", Reason="", readiness=false. Elapsed: 4.043358077s
+Feb 12 09:55:56.077: INFO: Pod "busybox-user-65534-789b0fd3-5c59-4e48-b655-94ab1ed0f364": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.049893543s
+Feb 12 09:55:56.077: INFO: Pod "busybox-user-65534-789b0fd3-5c59-4e48-b655-94ab1ed0f364" satisfied condition "success or failure"
+[AfterEach] [k8s.io] Security Context
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 09:55:56.077: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "security-context-test-2792" for this suite.
+
+• [SLOW TEST:6.261 seconds]
+[k8s.io] Security Context
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  When creating a container with runAsUser
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:43
+    should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [k8s.io] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":16,"skipped":247,"failed":0}
+SSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl run deployment 
-  should create a deployment from an image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  works for multiple CRDs of same group and version but different kinds [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:35:29.669: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2205
+Feb 12 09:55:56.093: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename crd-publish-openapi
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-3802
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[BeforeEach] Kubectl run deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1626
-[It] should create a deployment from an image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: running the image docker.io/library/httpd:2.4.38-alpine
-Aug 30 16:35:29.830: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 run e2e-test-httpd-deployment --image=docker.io/library/httpd:2.4.38-alpine --generator=deployment/apps.v1 --namespace=kubectl-2205'
-Aug 30 16:35:29.926: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-Aug 30 16:35:29.926: INFO: stdout: "deployment.apps/e2e-test-httpd-deployment created\n"
-STEP: verifying the deployment e2e-test-httpd-deployment was created
-STEP: verifying the pod controlled by deployment e2e-test-httpd-deployment was created
-[AfterEach] Kubectl run deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1631
-Aug 30 16:35:31.952: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete deployment e2e-test-httpd-deployment --namespace=kubectl-2205'
-Aug 30 16:35:32.026: INFO: stderr: ""
-Aug 30 16:35:32.026: INFO: stdout: "deployment.apps \"e2e-test-httpd-deployment\" deleted\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:35:32.026: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-2205" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run deployment should create a deployment from an image  [Conformance]","total":280,"completed":17,"skipped":242,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[It] works for multiple CRDs of same group and version but different kinds [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: CRs in the same group and version but different kinds (two CRDs) show up in OpenAPI documentation
+Feb 12 09:55:56.261: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 09:55:59.183: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 09:56:11.565: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-3802" for this suite.
+
+• [SLOW TEST:15.487 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  works for multiple CRDs of same group and version but different kinds [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:35:32.040: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-9275
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0644 on tmpfs
-Aug 30 16:35:32.223: INFO: Waiting up to 5m0s for pod "pod-96fa5c45-57f0-4d25-ae74-e26638eecfa2" in namespace "emptydir-9275" to be "success or failure"
-Aug 30 16:35:32.227: INFO: Pod "pod-96fa5c45-57f0-4d25-ae74-e26638eecfa2": Phase="Pending", Reason="", readiness=false. Elapsed: 4.614183ms
-Aug 30 16:35:34.232: INFO: Pod "pod-96fa5c45-57f0-4d25-ae74-e26638eecfa2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.009352529s
-Aug 30 16:35:36.238: INFO: Pod "pod-96fa5c45-57f0-4d25-ae74-e26638eecfa2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015030796s
-STEP: Saw pod success
-Aug 30 16:35:36.238: INFO: Pod "pod-96fa5c45-57f0-4d25-ae74-e26638eecfa2" satisfied condition "success or failure"
-Aug 30 16:35:36.242: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod pod-96fa5c45-57f0-4d25-ae74-e26638eecfa2 container test-container: 
-STEP: delete the pod
-Aug 30 16:35:36.315: INFO: Waiting for pod pod-96fa5c45-57f0-4d25-ae74-e26638eecfa2 to disappear
-Aug 30 16:35:36.319: INFO: Pod pod-96fa5c45-57f0-4d25-ae74-e26638eecfa2 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:35:36.319: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-9275" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":18,"skipped":301,"failed":0}
-SSSSSS
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance]","total":280,"completed":17,"skipped":266,"failed":0}
+SSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-node] ConfigMap 
+  should fail to create ConfigMap with empty key [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-node] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:35:36.332: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-2665
+Feb 12 09:56:11.581: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-2343
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0666 on node default medium
-Aug 30 16:35:36.530: INFO: Waiting up to 5m0s for pod "pod-ec5f4dfc-6bb3-40de-8689-f13268185958" in namespace "emptydir-2665" to be "success or failure"
-Aug 30 16:35:36.540: INFO: Pod "pod-ec5f4dfc-6bb3-40de-8689-f13268185958": Phase="Pending", Reason="", readiness=false. Elapsed: 9.284805ms
-Aug 30 16:35:38.545: INFO: Pod "pod-ec5f4dfc-6bb3-40de-8689-f13268185958": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014578115s
-STEP: Saw pod success
-Aug 30 16:35:38.545: INFO: Pod "pod-ec5f4dfc-6bb3-40de-8689-f13268185958" satisfied condition "success or failure"
-Aug 30 16:35:38.554: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod pod-ec5f4dfc-6bb3-40de-8689-f13268185958 container test-container: 
-STEP: delete the pod
-Aug 30 16:35:38.619: INFO: Waiting for pod pod-ec5f4dfc-6bb3-40de-8689-f13268185958 to disappear
-Aug 30 16:35:38.623: INFO: Pod pod-ec5f4dfc-6bb3-40de-8689-f13268185958 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:35:38.623: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-2665" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":19,"skipped":307,"failed":0}
-SSSSSSSSSSS
+[It] should fail to create ConfigMap with empty key [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap that has name configmap-test-emptyKey-88bec098-3b36-4e01-bf6d-d8dd1a2986f0
+[AfterEach] [sig-node] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 09:56:11.748: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-2343" for this suite.
+•{"msg":"PASSED [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance]","total":280,"completed":18,"skipped":275,"failed":0}
+SS
 ------------------------------
-[k8s.io] Docker Containers 
-  should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Docker Containers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] DNS 
+  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:35:38.637: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename containers
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-1208
+Feb 12 09:56:11.764: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename dns
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-7351
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test override arguments
-Aug 30 16:35:38.853: INFO: Waiting up to 5m0s for pod "client-containers-e1247519-d5fe-4cad-b30d-17c4a3d4be23" in namespace "containers-1208" to be "success or failure"
-Aug 30 16:35:38.861: INFO: Pod "client-containers-e1247519-d5fe-4cad-b30d-17c4a3d4be23": Phase="Pending", Reason="", readiness=false. Elapsed: 7.158436ms
-Aug 30 16:35:40.866: INFO: Pod "client-containers-e1247519-d5fe-4cad-b30d-17c4a3d4be23": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012555535s
-Aug 30 16:35:42.872: INFO: Pod "client-containers-e1247519-d5fe-4cad-b30d-17c4a3d4be23": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018374469s
-Aug 30 16:35:44.878: INFO: Pod "client-containers-e1247519-d5fe-4cad-b30d-17c4a3d4be23": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.024668885s
-STEP: Saw pod success
-Aug 30 16:35:44.878: INFO: Pod "client-containers-e1247519-d5fe-4cad-b30d-17c4a3d4be23" satisfied condition "success or failure"
-Aug 30 16:35:44.884: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod client-containers-e1247519-d5fe-4cad-b30d-17c4a3d4be23 container test-container: 
-STEP: delete the pod
-Aug 30 16:35:44.958: INFO: Waiting for pod client-containers-e1247519-d5fe-4cad-b30d-17c4a3d4be23 to disappear
-Aug 30 16:35:44.962: INFO: Pod client-containers-e1247519-d5fe-4cad-b30d-17c4a3d4be23 no longer exists
-[AfterEach] [k8s.io] Docker Containers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:35:44.962: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "containers-1208" for this suite.
+[It] should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-7351.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-7351.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-7351.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
 
-• [SLOW TEST:6.341 seconds]
-[k8s.io] Docker Containers
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-7351.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-7351.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-7351.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+
+STEP: creating a pod to probe /etc/hosts
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Feb 12 09:56:24.722: INFO: DNS probes using dns-7351/dns-test-3da9bd6a-2d86-4345-92d3-818fac4b5312 succeeded
+
+STEP: deleting the pod
+[AfterEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 09:56:24.772: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-7351" for this suite.
+
+• [SLOW TEST:13.027 seconds]
+[sig-network] DNS
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]","total":280,"completed":20,"skipped":318,"failed":0}
-SSSSSS
+{"msg":"PASSED [sig-network] DNS should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]","total":280,"completed":19,"skipped":277,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] Deployment 
-  RollingUpdateDeployment should delete old pods and create new ones [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] ConfigMap 
+  updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:35:44.977: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename deployment
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-7703
+Feb 12 09:56:24.796: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-5765
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
-[It] RollingUpdateDeployment should delete old pods and create new ones [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 16:35:45.139: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted)
-Aug 30 16:35:45.153: INFO: Pod name sample-pod: Found 0 pods out of 1
-Aug 30 16:35:50.159: INFO: Pod name sample-pod: Found 1 pods out of 1
-STEP: ensuring each pod is running
-Aug 30 16:35:50.159: INFO: Creating deployment "test-rolling-update-deployment"
-Aug 30 16:35:50.167: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has
-Aug 30 16:35:50.175: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created
-Aug 30 16:35:52.192: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected
-Aug 30 16:35:52.199: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:2, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402150, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402150, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402152, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402150, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-67cf4f6444\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Aug 30 16:35:54.204: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted)
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
-Aug 30 16:35:54.217: INFO: Deployment "test-rolling-update-deployment":
-&Deployment{ObjectMeta:{test-rolling-update-deployment  deployment-7703 /apis/apps/v1/namespaces/deployment-7703/deployments/test-rolling-update-deployment 0d319f55-eb95-4372-8547-f6a910ff43eb 4011 1 2020-08-30 16:35:50 +0000 UTC   map[name:sample-pod] map[deployment.kubernetes.io/revision:3546343826724305833] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00270c4c8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-08-30 16:35:50 +0000 UTC,LastTransitionTime:2020-08-30 16:35:50 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rolling-update-deployment-67cf4f6444" has successfully progressed.,LastUpdateTime:2020-08-30 16:35:52 +0000 UTC,LastTransitionTime:2020-08-30 16:35:50 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},}
-
-Aug 30 16:35:54.225: INFO: New ReplicaSet "test-rolling-update-deployment-67cf4f6444" of Deployment "test-rolling-update-deployment":
-&ReplicaSet{ObjectMeta:{test-rolling-update-deployment-67cf4f6444  deployment-7703 /apis/apps/v1/namespaces/deployment-7703/replicasets/test-rolling-update-deployment-67cf4f6444 484db303-d776-4217-a4cb-61fcd92b29df 4000 1 2020-08-30 16:35:50 +0000 UTC   map[name:sample-pod pod-template-hash:67cf4f6444] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305833] [{apps/v1 Deployment test-rolling-update-deployment 0d319f55-eb95-4372-8547-f6a910ff43eb 0xc00270c977 0xc00270c978}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 67cf4f6444,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod pod-template-hash:67cf4f6444] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00270c9e8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},}
-Aug 30 16:35:54.225: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment":
-Aug 30 16:35:54.225: INFO: &ReplicaSet{ObjectMeta:{test-rolling-update-controller  deployment-7703 /apis/apps/v1/namespaces/deployment-7703/replicasets/test-rolling-update-controller f3737a56-3430-4db2-8c46-ae6e09459458 4010 2 2020-08-30 16:35:45 +0000 UTC   map[name:sample-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305832] [{apps/v1 Deployment test-rolling-update-deployment 0d319f55-eb95-4372-8547-f6a910ff43eb 0xc00270c89f 0xc00270c8b0}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod pod:httpd] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc00270c918  ClusterFirst map[]     false false false  PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
-Aug 30 16:35:54.231: INFO: Pod "test-rolling-update-deployment-67cf4f6444-j2cph" is available:
-&Pod{ObjectMeta:{test-rolling-update-deployment-67cf4f6444-j2cph test-rolling-update-deployment-67cf4f6444- deployment-7703 /api/v1/namespaces/deployment-7703/pods/test-rolling-update-deployment-67cf4f6444-j2cph 1926da9e-826b-4f3a-a520-186fb324c8e8 3998 0 2020-08-30 16:35:50 +0000 UTC   map[name:sample-pod pod-template-hash:67cf4f6444] map[cni.projectcalico.org/podIP:172.25.0.22/32] [{apps/v1 ReplicaSet test-rolling-update-deployment-67cf4f6444 484db303-d776-4217-a4cb-61fcd92b29df 0xc00270ce57 0xc00270ce58}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-7hw7x,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-7hw7x,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-7hw7x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:35:50 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:35:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:35:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:35:50 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:172.25.0.22,StartTime:2020-08-30 16:35:50 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-08-30 16:35:51 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/agnhost@sha256:daf5332100521b1256d0e3c56d697a238eaec3af48897ed9167cbadd426773b5,ContainerID:docker://0023a7180c84950d710b154874ef8a1c76343ef074258908c16c6d5a41270990,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.0.22,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:35:54.231: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "deployment-7703" for this suite.
+[It] updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name configmap-test-upd-c27a51d9-523f-4551-bd8f-79319d1f4573
+STEP: Creating the pod
+STEP: Updating configmap configmap-test-upd-c27a51d9-523f-4551-bd8f-79319d1f4573
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 09:57:56.137: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-5765" for this suite.
 
-• [SLOW TEST:9.266 seconds]
-[sig-apps] Deployment
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  RollingUpdateDeployment should delete old pods and create new ones [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:91.365 seconds]
+[sig-storage] ConfigMap
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
+  updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance]","total":280,"completed":21,"skipped":324,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":20,"skipped":320,"failed":0}
+SSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Container Runtime blackbox test on terminated container 
+  should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Container Runtime
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:35:54.245: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-267
+Feb 12 09:57:56.161: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-runtime
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-1765
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating projection with secret that has name projected-secret-test-3334840d-d4d8-4c7c-bc43-319194a79a1f
-STEP: Creating a pod to test consume secrets
-Aug 30 16:35:54.431: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-f3a9b070-e56f-4950-b6d5-2a3f146979f6" in namespace "projected-267" to be "success or failure"
-Aug 30 16:35:54.436: INFO: Pod "pod-projected-secrets-f3a9b070-e56f-4950-b6d5-2a3f146979f6": Phase="Pending", Reason="", readiness=false. Elapsed: 5.677535ms
-Aug 30 16:35:56.442: INFO: Pod "pod-projected-secrets-f3a9b070-e56f-4950-b6d5-2a3f146979f6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011651506s
-Aug 30 16:35:58.448: INFO: Pod "pod-projected-secrets-f3a9b070-e56f-4950-b6d5-2a3f146979f6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017324748s
-STEP: Saw pod success
-Aug 30 16:35:58.448: INFO: Pod "pod-projected-secrets-f3a9b070-e56f-4950-b6d5-2a3f146979f6" satisfied condition "success or failure"
-Aug 30 16:35:58.452: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-projected-secrets-f3a9b070-e56f-4950-b6d5-2a3f146979f6 container projected-secret-volume-test: 
-STEP: delete the pod
-Aug 30 16:35:58.530: INFO: Waiting for pod pod-projected-secrets-f3a9b070-e56f-4950-b6d5-2a3f146979f6 to disappear
-Aug 30 16:35:58.534: INFO: Pod pod-projected-secrets-f3a9b070-e56f-4950-b6d5-2a3f146979f6 no longer exists
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:35:58.534: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-267" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":22,"skipped":418,"failed":0}
-SSSSSSSSSSSSS
+[It] should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: create the container
+STEP: wait for the container to reach Succeeded
+STEP: get the container status
+STEP: the container should be terminated
+STEP: the termination message should be set
+Feb 12 09:57:58.388: INFO: Expected: &{} to match Container's Termination Message:  --
+STEP: delete the container
+[AfterEach] [k8s.io] Container Runtime
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 09:57:58.410: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-runtime-1765" for this suite.
+•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":280,"completed":21,"skipped":335,"failed":0}
+SSSSSSSS
 ------------------------------
-[sig-network] Services 
-  should serve multiport endpoints from pods  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Security Context When creating a pod with readOnlyRootFilesystem 
+  should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Security Context
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:35:58.548: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename services
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-8854
+Feb 12 09:57:58.426: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename security-context-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in security-context-test-8141
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should serve multiport endpoints from pods  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating service multi-endpoint-test in namespace services-8854
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8854 to expose endpoints map[]
-Aug 30 16:35:58.733: INFO: successfully validated that service multi-endpoint-test in namespace services-8854 exposes endpoints map[] (9.310679ms elapsed)
-STEP: Creating pod pod1 in namespace services-8854
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8854 to expose endpoints map[pod1:[100]]
-Aug 30 16:36:00.799: INFO: successfully validated that service multi-endpoint-test in namespace services-8854 exposes endpoints map[pod1:[100]] (2.049916569s elapsed)
-STEP: Creating pod pod2 in namespace services-8854
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8854 to expose endpoints map[pod1:[100] pod2:[101]]
-Aug 30 16:36:03.872: INFO: successfully validated that service multi-endpoint-test in namespace services-8854 exposes endpoints map[pod1:[100] pod2:[101]] (3.065281094s elapsed)
-STEP: Deleting pod pod1 in namespace services-8854
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8854 to expose endpoints map[pod2:[101]]
-Aug 30 16:36:03.896: INFO: successfully validated that service multi-endpoint-test in namespace services-8854 exposes endpoints map[pod2:[101]] (13.877505ms elapsed)
-STEP: Deleting pod pod2 in namespace services-8854
-STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-8854 to expose endpoints map[]
-Aug 30 16:36:04.919: INFO: successfully validated that service multi-endpoint-test in namespace services-8854 exposes endpoints map[] (1.01204984s elapsed)
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:36:04.942: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-8854" for this suite.
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
-
-• [SLOW TEST:6.412 seconds]
-[sig-network] Services
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should serve multiport endpoints from pods  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] Services should serve multiport endpoints from pods  [Conformance]","total":280,"completed":23,"skipped":431,"failed":0}
-SSSSSSSSSSSS
+[BeforeEach] [k8s.io] Security Context
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39
+[It] should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 09:57:58.614: INFO: Waiting up to 5m0s for pod "busybox-readonly-false-9e823271-0cdb-475b-bd03-179afced60ed" in namespace "security-context-test-8141" to be "success or failure"
+Feb 12 09:57:58.622: INFO: Pod "busybox-readonly-false-9e823271-0cdb-475b-bd03-179afced60ed": Phase="Pending", Reason="", readiness=false. Elapsed: 8.375714ms
+Feb 12 09:58:00.628: INFO: Pod "busybox-readonly-false-9e823271-0cdb-475b-bd03-179afced60ed": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014406091s
+Feb 12 09:58:02.636: INFO: Pod "busybox-readonly-false-9e823271-0cdb-475b-bd03-179afced60ed": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022242236s
+Feb 12 09:58:02.636: INFO: Pod "busybox-readonly-false-9e823271-0cdb-475b-bd03-179afced60ed" satisfied condition "success or failure"
+[AfterEach] [k8s.io] Security Context
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 09:58:02.636: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "security-context-test-8141" for this suite.
+•{"msg":"PASSED [k8s.io] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]","total":280,"completed":22,"skipped":343,"failed":0}
+SSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl replace 
-  should update a single-container pod's image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[sig-cli] Kubectl client Kubectl run --rm job 
+  should create a job from an image, then delete the job [Deprecated] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:36:04.960: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 09:58:02.672: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-6018
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-4867
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[BeforeEach] Kubectl replace
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1790
-[It] should update a single-container pod's image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: running the image docker.io/library/httpd:2.4.38-alpine
-Aug 30 16:36:05.126: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 run e2e-test-httpd-pod --generator=run-pod/v1 --image=docker.io/library/httpd:2.4.38-alpine --labels=run=e2e-test-httpd-pod --namespace=kubectl-6018'
-Aug 30 16:36:05.218: INFO: stderr: ""
-Aug 30 16:36:05.218: INFO: stdout: "pod/e2e-test-httpd-pod created\n"
-STEP: verifying the pod e2e-test-httpd-pod is running
-STEP: verifying the pod e2e-test-httpd-pod was created
-Aug 30 16:36:10.269: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pod e2e-test-httpd-pod --namespace=kubectl-6018 -o json'
-Aug 30 16:36:10.339: INFO: stderr: ""
-Aug 30 16:36:10.339: INFO: stdout: "{\n    \"apiVersion\": \"v1\",\n    \"kind\": \"Pod\",\n    \"metadata\": {\n        \"annotations\": {\n            \"cni.projectcalico.org/podIP\": \"172.25.0.25/32\"\n        },\n        \"creationTimestamp\": \"2020-08-30T16:36:05Z\",\n        \"labels\": {\n            \"run\": \"e2e-test-httpd-pod\"\n        },\n        \"name\": \"e2e-test-httpd-pod\",\n        \"namespace\": \"kubectl-6018\",\n        \"resourceVersion\": \"4197\",\n        \"selfLink\": \"/api/v1/namespaces/kubectl-6018/pods/e2e-test-httpd-pod\",\n        \"uid\": \"349d9913-c5a6-4c56-96ed-19e7c7403449\"\n    },\n    \"spec\": {\n        \"containers\": [\n            {\n                \"image\": \"docker.io/library/httpd:2.4.38-alpine\",\n                \"imagePullPolicy\": \"IfNotPresent\",\n                \"name\": \"e2e-test-httpd-pod\",\n                \"resources\": {},\n                \"terminationMessagePath\": \"/dev/termination-log\",\n                \"terminationMessagePolicy\": \"File\",\n                \"volumeMounts\": [\n                    {\n                        \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n                        \"name\": \"default-token-ntlvp\",\n                        \"readOnly\": true\n                    }\n                ]\n            }\n        ],\n        \"dnsPolicy\": \"ClusterFirst\",\n        \"enableServiceLinks\": true,\n        \"nodeName\": \"adoring-wozniak-54dcfd79fc-948mf\",\n        \"priority\": 0,\n        \"restartPolicy\": \"Always\",\n        \"schedulerName\": \"default-scheduler\",\n        \"securityContext\": {},\n        \"serviceAccount\": \"default\",\n        \"serviceAccountName\": \"default\",\n        \"terminationGracePeriodSeconds\": 30,\n        \"tolerations\": [\n            {\n                \"effect\": \"NoExecute\",\n                \"key\": \"node.kubernetes.io/not-ready\",\n                \"operator\": \"Exists\",\n                \"tolerationSeconds\": 300\n            },\n            {\n                \"effect\": \"NoExecute\",\n                \"key\": \"node.kubernetes.io/unreachable\",\n                \"operator\": \"Exists\",\n                \"tolerationSeconds\": 300\n            }\n        ],\n        \"volumes\": [\n            {\n                \"name\": \"default-token-ntlvp\",\n                \"secret\": {\n                    \"defaultMode\": 420,\n                    \"secretName\": \"default-token-ntlvp\"\n                }\n            }\n        ]\n    },\n    \"status\": {\n        \"conditions\": [\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2020-08-30T16:36:05Z\",\n                \"status\": \"True\",\n                \"type\": \"Initialized\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2020-08-30T16:36:07Z\",\n                \"status\": \"True\",\n                \"type\": \"Ready\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2020-08-30T16:36:07Z\",\n                \"status\": \"True\",\n                \"type\": \"ContainersReady\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2020-08-30T16:36:05Z\",\n                \"status\": \"True\",\n                \"type\": \"PodScheduled\"\n            }\n        ],\n        \"containerStatuses\": [\n            {\n                \"containerID\": \"docker://87d387d33b0b449db9e1ac96dede0e40eed305b8f75ca8eebd0a67e723a6ff5f\",\n                \"image\": \"httpd:2.4.38-alpine\",\n                \"imageID\": \"docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060\",\n                \"lastState\": {},\n                \"name\": \"e2e-test-httpd-pod\",\n                \"ready\": true,\n                \"restartCount\": 0,\n                \"started\": true,\n                \"state\": {\n                    \"running\": {\n                        \"startedAt\": \"2020-08-30T16:36:06Z\"\n                    }\n                }\n            }\n        ],\n        \"hostIP\": \"46.101.135.210\",\n        \"phase\": \"Running\",\n        \"podIP\": \"172.25.0.25\",\n        \"podIPs\": [\n            {\n                \"ip\": \"172.25.0.25\"\n            }\n        ],\n        \"qosClass\": \"BestEffort\",\n        \"startTime\": \"2020-08-30T16:36:05Z\"\n    }\n}\n"
-STEP: replace the image in the pod
-Aug 30 16:36:10.339: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 replace -f - --namespace=kubectl-6018'
-Aug 30 16:36:10.555: INFO: stderr: ""
-Aug 30 16:36:10.555: INFO: stdout: "pod/e2e-test-httpd-pod replaced\n"
-STEP: verifying the pod e2e-test-httpd-pod has the right image docker.io/library/busybox:1.29
-[AfterEach] Kubectl replace
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1795
-Aug 30 16:36:10.562: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete pods e2e-test-httpd-pod --namespace=kubectl-6018'
-Aug 30 16:36:14.276: INFO: stderr: ""
-Aug 30 16:36:14.276: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n"
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[It] should create a job from an image, then delete the job [Deprecated] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: executing a command with run --rm and attach with stdin
+Feb 12 09:58:02.838: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=kubectl-4867 run e2e-test-rm-busybox-job --image=docker.io/library/busybox:1.29 --rm=true --generator=job/v1 --restart=OnFailure --attach=true --stdin -- sh -c cat && echo 'stdin closed''
+Feb 12 09:58:06.168: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\nIf you don't see a command prompt, try pressing enter.\n"
+Feb 12 09:58:06.168: INFO: stdout: "abcd1234stdin closed\njob.batch \"e2e-test-rm-busybox-job\" deleted\n"
+STEP: verifying the job e2e-test-rm-busybox-job was deleted
 [AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:36:14.276: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-6018" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 09:58:08.184: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-4867" for this suite.
 
-• [SLOW TEST:9.330 seconds]
+• [SLOW TEST:5.540 seconds]
 [sig-cli] Kubectl client
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Kubectl replace
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1786
-    should update a single-container pod's image  [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Kubectl run --rm job
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1843
+    should create a job from an image, then delete the job [Deprecated] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Kubectl replace should update a single-container pod's image  [Conformance]","total":280,"completed":24,"skipped":443,"failed":0}
-S
+{"msg":"PASSED [sig-cli] Kubectl client Kubectl run --rm job should create a job from an image, then delete the job [Deprecated] [Conformance]","total":280,"completed":23,"skipped":349,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
-  should execute prestop exec hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Pods 
+  should support retrieving logs from the container over websockets [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:36:14.291: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-lifecycle-hook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-4176
+Feb 12 09:58:08.212: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-5672
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] when create a pod with lifecycle hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
-STEP: create the container to handle the HTTPGet hook request.
-[It] should execute prestop exec hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the pod with lifecycle hook
-STEP: delete the pod with lifecycle hook
-Aug 30 16:36:18.517: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Aug 30 16:36:18.522: INFO: Pod pod-with-prestop-exec-hook still exists
-Aug 30 16:36:20.522: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Aug 30 16:36:20.527: INFO: Pod pod-with-prestop-exec-hook still exists
-Aug 30 16:36:22.522: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
-Aug 30 16:36:22.528: INFO: Pod pod-with-prestop-exec-hook no longer exists
-STEP: check prestop hook
-[AfterEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:36:22.580: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-lifecycle-hook-4176" for this suite.
-
-• [SLOW TEST:8.303 seconds]
-[k8s.io] Container Lifecycle Hook
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  when create a pod with lifecycle hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
-    should execute prestop exec hook properly [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
+[It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 09:58:08.406: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+[AfterEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 09:58:12.594: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-5672" for this suite.
+•{"msg":"PASSED [k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]","total":280,"completed":24,"skipped":373,"failed":0}
+SSSSSSSSS
 ------------------------------
-{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance]","total":280,"completed":25,"skipped":444,"failed":0}
-SSS
+[sig-network] DNS 
+  should support configurable pod DNS nameservers [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 09:58:12.615: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename dns
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-7966
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support configurable pod DNS nameservers [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod with dnsPolicy=None and customized dnsConfig...
+Feb 12 09:58:12.803: INFO: Created pod &Pod{ObjectMeta:{dns-7966  dns-7966 /api/v1/namespaces/dns-7966/pods/dns-7966 7a76af8b-73a3-4b96-8f6d-0fb0e3f1835f 5712 0 2021-02-12 09:58:12 +0000 UTC   map[] map[] [] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-8xf64,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-8xf64,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[pause],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-8xf64,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:None,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:&PodDNSConfig{Nameservers:[1.1.1.1],Searches:[resolv.conf.local],Options:[]PodDNSConfigOption{},},ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+STEP: Verifying customized DNS suffix list is configured on pod...
+Feb 12 09:58:14.824: INFO: ExecWithOptions {Command:[/agnhost dns-suffix] Namespace:dns-7966 PodName:dns-7966 ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 09:58:14.824: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Verifying customized DNS server is configured on pod...
+Feb 12 09:58:15.316: INFO: ExecWithOptions {Command:[/agnhost dns-server-list] Namespace:dns-7966 PodName:dns-7966 ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 09:58:15.316: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 09:58:15.811: INFO: Deleting pod dns-7966...
+[AfterEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 09:58:15.830: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-7966" for this suite.
+•{"msg":"PASSED [sig-network] DNS should support configurable pod DNS nameservers [Conformance]","total":280,"completed":25,"skipped":382,"failed":0}
+SSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] HostPath 
-  should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] HostPath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:36:22.595: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename hostpath
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in hostpath-556
+Feb 12 09:58:15.853: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3858
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] HostPath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/host_path.go:37
-[It] should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test hostPath mode
-Aug 30 16:36:22.779: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "hostpath-556" to be "success or failure"
-Aug 30 16:36:22.786: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 6.828789ms
-Aug 30 16:36:24.791: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.01206146s
+[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name projected-configmap-test-volume-map-83e87f53-c9d8-4c9a-91aa-1b777b35fc38
+STEP: Creating a pod to test consume configMaps
+Feb 12 09:58:16.035: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-a5d8fdf7-2ffa-4115-a142-677e9066a760" in namespace "projected-3858" to be "success or failure"
+Feb 12 09:58:16.041: INFO: Pod "pod-projected-configmaps-a5d8fdf7-2ffa-4115-a142-677e9066a760": Phase="Pending", Reason="", readiness=false. Elapsed: 5.139998ms
+Feb 12 09:58:18.053: INFO: Pod "pod-projected-configmaps-a5d8fdf7-2ffa-4115-a142-677e9066a760": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.017562787s
 STEP: Saw pod success
-Aug 30 16:36:24.791: INFO: Pod "pod-host-path-test" satisfied condition "success or failure"
-Aug 30 16:36:24.796: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-host-path-test container test-container-1: 
+Feb 12 09:58:18.053: INFO: Pod "pod-projected-configmaps-a5d8fdf7-2ffa-4115-a142-677e9066a760" satisfied condition "success or failure"
+Feb 12 09:58:18.059: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-projected-configmaps-a5d8fdf7-2ffa-4115-a142-677e9066a760 container projected-configmap-volume-test: 
 STEP: delete the pod
-Aug 30 16:36:24.861: INFO: Waiting for pod pod-host-path-test to disappear
-Aug 30 16:36:24.865: INFO: Pod pod-host-path-test no longer exists
-[AfterEach] [sig-storage] HostPath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:36:24.865: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "hostpath-556" for this suite.
-•{"msg":"PASSED [sig-storage] HostPath should give a volume the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":26,"skipped":447,"failed":0}
-SSSSSSSSSSSSSSSSSSSS
+Feb 12 09:58:18.137: INFO: Waiting for pod pod-projected-configmaps-a5d8fdf7-2ffa-4115-a142-677e9066a760 to disappear
+Feb 12 09:58:18.142: INFO: Pod pod-projected-configmaps-a5d8fdf7-2ffa-4115-a142-677e9066a760 no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 09:58:18.142: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-3858" for this suite.
+•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":26,"skipped":400,"failed":0}
+SSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] Daemon set [Serial] 
-  should update pod when spec was updated and update strategy is RollingUpdate [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] ResourceQuota 
+  should verify ResourceQuota with best effort scope. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:36:24.879: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename daemonsets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-6694
+Feb 12 09:58:18.167: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename resourcequota
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-6501
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
-[It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 16:36:25.074: INFO: Creating simple daemon set daemon-set
-STEP: Check that daemon pods launch on every node of the cluster.
-Aug 30 16:36:25.094: INFO: Number of nodes with available pods: 0
-Aug 30 16:36:25.094: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:36:26.107: INFO: Number of nodes with available pods: 0
-Aug 30 16:36:26.107: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:36:27.104: INFO: Number of nodes with available pods: 2
-Aug 30 16:36:27.104: INFO: Number of running nodes: 2, number of available pods: 2
-STEP: Update daemon pods image.
-STEP: Check that daemon pods images are updated.
-Aug 30 16:36:27.151: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:27.151: INFO: Wrong image for pod: daemon-set-pmlp8. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:28.163: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:28.163: INFO: Wrong image for pod: daemon-set-pmlp8. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:29.165: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:29.165: INFO: Wrong image for pod: daemon-set-pmlp8. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:30.161: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:30.161: INFO: Wrong image for pod: daemon-set-pmlp8. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:30.161: INFO: Pod daemon-set-pmlp8 is not available
-Aug 30 16:36:31.163: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:31.163: INFO: Pod daemon-set-hqn6x is not available
-Aug 30 16:36:32.164: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:32.164: INFO: Pod daemon-set-hqn6x is not available
-Aug 30 16:36:33.163: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:34.163: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:34.163: INFO: Pod daemon-set-7pjdd is not available
-Aug 30 16:36:35.162: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:35.162: INFO: Pod daemon-set-7pjdd is not available
-Aug 30 16:36:36.164: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:36.164: INFO: Pod daemon-set-7pjdd is not available
-Aug 30 16:36:37.163: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:37.163: INFO: Pod daemon-set-7pjdd is not available
-Aug 30 16:36:38.163: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:38.163: INFO: Pod daemon-set-7pjdd is not available
-Aug 30 16:36:39.162: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:39.162: INFO: Pod daemon-set-7pjdd is not available
-Aug 30 16:36:40.163: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:40.163: INFO: Pod daemon-set-7pjdd is not available
-Aug 30 16:36:41.163: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:41.163: INFO: Pod daemon-set-7pjdd is not available
-Aug 30 16:36:42.164: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:42.164: INFO: Pod daemon-set-7pjdd is not available
-Aug 30 16:36:43.189: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:43.189: INFO: Pod daemon-set-7pjdd is not available
-Aug 30 16:36:44.163: INFO: Wrong image for pod: daemon-set-7pjdd. Expected: gcr.io/kubernetes-e2e-test-images/agnhost:2.8, got: docker.io/library/httpd:2.4.38-alpine.
-Aug 30 16:36:44.163: INFO: Pod daemon-set-7pjdd is not available
-Aug 30 16:36:45.163: INFO: Pod daemon-set-v5tnr is not available
-STEP: Check that daemon pods are still running on every node of the cluster.
-Aug 30 16:36:45.177: INFO: Number of nodes with available pods: 1
-Aug 30 16:36:45.177: INFO: Node adoring-wozniak-54dcfd79fc-948mf is running more than one daemon pod
-Aug 30 16:36:46.187: INFO: Number of nodes with available pods: 1
-Aug 30 16:36:46.187: INFO: Node adoring-wozniak-54dcfd79fc-948mf is running more than one daemon pod
-Aug 30 16:36:47.188: INFO: Number of nodes with available pods: 2
-Aug 30 16:36:47.188: INFO: Number of running nodes: 2, number of available pods: 2
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6694, will wait for the garbage collector to delete the pods
-Aug 30 16:36:47.277: INFO: Deleting DaemonSet.extensions daemon-set took: 9.846717ms
-Aug 30 16:36:47.777: INFO: Terminating DaemonSet.extensions daemon-set pods took: 500.155457ms
-Aug 30 16:36:59.982: INFO: Number of nodes with available pods: 0
-Aug 30 16:36:59.982: INFO: Number of running nodes: 0, number of available pods: 0
-Aug 30 16:36:59.987: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-6694/daemonsets","resourceVersion":"4610"},"items":null}
-
-Aug 30 16:36:59.991: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-6694/pods","resourceVersion":"4610"},"items":null}
-
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:37:00.005: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "daemonsets-6694" for this suite.
+[It] should verify ResourceQuota with best effort scope. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a ResourceQuota with best effort scope
+STEP: Ensuring ResourceQuota status is calculated
+STEP: Creating a ResourceQuota with not best effort scope
+STEP: Ensuring ResourceQuota status is calculated
+STEP: Creating a best-effort pod
+STEP: Ensuring resource quota with best effort scope captures the pod usage
+STEP: Ensuring resource quota with not best effort ignored the pod usage
+STEP: Deleting the pod
+STEP: Ensuring resource quota status released the pod usage
+STEP: Creating a not best-effort pod
+STEP: Ensuring resource quota with not best effort scope captures the pod usage
+STEP: Ensuring resource quota with best effort scope ignored the pod usage
+STEP: Deleting the pod
+STEP: Ensuring resource quota status released the pod usage
+[AfterEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 09:58:34.531: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-6501" for this suite.
 
-• [SLOW TEST:35.140 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should update pod when spec was updated and update strategy is RollingUpdate [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:16.382 seconds]
+[sig-api-machinery] ResourceQuota
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should verify ResourceQuota with best effort scope. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]","total":280,"completed":27,"skipped":467,"failed":0}
-S
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance]","total":280,"completed":27,"skipped":415,"failed":0}
+SSSSSSSSSSS
 ------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates that NodeSelector is respected if matching  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-scheduling] NoExecuteTaintManager Single Pod [Serial] 
+  removing taint cancels eviction [Disruptive] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-scheduling] NoExecuteTaintManager Single Pod [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:37:00.019: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename sched-pred
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-8955
+Feb 12 09:58:34.553: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename taint-single-pod
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in taint-single-pod-5394
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
-Aug 30 16:37:00.188: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-Aug 30 16:37:00.203: INFO: Waiting for terminating namespaces to be deleted...
-Aug 30 16:37:00.208: INFO: 
-Logging pods the kubelet thinks is on node adoring-wozniak-54dcfd79fc-6rshr before test
-Aug 30 16:37:00.266: INFO: canal-54glj from kube-system started at 2020-08-30 16:28:14 +0000 UTC (2 container statuses recorded)
-Aug 30 16:37:00.266: INFO: 	Container calico-node ready: true, restart count 0
-Aug 30 16:37:00.266: INFO: 	Container kube-flannel ready: true, restart count 0
-Aug 30 16:37:00.266: INFO: kube-proxy-wxdxv from kube-system started at 2020-08-30 16:28:14 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.266: INFO: 	Container kube-proxy ready: true, restart count 0
-Aug 30 16:37:00.266: INFO: user-ssh-keys-agent-xwrzj from kube-system started at 2020-08-30 16:28:14 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.266: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
-Aug 30 16:37:00.266: INFO: node-local-dns-cvxnf from kube-system started at 2020-08-30 16:28:34 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.266: INFO: 	Container node-cache ready: true, restart count 0
-Aug 30 16:37:00.266: INFO: logrotate-57bmz from kube-system started at 2020-08-30 16:28:34 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.266: INFO: 	Container logrotate ready: true, restart count 0
-Aug 30 16:37:00.266: INFO: sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-nd5nc from sonobuoy started at 2020-08-30 16:29:24 +0000 UTC (2 container statuses recorded)
-Aug 30 16:37:00.266: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 16:37:00.266: INFO: 	Container systemd-logs ready: true, restart count 0
-Aug 30 16:37:00.266: INFO: coredns-54457d966b-fbnz4 from kube-system started at 2020-08-30 16:28:39 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.266: INFO: 	Container coredns ready: true, restart count 0
-Aug 30 16:37:00.266: INFO: sonobuoy from sonobuoy started at 2020-08-30 16:29:18 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.266: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-Aug 30 16:37:00.266: INFO: sonobuoy-e2e-job-cf49606f646f4c8a from sonobuoy started at 2020-08-30 16:29:23 +0000 UTC (2 container statuses recorded)
-Aug 30 16:37:00.266: INFO: 	Container e2e ready: true, restart count 0
-Aug 30 16:37:00.266: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 16:37:00.266: INFO: 
-Logging pods the kubelet thinks is on node adoring-wozniak-54dcfd79fc-948mf before test
-Aug 30 16:37:00.324: INFO: dashboard-metrics-scraper-59bfc65dc9-94mm2 from kubernetes-dashboard started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.324: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
-Aug 30 16:37:00.324: INFO: coredns-54457d966b-vr27r from kube-system started at 2020-08-30 16:28:39 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.324: INFO: 	Container coredns ready: true, restart count 0
-Aug 30 16:37:00.324: INFO: canal-lg5hn from kube-system started at 2020-08-30 16:28:09 +0000 UTC (2 container statuses recorded)
-Aug 30 16:37:00.324: INFO: 	Container calico-node ready: true, restart count 0
-Aug 30 16:37:00.324: INFO: 	Container kube-flannel ready: true, restart count 0
-Aug 30 16:37:00.324: INFO: user-ssh-keys-agent-vkbs9 from kube-system started at 2020-08-30 16:28:09 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.324: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
-Aug 30 16:37:00.324: INFO: openvpn-client-78d595f58b-pht8r from kube-system started at 2020-08-30 16:28:29 +0000 UTC (2 container statuses recorded)
-Aug 30 16:37:00.324: INFO: 	Container dnat-controller ready: true, restart count 0
-Aug 30 16:37:00.324: INFO: 	Container openvpn-client ready: true, restart count 0
-Aug 30 16:37:00.324: INFO: dashboard-metrics-scraper-59bfc65dc9-fhfrl from kubernetes-dashboard started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.324: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
-Aug 30 16:37:00.324: INFO: node-local-dns-wrs2k from kube-system started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.324: INFO: 	Container node-cache ready: true, restart count 0
-Aug 30 16:37:00.324: INFO: logrotate-pw2ff from kube-system started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.324: INFO: 	Container logrotate ready: true, restart count 0
-Aug 30 16:37:00.324: INFO: sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-cdm8v from sonobuoy started at 2020-08-30 16:29:23 +0000 UTC (2 container statuses recorded)
-Aug 30 16:37:00.324: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 16:37:00.324: INFO: 	Container systemd-logs ready: true, restart count 0
-Aug 30 16:37:00.324: INFO: kube-proxy-v88gx from kube-system started at 2020-08-30 16:28:09 +0000 UTC (1 container statuses recorded)
-Aug 30 16:37:00.324: INFO: 	Container kube-proxy ready: true, restart count 0
-[It] validates that NodeSelector is respected if matching  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Trying to launch a pod without a label to get a node which can launch it.
-STEP: Explicitly delete pod here to free the resource it takes.
-STEP: Trying to apply a random label on the found node.
-STEP: verifying the node has the label kubernetes.io/e2e-326c569c-5555-4d60-af2a-9b273bd14ec5 42
-STEP: Trying to relaunch the pod, now with labels.
-STEP: removing the label kubernetes.io/e2e-326c569c-5555-4d60-af2a-9b273bd14ec5 off the node adoring-wozniak-54dcfd79fc-948mf
-STEP: verifying the node doesn't have the label kubernetes.io/e2e-326c569c-5555-4d60-af2a-9b273bd14ec5
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:37:06.425: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "sched-pred-8955" for this suite.
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
+[BeforeEach] [sig-scheduling] NoExecuteTaintManager Single Pod [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/taints.go:163
+Feb 12 09:58:34.730: INFO: Waiting up to 1m0s for all nodes to be ready
+Feb 12 09:59:34.759: INFO: Waiting for terminating namespaces to be deleted...
+[It] removing taint cancels eviction [Disruptive] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 09:59:34.765: INFO: Starting informer...
+STEP: Starting pod...
+Feb 12 09:59:34.987: INFO: Pod is running on nostalgic-fermat-cdc5d8777-xg74t. Tainting Node
+STEP: Trying to apply a taint on the Node
+STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute
+STEP: Waiting short time to make sure Pod is queued for deletion
+Feb 12 09:59:35.009: INFO: Pod wasn't evicted. Proceeding
+Feb 12 09:59:35.009: INFO: Removing taint from Node
+STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute
+STEP: Waiting some time to make sure that toleration time passed.
+Feb 12 10:00:50.079: INFO: Pod wasn't evicted. Test successful
+[AfterEach] [sig-scheduling] NoExecuteTaintManager Single Pod [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:00:50.079: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "taint-single-pod-5394" for this suite.
 
-• [SLOW TEST:6.419 seconds]
-[sig-scheduling] SchedulerPredicates [Serial]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
-  validates that NodeSelector is respected if matching  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching  [Conformance]","total":280,"completed":28,"skipped":468,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+• [SLOW TEST:135.545 seconds]
+[sig-scheduling] NoExecuteTaintManager Single Pod [Serial]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
+  removing taint cancels eviction [Disruptive] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] Probing container 
-  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-scheduling] NoExecuteTaintManager Single Pod [Serial] removing taint cancels eviction [Disruptive] [Conformance]","total":280,"completed":28,"skipped":426,"failed":0}
+[k8s.io] Pods 
+  should be updated [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:37:06.439: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-7084
+Feb 12 10:00:50.097: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-8454
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod busybox-74156fc0-7904-4b99-92b4-13943f680892 in namespace container-probe-7084
-Aug 30 16:37:10.630: INFO: Started pod busybox-74156fc0-7904-4b99-92b4-13943f680892 in namespace container-probe-7084
-STEP: checking the pod's current state and verifying that restartCount is present
-Aug 30 16:37:10.636: INFO: Initial restart count of pod busybox-74156fc0-7904-4b99-92b4-13943f680892 is 0
-STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:41:11.385: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-7084" for this suite.
-
-• [SLOW TEST:244.962 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":280,"completed":29,"skipped":498,"failed":0}
-S
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
+[It] should be updated [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+STEP: updating the pod
+Feb 12 10:00:52.836: INFO: Successfully updated pod "pod-update-f846027a-090c-4bcb-a8ee-9dc8a193f3bb"
+STEP: verifying the updated pod is in kubernetes
+Feb 12 10:00:52.848: INFO: Pod update OK
+[AfterEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:00:52.848: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-8454" for this suite.
+•{"msg":"PASSED [k8s.io] Pods should be updated [NodeConformance] [Conformance]","total":280,"completed":29,"skipped":426,"failed":0}
+SSSSSSS
 ------------------------------
-[sig-apps] Deployment 
-  deployment should support rollover [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:41:11.401: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename deployment
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-8348
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
-[It] deployment should support rollover [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 16:41:11.570: INFO: Pod name rollover-pod: Found 0 pods out of 1
-Aug 30 16:41:16.575: INFO: Pod name rollover-pod: Found 1 pods out of 1
-STEP: ensuring each pod is running
-Aug 30 16:41:16.575: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready
-Aug 30 16:41:18.581: INFO: Creating deployment "test-rollover-deployment"
-Aug 30 16:41:18.593: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations
-Aug 30 16:41:20.604: INFO: Check revision of new replica set for deployment "test-rollover-deployment"
-Aug 30 16:41:20.614: INFO: Ensure that both replica sets have 1 created replica
-Aug 30 16:41:20.624: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update
-Aug 30 16:41:20.635: INFO: Updating deployment test-rollover-deployment
-Aug 30 16:41:20.635: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller
-Aug 30 16:41:22.646: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2
-Aug 30 16:41:22.657: INFO: Make sure deployment "test-rollover-deployment" is complete
-Aug 30 16:41:22.667: INFO: all replica sets need to contain the pod-template-hash label
-Aug 30 16:41:22.667: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402480, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Aug 30 16:41:24.677: INFO: all replica sets need to contain the pod-template-hash label
-Aug 30 16:41:24.678: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402482, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Aug 30 16:41:26.679: INFO: all replica sets need to contain the pod-template-hash label
-Aug 30 16:41:26.680: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402482, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Aug 30 16:41:28.677: INFO: all replica sets need to contain the pod-template-hash label
-Aug 30 16:41:28.677: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402482, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Aug 30 16:41:30.678: INFO: all replica sets need to contain the pod-template-hash label
-Aug 30 16:41:30.678: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402482, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Aug 30 16:41:32.679: INFO: all replica sets need to contain the pod-template-hash label
-Aug 30 16:41:32.680: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402482, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402478, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Aug 30 16:41:34.679: INFO: 
-Aug 30 16:41:34.679: INFO: Ensure that both old replica sets have no replicas
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
-Aug 30 16:41:34.908: INFO: Deployment "test-rollover-deployment":
-&Deployment{ObjectMeta:{test-rollover-deployment  deployment-8348 /apis/apps/v1/namespaces/deployment-8348/deployments/test-rollover-deployment 55db255b-9dc1-484b-ba21-a5e1e8916064 5756 2 2020-08-30 16:41:18 +0000 UTC   map[name:rollover-pod] map[deployment.kubernetes.io/revision:2] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0000547c8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2020-08-30 16:41:18 +0000 UTC,LastTransitionTime:2020-08-30 16:41:18 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rollover-deployment-574d6dfbff" has successfully progressed.,LastUpdateTime:2020-08-30 16:41:32 +0000 UTC,LastTransitionTime:2020-08-30 16:41:18 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},}
-
-Aug 30 16:41:34.913: INFO: New ReplicaSet "test-rollover-deployment-574d6dfbff" of Deployment "test-rollover-deployment":
-&ReplicaSet{ObjectMeta:{test-rollover-deployment-574d6dfbff  deployment-8348 /apis/apps/v1/namespaces/deployment-8348/replicasets/test-rollover-deployment-574d6dfbff 95ae23a1-fdbc-43ee-b549-3241c7546eb4 5745 2 2020-08-30 16:41:20 +0000 UTC   map[name:rollover-pod pod-template-hash:574d6dfbff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-rollover-deployment 55db255b-9dc1-484b-ba21-a5e1e8916064 0xc0000552d7 0xc0000552d8}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 574d6dfbff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod pod-template-hash:574d6dfbff] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0000553a8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},}
-Aug 30 16:41:34.913: INFO: All old ReplicaSets of Deployment "test-rollover-deployment":
-Aug 30 16:41:34.913: INFO: &ReplicaSet{ObjectMeta:{test-rollover-controller  deployment-8348 /apis/apps/v1/namespaces/deployment-8348/replicasets/test-rollover-controller 5f130d62-ae1d-4830-b32e-368e38c61ffd 5755 2 2020-08-30 16:41:11 +0000 UTC   map[name:rollover-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2] [{apps/v1 Deployment test-rollover-deployment 55db255b-9dc1-484b-ba21-a5e1e8916064 0xc00005506f 0xc0000550d0}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod pod:httpd] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc000055168  ClusterFirst map[]     false false false  PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
-Aug 30 16:41:34.913: INFO: &ReplicaSet{ObjectMeta:{test-rollover-deployment-f6c94f66c  deployment-8348 /apis/apps/v1/namespaces/deployment-8348/replicasets/test-rollover-deployment-f6c94f66c 17632baf-bd21-4657-9047-262037f9f705 5691 2 2020-08-30 16:41:18 +0000 UTC   map[name:rollover-pod pod-template-hash:f6c94f66c] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-rollover-deployment 55db255b-9dc1-484b-ba21-a5e1e8916064 0xc0000554a0 0xc0000554a1}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: f6c94f66c,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod pod-template-hash:f6c94f66c] map[] [] []  []} {[] [] [{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc000055578  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
-Aug 30 16:41:35.140: INFO: Pod "test-rollover-deployment-574d6dfbff-98nwc" is available:
-&Pod{ObjectMeta:{test-rollover-deployment-574d6dfbff-98nwc test-rollover-deployment-574d6dfbff- deployment-8348 /api/v1/namespaces/deployment-8348/pods/test-rollover-deployment-574d6dfbff-98nwc bf49bf93-e84f-4685-b47f-ed5b53c2deaf 5706 0 2020-08-30 16:41:20 +0000 UTC   map[name:rollover-pod pod-template-hash:574d6dfbff] map[cni.projectcalico.org/podIP:172.25.1.16/32] [{apps/v1 ReplicaSet test-rollover-deployment-574d6dfbff 95ae23a1-fdbc-43ee-b549-3241c7546eb4 0xc000786827 0xc000786828}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-pmpwl,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-pmpwl,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-pmpwl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:41:20 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:41:22 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:41:22 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:41:20 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:172.25.1.16,StartTime:2020-08-30 16:41:20 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-08-30 16:41:21 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/agnhost@sha256:daf5332100521b1256d0e3c56d697a238eaec3af48897ed9167cbadd426773b5,ContainerID:docker://d0dda35f351dd9b7f0522f5680db98e55e2a84af41f7e7a6ce765ddf0e2faeea,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.16,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:41:35.140: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "deployment-8348" for this suite.
-
-• [SLOW TEST:23.756 seconds]
-[sig-apps] Deployment
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  deployment should support rollover [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] Deployment deployment should support rollover [Conformance]","total":280,"completed":30,"skipped":499,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-auth] ServiceAccounts 
-  should allow opting out of API token automount  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-auth] ServiceAccounts
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:41:35.158: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename svcaccounts
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svcaccounts-3676
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should allow opting out of API token automount  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: getting the auto-created API token
-Aug 30 16:41:36.604: INFO: created pod pod-service-account-defaultsa
-Aug 30 16:41:36.604: INFO: pod pod-service-account-defaultsa service account token volume mount: true
-Aug 30 16:41:36.613: INFO: created pod pod-service-account-mountsa
-Aug 30 16:41:36.613: INFO: pod pod-service-account-mountsa service account token volume mount: true
-Aug 30 16:41:36.621: INFO: created pod pod-service-account-nomountsa
-Aug 30 16:41:36.621: INFO: pod pod-service-account-nomountsa service account token volume mount: false
-Aug 30 16:41:36.630: INFO: created pod pod-service-account-defaultsa-mountspec
-Aug 30 16:41:36.630: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true
-Aug 30 16:41:36.641: INFO: created pod pod-service-account-mountsa-mountspec
-Aug 30 16:41:36.641: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true
-Aug 30 16:41:36.664: INFO: created pod pod-service-account-nomountsa-mountspec
-Aug 30 16:41:36.664: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true
-Aug 30 16:41:36.681: INFO: created pod pod-service-account-defaultsa-nomountspec
-Aug 30 16:41:36.681: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false
-Aug 30 16:41:36.701: INFO: created pod pod-service-account-mountsa-nomountspec
-Aug 30 16:41:36.702: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false
-Aug 30 16:41:36.710: INFO: created pod pod-service-account-nomountsa-nomountspec
-Aug 30 16:41:36.710: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false
-[AfterEach] [sig-auth] ServiceAccounts
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:41:36.710: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "svcaccounts-3676" for this suite.
-•{"msg":"PASSED [sig-auth] ServiceAccounts should allow opting out of API token automount  [Conformance]","total":280,"completed":31,"skipped":527,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-storage] Downward API volume 
-  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Projected secret 
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:41:36.734: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-8207
+Feb 12 10:00:52.869: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9352
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 16:41:36.955: INFO: Waiting up to 5m0s for pod "downwardapi-volume-1388654e-32f8-48ef-8d90-3fe1f2c32367" in namespace "downward-api-8207" to be "success or failure"
-Aug 30 16:41:36.965: INFO: Pod "downwardapi-volume-1388654e-32f8-48ef-8d90-3fe1f2c32367": Phase="Pending", Reason="", readiness=false. Elapsed: 10.360808ms
-Aug 30 16:41:38.971: INFO: Pod "downwardapi-volume-1388654e-32f8-48ef-8d90-3fe1f2c32367": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015932416s
-Aug 30 16:41:40.977: INFO: Pod "downwardapi-volume-1388654e-32f8-48ef-8d90-3fe1f2c32367": Phase="Pending", Reason="", readiness=false. Elapsed: 4.021601996s
-Aug 30 16:41:42.982: INFO: Pod "downwardapi-volume-1388654e-32f8-48ef-8d90-3fe1f2c32367": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.02662219s
-STEP: Saw pod success
-Aug 30 16:41:42.982: INFO: Pod "downwardapi-volume-1388654e-32f8-48ef-8d90-3fe1f2c32367" satisfied condition "success or failure"
-Aug 30 16:41:42.987: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod downwardapi-volume-1388654e-32f8-48ef-8d90-3fe1f2c32367 container client-container: 
-STEP: delete the pod
-Aug 30 16:41:43.054: INFO: Waiting for pod downwardapi-volume-1388654e-32f8-48ef-8d90-3fe1f2c32367 to disappear
-Aug 30 16:41:43.058: INFO: Pod downwardapi-volume-1388654e-32f8-48ef-8d90-3fe1f2c32367 no longer exists
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:41:43.058: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-8207" for this suite.
-
-• [SLOW TEST:6.337 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:35
-  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":32,"skipped":598,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating secret with name s-test-opt-del-e69a447e-796b-447f-8eaf-b3dc2f783602
+STEP: Creating secret with name s-test-opt-upd-3a53294b-41f3-4036-a9dd-dd1c827a8e6c
+STEP: Creating the pod
+STEP: Deleting secret s-test-opt-del-e69a447e-796b-447f-8eaf-b3dc2f783602
+STEP: Updating secret s-test-opt-upd-3a53294b-41f3-4036-a9dd-dd1c827a8e6c
+STEP: Creating secret with name s-test-opt-create-201c9e53-f0fb-4c0d-9c91-8034b791fda6
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] Projected secret
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:00:57.629: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9352" for this suite.
+•{"msg":"PASSED [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":30,"skipped":433,"failed":0}
+SSSSSSSSSSSSSS
 ------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] ResourceQuota 
+  should create a ResourceQuota and capture the life of a replication controller. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:41:43.072: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename sched-pred
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-8930
+Feb 12 10:00:57.650: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename resourcequota
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-1581
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
-Aug 30 16:41:43.237: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-Aug 30 16:41:43.254: INFO: Waiting for terminating namespaces to be deleted...
-Aug 30 16:41:43.258: INFO: 
-Logging pods the kubelet thinks is on node adoring-wozniak-54dcfd79fc-6rshr before test
-Aug 30 16:41:43.314: INFO: sonobuoy from sonobuoy started at 2020-08-30 16:29:18 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.314: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-Aug 30 16:41:43.314: INFO: logrotate-57bmz from kube-system started at 2020-08-30 16:28:34 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.314: INFO: 	Container logrotate ready: true, restart count 0
-Aug 30 16:41:43.314: INFO: sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-nd5nc from sonobuoy started at 2020-08-30 16:29:24 +0000 UTC (2 container statuses recorded)
-Aug 30 16:41:43.314: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 16:41:43.314: INFO: 	Container systemd-logs ready: true, restart count 0
-Aug 30 16:41:43.314: INFO: kube-proxy-wxdxv from kube-system started at 2020-08-30 16:28:14 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.314: INFO: 	Container kube-proxy ready: true, restart count 0
-Aug 30 16:41:43.314: INFO: node-local-dns-cvxnf from kube-system started at 2020-08-30 16:28:34 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.314: INFO: 	Container node-cache ready: true, restart count 0
-Aug 30 16:41:43.314: INFO: coredns-54457d966b-fbnz4 from kube-system started at 2020-08-30 16:28:39 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.314: INFO: 	Container coredns ready: true, restart count 0
-Aug 30 16:41:43.314: INFO: sonobuoy-e2e-job-cf49606f646f4c8a from sonobuoy started at 2020-08-30 16:29:23 +0000 UTC (2 container statuses recorded)
-Aug 30 16:41:43.314: INFO: 	Container e2e ready: true, restart count 0
-Aug 30 16:41:43.314: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 16:41:43.314: INFO: canal-54glj from kube-system started at 2020-08-30 16:28:14 +0000 UTC (2 container statuses recorded)
-Aug 30 16:41:43.314: INFO: 	Container calico-node ready: true, restart count 0
-Aug 30 16:41:43.314: INFO: 	Container kube-flannel ready: true, restart count 0
-Aug 30 16:41:43.314: INFO: user-ssh-keys-agent-xwrzj from kube-system started at 2020-08-30 16:28:14 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.314: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
-Aug 30 16:41:43.314: INFO: 
-Logging pods the kubelet thinks is on node adoring-wozniak-54dcfd79fc-948mf before test
-Aug 30 16:41:43.412: INFO: dashboard-metrics-scraper-59bfc65dc9-94mm2 from kubernetes-dashboard started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.412: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
-Aug 30 16:41:43.412: INFO: kube-proxy-v88gx from kube-system started at 2020-08-30 16:28:09 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.412: INFO: 	Container kube-proxy ready: true, restart count 0
-Aug 30 16:41:43.412: INFO: user-ssh-keys-agent-vkbs9 from kube-system started at 2020-08-30 16:28:09 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.412: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
-Aug 30 16:41:43.412: INFO: openvpn-client-78d595f58b-pht8r from kube-system started at 2020-08-30 16:28:29 +0000 UTC (2 container statuses recorded)
-Aug 30 16:41:43.412: INFO: 	Container dnat-controller ready: true, restart count 0
-Aug 30 16:41:43.412: INFO: 	Container openvpn-client ready: true, restart count 0
-Aug 30 16:41:43.412: INFO: dashboard-metrics-scraper-59bfc65dc9-fhfrl from kubernetes-dashboard started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.412: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
-Aug 30 16:41:43.412: INFO: logrotate-pw2ff from kube-system started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.412: INFO: 	Container logrotate ready: true, restart count 0
-Aug 30 16:41:43.412: INFO: sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-cdm8v from sonobuoy started at 2020-08-30 16:29:23 +0000 UTC (2 container statuses recorded)
-Aug 30 16:41:43.412: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 16:41:43.412: INFO: 	Container systemd-logs ready: true, restart count 0
-Aug 30 16:41:43.412: INFO: canal-lg5hn from kube-system started at 2020-08-30 16:28:09 +0000 UTC (2 container statuses recorded)
-Aug 30 16:41:43.412: INFO: 	Container calico-node ready: true, restart count 0
-Aug 30 16:41:43.412: INFO: 	Container kube-flannel ready: true, restart count 0
-Aug 30 16:41:43.412: INFO: coredns-54457d966b-vr27r from kube-system started at 2020-08-30 16:28:39 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.412: INFO: 	Container coredns ready: true, restart count 0
-Aug 30 16:41:43.412: INFO: node-local-dns-wrs2k from kube-system started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:41:43.412: INFO: 	Container node-cache ready: true, restart count 0
-[It] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Trying to launch a pod without a label to get a node which can launch it.
-STEP: Explicitly delete pod here to free the resource it takes.
-STEP: Trying to apply a random label on the found node.
-STEP: verifying the node has the label kubernetes.io/e2e-3d26398b-7c08-48c2-8a84-60b584cd313f 95
-STEP: Trying to create a pod(pod4) with hostport 54322 and hostIP 0.0.0.0(empty string here) and expect scheduled
-STEP: Trying to create another pod(pod5) with hostport 54322 but hostIP 127.0.0.1 on the node which pod4 resides and expect not scheduled
-STEP: removing the label kubernetes.io/e2e-3d26398b-7c08-48c2-8a84-60b584cd313f off the node adoring-wozniak-54dcfd79fc-948mf
-STEP: verifying the node doesn't have the label kubernetes.io/e2e-3d26398b-7c08-48c2-8a84-60b584cd313f
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:46:47.555: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "sched-pred-8930" for this suite.
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
+[It] should create a ResourceQuota and capture the life of a replication controller. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Counting existing ResourceQuota
+STEP: Creating a ResourceQuota
+STEP: Ensuring resource quota status is calculated
+STEP: Creating a ReplicationController
+STEP: Ensuring resource quota status captures replication controller creation
+STEP: Deleting a ReplicationController
+STEP: Ensuring resource quota status released usage
+[AfterEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:01:08.914: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-1581" for this suite.
 
-• [SLOW TEST:304.497 seconds]
-[sig-scheduling] SchedulerPredicates [Serial]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
-  validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:11.287 seconds]
+[sig-api-machinery] ResourceQuota
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should create a ResourceQuota and capture the life of a replication controller. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]","total":280,"completed":33,"skipped":652,"failed":0}
-SSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance]","total":280,"completed":31,"skipped":447,"failed":0}
+SSSSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide container's cpu request [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition 
+  listing custom resource definition objects works  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:46:47.569: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-4100
+Feb 12 10:01:08.938: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename custom-resource-definition
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-730
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should provide container's cpu request [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 16:46:47.770: INFO: Waiting up to 5m0s for pod "downwardapi-volume-69f49be4-38bf-416a-a85f-27586485f6d6" in namespace "downward-api-4100" to be "success or failure"
-Aug 30 16:46:47.776: INFO: Pod "downwardapi-volume-69f49be4-38bf-416a-a85f-27586485f6d6": Phase="Pending", Reason="", readiness=false. Elapsed: 6.555073ms
-Aug 30 16:46:49.782: INFO: Pod "downwardapi-volume-69f49be4-38bf-416a-a85f-27586485f6d6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012197904s
-Aug 30 16:46:51.788: INFO: Pod "downwardapi-volume-69f49be4-38bf-416a-a85f-27586485f6d6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018192108s
-STEP: Saw pod success
-Aug 30 16:46:51.788: INFO: Pod "downwardapi-volume-69f49be4-38bf-416a-a85f-27586485f6d6" satisfied condition "success or failure"
-Aug 30 16:46:51.793: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-69f49be4-38bf-416a-a85f-27586485f6d6 container client-container: 
-STEP: delete the pod
-Aug 30 16:46:51.863: INFO: Waiting for pod downwardapi-volume-69f49be4-38bf-416a-a85f-27586485f6d6 to disappear
-Aug 30 16:46:51.867: INFO: Pod downwardapi-volume-69f49be4-38bf-416a-a85f-27586485f6d6 no longer exists
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:46:51.867: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-4100" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance]","total":280,"completed":34,"skipped":666,"failed":0}
-SSSSSSSSS
+[It] listing custom resource definition objects works  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:01:09.227: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:01:13.012: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "custom-resource-definition-730" for this suite.
+•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works  [Conformance]","total":280,"completed":32,"skipped":453,"failed":0}
+SSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] Service endpoints latency 
+  should not be very high  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Service endpoints latency
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:46:51.889: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-8680
+Feb 12 10:01:13.039: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename svc-latency
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svc-latency-2063
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name cm-test-opt-del-162574cd-903d-4d50-a5e8-1cbf4afce40a
-STEP: Creating configMap with name cm-test-opt-upd-8220f7d1-abe6-4029-b935-9c73e5c83adb
-STEP: Creating the pod
-STEP: Deleting configmap cm-test-opt-del-162574cd-903d-4d50-a5e8-1cbf4afce40a
-STEP: Updating configmap cm-test-opt-upd-8220f7d1-abe6-4029-b935-9c73e5c83adb
-STEP: Creating configMap with name cm-test-opt-create-c29ce3f8-28c0-46e3-9338-78377a0fd8f4
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:48:27.354: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-8680" for this suite.
+[It] should not be very high  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:01:13.213: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: creating replication controller svc-latency-rc in namespace svc-latency-2063
+I0212 10:01:13.230738      21 runners.go:189] Created replication controller with name: svc-latency-rc, namespace: svc-latency-2063, replica count: 1
+I0212 10:01:14.281754      21 runners.go:189] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+I0212 10:01:15.281992      21 runners.go:189] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+I0212 10:01:16.282214      21 runners.go:189] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+Feb 12 10:01:16.401: INFO: Created: latency-svc-mcnjb
+Feb 12 10:01:16.417: INFO: Got endpoints: latency-svc-mcnjb [35.400395ms]
+Feb 12 10:01:16.463: INFO: Created: latency-svc-5dn5g
+Feb 12 10:01:16.471: INFO: Created: latency-svc-5jnl5
+Feb 12 10:01:16.474: INFO: Got endpoints: latency-svc-5dn5g [56.340677ms]
+Feb 12 10:01:16.477: INFO: Created: latency-svc-5nx6c
+Feb 12 10:01:16.484: INFO: Got endpoints: latency-svc-5jnl5 [65.706771ms]
+Feb 12 10:01:16.489: INFO: Got endpoints: latency-svc-5nx6c [71.284207ms]
+Feb 12 10:01:16.493: INFO: Created: latency-svc-9xcb5
+Feb 12 10:01:16.498: INFO: Got endpoints: latency-svc-9xcb5 [80.336984ms]
+Feb 12 10:01:16.501: INFO: Created: latency-svc-k2rcc
+Feb 12 10:01:16.510: INFO: Created: latency-svc-5xz5b
+Feb 12 10:01:16.514: INFO: Got endpoints: latency-svc-k2rcc [95.200899ms]
+Feb 12 10:01:16.520: INFO: Got endpoints: latency-svc-5xz5b [101.40848ms]
+Feb 12 10:01:16.524: INFO: Created: latency-svc-mjqzz
+Feb 12 10:01:16.534: INFO: Got endpoints: latency-svc-mjqzz [115.156862ms]
+Feb 12 10:01:16.534: INFO: Created: latency-svc-lhrk5
+Feb 12 10:01:16.548: INFO: Got endpoints: latency-svc-lhrk5 [34.262356ms]
+Feb 12 10:01:16.548: INFO: Created: latency-svc-d2gw6
+Feb 12 10:01:16.557: INFO: Got endpoints: latency-svc-d2gw6 [138.07277ms]
+Feb 12 10:01:16.568: INFO: Created: latency-svc-xvp9w
+Feb 12 10:01:16.594: INFO: Created: latency-svc-tdb7k
+Feb 12 10:01:16.596: INFO: Got endpoints: latency-svc-xvp9w [177.488871ms]
+Feb 12 10:01:16.600: INFO: Got endpoints: latency-svc-tdb7k [179.846986ms]
+Feb 12 10:01:16.603: INFO: Created: latency-svc-7lcv8
+Feb 12 10:01:16.610: INFO: Got endpoints: latency-svc-7lcv8 [191.190222ms]
+Feb 12 10:01:16.612: INFO: Created: latency-svc-s48gr
+Feb 12 10:01:16.619: INFO: Got endpoints: latency-svc-s48gr [198.763389ms]
+Feb 12 10:01:16.622: INFO: Created: latency-svc-krsfr
+Feb 12 10:01:16.630: INFO: Got endpoints: latency-svc-krsfr [209.894152ms]
+Feb 12 10:01:16.636: INFO: Created: latency-svc-b5c2w
+Feb 12 10:01:16.639: INFO: Got endpoints: latency-svc-b5c2w [218.129055ms]
+Feb 12 10:01:16.650: INFO: Created: latency-svc-wttsd
+Feb 12 10:01:16.657: INFO: Got endpoints: latency-svc-wttsd [235.76627ms]
+Feb 12 10:01:16.667: INFO: Created: latency-svc-b4w8g
+Feb 12 10:01:16.668: INFO: Created: latency-svc-zfbc5
+Feb 12 10:01:16.670: INFO: Got endpoints: latency-svc-b4w8g [195.928223ms]
+Feb 12 10:01:16.677: INFO: Got endpoints: latency-svc-zfbc5 [193.655372ms]
+Feb 12 10:01:16.679: INFO: Created: latency-svc-2tfwg
+Feb 12 10:01:16.686: INFO: Got endpoints: latency-svc-2tfwg [197.143884ms]
+Feb 12 10:01:16.688: INFO: Created: latency-svc-m7k7r
+Feb 12 10:01:16.697: INFO: Created: latency-svc-7bc5b
+Feb 12 10:01:16.698: INFO: Got endpoints: latency-svc-m7k7r [199.134311ms]
+Feb 12 10:01:16.708: INFO: Got endpoints: latency-svc-7bc5b [187.444769ms]
+Feb 12 10:01:16.715: INFO: Created: latency-svc-r6hkl
+Feb 12 10:01:16.717: INFO: Got endpoints: latency-svc-r6hkl [183.196079ms]
+Feb 12 10:01:16.720: INFO: Created: latency-svc-c9f85
+Feb 12 10:01:16.727: INFO: Got endpoints: latency-svc-c9f85 [179.406789ms]
+Feb 12 10:01:16.731: INFO: Created: latency-svc-rs9vt
+Feb 12 10:01:16.739: INFO: Got endpoints: latency-svc-rs9vt [181.841958ms]
+Feb 12 10:01:16.744: INFO: Created: latency-svc-phclh
+Feb 12 10:01:16.751: INFO: Got endpoints: latency-svc-phclh [154.673879ms]
+Feb 12 10:01:16.754: INFO: Created: latency-svc-qbnp2
+Feb 12 10:01:16.760: INFO: Got endpoints: latency-svc-qbnp2 [159.85863ms]
+Feb 12 10:01:16.764: INFO: Created: latency-svc-vqbrr
+Feb 12 10:01:16.776: INFO: Got endpoints: latency-svc-vqbrr [165.891996ms]
+Feb 12 10:01:16.779: INFO: Created: latency-svc-fbbjp
+Feb 12 10:01:16.787: INFO: Got endpoints: latency-svc-fbbjp [167.627687ms]
+Feb 12 10:01:16.789: INFO: Created: latency-svc-wr62p
+Feb 12 10:01:16.796: INFO: Got endpoints: latency-svc-wr62p [165.545937ms]
+Feb 12 10:01:16.799: INFO: Created: latency-svc-96kfx
+Feb 12 10:01:16.811: INFO: Created: latency-svc-29v27
+Feb 12 10:01:16.811: INFO: Got endpoints: latency-svc-96kfx [171.816617ms]
+Feb 12 10:01:16.817: INFO: Got endpoints: latency-svc-29v27 [159.765072ms]
+Feb 12 10:01:16.831: INFO: Created: latency-svc-fbs8g
+Feb 12 10:01:16.844: INFO: Created: latency-svc-7xrkx
+Feb 12 10:01:16.844: INFO: Got endpoints: latency-svc-fbs8g [173.877715ms]
+Feb 12 10:01:16.852: INFO: Got endpoints: latency-svc-7xrkx [174.216326ms]
+Feb 12 10:01:16.855: INFO: Created: latency-svc-gn72t
+Feb 12 10:01:16.862: INFO: Created: latency-svc-ljdsb
+Feb 12 10:01:16.866: INFO: Got endpoints: latency-svc-gn72t [179.887899ms]
+Feb 12 10:01:16.869: INFO: Created: latency-svc-sm9vx
+Feb 12 10:01:16.879: INFO: Got endpoints: latency-svc-ljdsb [181.409711ms]
+Feb 12 10:01:16.896: INFO: Got endpoints: latency-svc-sm9vx [188.361973ms]
+Feb 12 10:01:16.902: INFO: Created: latency-svc-b9j86
+Feb 12 10:01:16.913: INFO: Created: latency-svc-z9gm7
+Feb 12 10:01:16.913: INFO: Got endpoints: latency-svc-b9j86 [196.396308ms]
+Feb 12 10:01:16.921: INFO: Created: latency-svc-v9cz8
+Feb 12 10:01:16.922: INFO: Got endpoints: latency-svc-z9gm7 [194.988195ms]
+Feb 12 10:01:16.934: INFO: Got endpoints: latency-svc-v9cz8 [195.604563ms]
+Feb 12 10:01:16.942: INFO: Created: latency-svc-ht8c5
+Feb 12 10:01:16.972: INFO: Created: latency-svc-qr6k2
+Feb 12 10:01:16.972: INFO: Got endpoints: latency-svc-ht8c5 [221.050463ms]
+Feb 12 10:01:16.972: INFO: Created: latency-svc-plqnn
+Feb 12 10:01:16.979: INFO: Created: latency-svc-k8s4r
+Feb 12 10:01:16.989: INFO: Created: latency-svc-j4hv7
+Feb 12 10:01:16.996: INFO: Created: latency-svc-jcgdb
+Feb 12 10:01:17.005: INFO: Created: latency-svc-vfqqd
+Feb 12 10:01:17.009: INFO: Got endpoints: latency-svc-plqnn [248.761994ms]
+Feb 12 10:01:17.018: INFO: Created: latency-svc-whknv
+Feb 12 10:01:17.022: INFO: Created: latency-svc-g52nz
+Feb 12 10:01:17.030: INFO: Created: latency-svc-2md6l
+Feb 12 10:01:17.044: INFO: Created: latency-svc-w544w
+Feb 12 10:01:17.049: INFO: Created: latency-svc-m8wwq
+Feb 12 10:01:17.062: INFO: Created: latency-svc-8h46q
+Feb 12 10:01:17.062: INFO: Got endpoints: latency-svc-qr6k2 [286.004ms]
+Feb 12 10:01:17.072: INFO: Created: latency-svc-6v6gd
+Feb 12 10:01:17.080: INFO: Created: latency-svc-hzl86
+Feb 12 10:01:17.088: INFO: Created: latency-svc-2c4g2
+Feb 12 10:01:17.099: INFO: Created: latency-svc-4m5l5
+Feb 12 10:01:17.101: INFO: Created: latency-svc-9xz5p
+Feb 12 10:01:17.111: INFO: Got endpoints: latency-svc-k8s4r [323.941366ms]
+Feb 12 10:01:17.128: INFO: Created: latency-svc-dkrkw
+Feb 12 10:01:17.159: INFO: Got endpoints: latency-svc-j4hv7 [362.780374ms]
+Feb 12 10:01:17.175: INFO: Created: latency-svc-s5stg
+Feb 12 10:01:17.212: INFO: Got endpoints: latency-svc-jcgdb [401.133673ms]
+Feb 12 10:01:17.229: INFO: Created: latency-svc-hhtqf
+Feb 12 10:01:17.261: INFO: Got endpoints: latency-svc-vfqqd [444.76335ms]
+Feb 12 10:01:17.285: INFO: Created: latency-svc-8t69b
+Feb 12 10:01:17.308: INFO: Got endpoints: latency-svc-whknv [464.444243ms]
+Feb 12 10:01:17.329: INFO: Created: latency-svc-9cpzq
+Feb 12 10:01:17.363: INFO: Got endpoints: latency-svc-g52nz [511.205968ms]
+Feb 12 10:01:17.381: INFO: Created: latency-svc-cd6m7
+Feb 12 10:01:17.408: INFO: Got endpoints: latency-svc-2md6l [542.486863ms]
+Feb 12 10:01:17.426: INFO: Created: latency-svc-jqjkr
+Feb 12 10:01:17.460: INFO: Got endpoints: latency-svc-w544w [580.504494ms]
+Feb 12 10:01:17.481: INFO: Created: latency-svc-t52p9
+Feb 12 10:01:17.508: INFO: Got endpoints: latency-svc-m8wwq [611.926688ms]
+Feb 12 10:01:17.522: INFO: Created: latency-svc-hp5tr
+Feb 12 10:01:17.559: INFO: Got endpoints: latency-svc-8h46q [645.707222ms]
+Feb 12 10:01:17.579: INFO: Created: latency-svc-pl4qm
+Feb 12 10:01:17.608: INFO: Got endpoints: latency-svc-6v6gd [685.586515ms]
+Feb 12 10:01:17.624: INFO: Created: latency-svc-4ddph
+Feb 12 10:01:17.659: INFO: Got endpoints: latency-svc-hzl86 [724.566477ms]
+Feb 12 10:01:17.673: INFO: Created: latency-svc-jkwm2
+Feb 12 10:01:17.708: INFO: Got endpoints: latency-svc-2c4g2 [735.881376ms]
+Feb 12 10:01:17.727: INFO: Created: latency-svc-vz7vq
+Feb 12 10:01:17.763: INFO: Got endpoints: latency-svc-4m5l5 [754.220729ms]
+Feb 12 10:01:17.779: INFO: Created: latency-svc-gfvdd
+Feb 12 10:01:17.809: INFO: Got endpoints: latency-svc-9xz5p [746.654595ms]
+Feb 12 10:01:17.826: INFO: Created: latency-svc-fkpkb
+Feb 12 10:01:17.859: INFO: Got endpoints: latency-svc-dkrkw [748.068551ms]
+Feb 12 10:01:17.876: INFO: Created: latency-svc-l2pjl
+Feb 12 10:01:17.908: INFO: Got endpoints: latency-svc-s5stg [749.235726ms]
+Feb 12 10:01:17.927: INFO: Created: latency-svc-x7wlf
+Feb 12 10:01:17.958: INFO: Got endpoints: latency-svc-hhtqf [745.886725ms]
+Feb 12 10:01:17.975: INFO: Created: latency-svc-p2c88
+Feb 12 10:01:18.009: INFO: Got endpoints: latency-svc-8t69b [747.300153ms]
+Feb 12 10:01:18.028: INFO: Created: latency-svc-2lrrz
+Feb 12 10:01:18.060: INFO: Got endpoints: latency-svc-9cpzq [751.286034ms]
+Feb 12 10:01:18.080: INFO: Created: latency-svc-js7vz
+Feb 12 10:01:18.107: INFO: Got endpoints: latency-svc-cd6m7 [743.743849ms]
+Feb 12 10:01:18.123: INFO: Created: latency-svc-tr8cx
+Feb 12 10:01:18.159: INFO: Got endpoints: latency-svc-jqjkr [750.920833ms]
+Feb 12 10:01:18.182: INFO: Created: latency-svc-tn56h
+Feb 12 10:01:18.208: INFO: Got endpoints: latency-svc-t52p9 [748.405067ms]
+Feb 12 10:01:18.225: INFO: Created: latency-svc-ql9j2
+Feb 12 10:01:18.259: INFO: Got endpoints: latency-svc-hp5tr [750.469081ms]
+Feb 12 10:01:18.274: INFO: Created: latency-svc-bwdsq
+Feb 12 10:01:18.308: INFO: Got endpoints: latency-svc-pl4qm [748.644845ms]
+Feb 12 10:01:18.321: INFO: Created: latency-svc-xj824
+Feb 12 10:01:18.361: INFO: Got endpoints: latency-svc-4ddph [752.639311ms]
+Feb 12 10:01:18.377: INFO: Created: latency-svc-mm48t
+Feb 12 10:01:18.408: INFO: Got endpoints: latency-svc-jkwm2 [749.526151ms]
+Feb 12 10:01:18.424: INFO: Created: latency-svc-fpj6n
+Feb 12 10:01:18.457: INFO: Got endpoints: latency-svc-vz7vq [748.765592ms]
+Feb 12 10:01:18.472: INFO: Created: latency-svc-w4h4w
+Feb 12 10:01:18.508: INFO: Got endpoints: latency-svc-gfvdd [744.428009ms]
+Feb 12 10:01:18.523: INFO: Created: latency-svc-t65df
+Feb 12 10:01:18.558: INFO: Got endpoints: latency-svc-fkpkb [748.558501ms]
+Feb 12 10:01:18.586: INFO: Created: latency-svc-czk8l
+Feb 12 10:01:18.609: INFO: Got endpoints: latency-svc-l2pjl [749.620357ms]
+Feb 12 10:01:18.627: INFO: Created: latency-svc-jxmbf
+Feb 12 10:01:18.669: INFO: Got endpoints: latency-svc-x7wlf [760.61899ms]
+Feb 12 10:01:18.686: INFO: Created: latency-svc-n4ck7
+Feb 12 10:01:18.708: INFO: Got endpoints: latency-svc-p2c88 [749.875783ms]
+Feb 12 10:01:18.736: INFO: Created: latency-svc-l9kj7
+Feb 12 10:01:18.767: INFO: Got endpoints: latency-svc-2lrrz [758.117611ms]
+Feb 12 10:01:18.792: INFO: Created: latency-svc-7cs7h
+Feb 12 10:01:18.813: INFO: Got endpoints: latency-svc-js7vz [750.969601ms]
+Feb 12 10:01:18.832: INFO: Created: latency-svc-mztzp
+Feb 12 10:01:18.862: INFO: Got endpoints: latency-svc-tr8cx [755.507746ms]
+Feb 12 10:01:18.893: INFO: Created: latency-svc-mf7rz
+Feb 12 10:01:18.910: INFO: Got endpoints: latency-svc-tn56h [750.348502ms]
+Feb 12 10:01:18.944: INFO: Created: latency-svc-6f9lg
+Feb 12 10:01:18.982: INFO: Got endpoints: latency-svc-ql9j2 [773.504893ms]
+Feb 12 10:01:19.013: INFO: Got endpoints: latency-svc-bwdsq [753.583203ms]
+Feb 12 10:01:19.017: INFO: Created: latency-svc-wnh4x
+Feb 12 10:01:19.034: INFO: Created: latency-svc-t8xzl
+Feb 12 10:01:19.058: INFO: Got endpoints: latency-svc-xj824 [750.433428ms]
+Feb 12 10:01:19.078: INFO: Created: latency-svc-vbn6q
+Feb 12 10:01:19.107: INFO: Got endpoints: latency-svc-mm48t [745.780573ms]
+Feb 12 10:01:19.120: INFO: Created: latency-svc-dnw9d
+Feb 12 10:01:19.166: INFO: Got endpoints: latency-svc-fpj6n [757.359004ms]
+Feb 12 10:01:19.185: INFO: Created: latency-svc-mkg8l
+Feb 12 10:01:19.206: INFO: Got endpoints: latency-svc-w4h4w [749.132592ms]
+Feb 12 10:01:19.220: INFO: Created: latency-svc-r2nl6
+Feb 12 10:01:19.266: INFO: Got endpoints: latency-svc-t65df [758.575063ms]
+Feb 12 10:01:19.305: INFO: Created: latency-svc-9wjzm
+Feb 12 10:01:19.309: INFO: Got endpoints: latency-svc-czk8l [751.721359ms]
+Feb 12 10:01:19.328: INFO: Created: latency-svc-89v72
+Feb 12 10:01:19.357: INFO: Got endpoints: latency-svc-jxmbf [747.914796ms]
+Feb 12 10:01:19.372: INFO: Created: latency-svc-n2nzj
+Feb 12 10:01:19.407: INFO: Got endpoints: latency-svc-n4ck7 [737.719909ms]
+Feb 12 10:01:19.424: INFO: Created: latency-svc-nq5x6
+Feb 12 10:01:19.460: INFO: Got endpoints: latency-svc-l9kj7 [751.577877ms]
+Feb 12 10:01:19.478: INFO: Created: latency-svc-qf9mm
+Feb 12 10:01:19.507: INFO: Got endpoints: latency-svc-7cs7h [739.372816ms]
+Feb 12 10:01:19.522: INFO: Created: latency-svc-nfs4m
+Feb 12 10:01:19.557: INFO: Got endpoints: latency-svc-mztzp [743.816116ms]
+Feb 12 10:01:19.574: INFO: Created: latency-svc-5wb8b
+Feb 12 10:01:19.607: INFO: Got endpoints: latency-svc-mf7rz [744.77742ms]
+Feb 12 10:01:19.638: INFO: Created: latency-svc-rh245
+Feb 12 10:01:19.658: INFO: Got endpoints: latency-svc-6f9lg [748.124775ms]
+Feb 12 10:01:19.680: INFO: Created: latency-svc-47qpx
+Feb 12 10:01:19.706: INFO: Got endpoints: latency-svc-wnh4x [724.419229ms]
+Feb 12 10:01:19.724: INFO: Created: latency-svc-wd224
+Feb 12 10:01:19.757: INFO: Got endpoints: latency-svc-t8xzl [743.96967ms]
+Feb 12 10:01:19.770: INFO: Created: latency-svc-6mzkc
+Feb 12 10:01:19.808: INFO: Got endpoints: latency-svc-vbn6q [749.491395ms]
+Feb 12 10:01:19.828: INFO: Created: latency-svc-q5gmk
+Feb 12 10:01:19.858: INFO: Got endpoints: latency-svc-dnw9d [751.636725ms]
+Feb 12 10:01:19.875: INFO: Created: latency-svc-7t9rh
+Feb 12 10:01:19.906: INFO: Got endpoints: latency-svc-mkg8l [740.270919ms]
+Feb 12 10:01:19.923: INFO: Created: latency-svc-hw85x
+Feb 12 10:01:19.956: INFO: Got endpoints: latency-svc-r2nl6 [749.642105ms]
+Feb 12 10:01:19.970: INFO: Created: latency-svc-kw4kk
+Feb 12 10:01:20.006: INFO: Got endpoints: latency-svc-9wjzm [739.301164ms]
+Feb 12 10:01:20.022: INFO: Created: latency-svc-qlsjt
+Feb 12 10:01:20.059: INFO: Got endpoints: latency-svc-89v72 [749.429354ms]
+Feb 12 10:01:20.073: INFO: Created: latency-svc-5jfrq
+Feb 12 10:01:20.106: INFO: Got endpoints: latency-svc-n2nzj [749.649589ms]
+Feb 12 10:01:20.124: INFO: Created: latency-svc-hdmvb
+Feb 12 10:01:20.171: INFO: Got endpoints: latency-svc-nq5x6 [764.786866ms]
+Feb 12 10:01:20.189: INFO: Created: latency-svc-td578
+Feb 12 10:01:20.207: INFO: Got endpoints: latency-svc-qf9mm [747.283631ms]
+Feb 12 10:01:20.220: INFO: Created: latency-svc-djcd6
+Feb 12 10:01:20.266: INFO: Got endpoints: latency-svc-nfs4m [759.522244ms]
+Feb 12 10:01:20.283: INFO: Created: latency-svc-2srcb
+Feb 12 10:01:20.307: INFO: Got endpoints: latency-svc-5wb8b [749.944619ms]
+Feb 12 10:01:20.323: INFO: Created: latency-svc-lbwmw
+Feb 12 10:01:20.363: INFO: Got endpoints: latency-svc-rh245 [755.953456ms]
+Feb 12 10:01:20.375: INFO: Created: latency-svc-cw2dq
+Feb 12 10:01:20.407: INFO: Got endpoints: latency-svc-47qpx [748.639412ms]
+Feb 12 10:01:20.422: INFO: Created: latency-svc-lkfdk
+Feb 12 10:01:20.457: INFO: Got endpoints: latency-svc-wd224 [750.681828ms]
+Feb 12 10:01:20.480: INFO: Created: latency-svc-5v5db
+Feb 12 10:01:20.506: INFO: Got endpoints: latency-svc-6mzkc [749.478463ms]
+Feb 12 10:01:20.519: INFO: Created: latency-svc-szxh5
+Feb 12 10:01:20.558: INFO: Got endpoints: latency-svc-q5gmk [750.054819ms]
+Feb 12 10:01:20.581: INFO: Created: latency-svc-mk424
+Feb 12 10:01:20.606: INFO: Got endpoints: latency-svc-7t9rh [747.471107ms]
+Feb 12 10:01:20.621: INFO: Created: latency-svc-c66zp
+Feb 12 10:01:20.658: INFO: Got endpoints: latency-svc-hw85x [752.049985ms]
+Feb 12 10:01:20.677: INFO: Created: latency-svc-8gnrr
+Feb 12 10:01:20.708: INFO: Got endpoints: latency-svc-kw4kk [752.067173ms]
+Feb 12 10:01:20.724: INFO: Created: latency-svc-hwpr8
+Feb 12 10:01:20.764: INFO: Got endpoints: latency-svc-qlsjt [758.103389ms]
+Feb 12 10:01:20.782: INFO: Created: latency-svc-87vcn
+Feb 12 10:01:20.806: INFO: Got endpoints: latency-svc-5jfrq [747.273706ms]
+Feb 12 10:01:20.824: INFO: Created: latency-svc-9cwzs
+Feb 12 10:01:20.857: INFO: Got endpoints: latency-svc-hdmvb [750.2248ms]
+Feb 12 10:01:20.876: INFO: Created: latency-svc-wj2gf
+Feb 12 10:01:20.906: INFO: Got endpoints: latency-svc-td578 [734.260134ms]
+Feb 12 10:01:20.928: INFO: Created: latency-svc-7t8br
+Feb 12 10:01:20.955: INFO: Got endpoints: latency-svc-djcd6 [747.758671ms]
+Feb 12 10:01:20.970: INFO: Created: latency-svc-ql4w5
+Feb 12 10:01:21.008: INFO: Got endpoints: latency-svc-2srcb [741.286619ms]
+Feb 12 10:01:21.025: INFO: Created: latency-svc-t8482
+Feb 12 10:01:21.061: INFO: Got endpoints: latency-svc-lbwmw [753.217309ms]
+Feb 12 10:01:21.078: INFO: Created: latency-svc-gkvtv
+Feb 12 10:01:21.105: INFO: Got endpoints: latency-svc-cw2dq [742.011439ms]
+Feb 12 10:01:21.120: INFO: Created: latency-svc-ng4kl
+Feb 12 10:01:21.173: INFO: Got endpoints: latency-svc-lkfdk [765.434339ms]
+Feb 12 10:01:21.187: INFO: Created: latency-svc-5vl7z
+Feb 12 10:01:21.206: INFO: Got endpoints: latency-svc-5v5db [748.667905ms]
+Feb 12 10:01:21.221: INFO: Created: latency-svc-vbpv6
+Feb 12 10:01:21.282: INFO: Got endpoints: latency-svc-szxh5 [775.169428ms]
+Feb 12 10:01:21.297: INFO: Created: latency-svc-vnl2w
+Feb 12 10:01:21.305: INFO: Got endpoints: latency-svc-mk424 [746.795715ms]
+Feb 12 10:01:21.321: INFO: Created: latency-svc-9ccq8
+Feb 12 10:01:21.363: INFO: Got endpoints: latency-svc-c66zp [757.505812ms]
+Feb 12 10:01:21.386: INFO: Created: latency-svc-4x97q
+Feb 12 10:01:21.405: INFO: Got endpoints: latency-svc-8gnrr [746.602472ms]
+Feb 12 10:01:21.431: INFO: Created: latency-svc-6b67z
+Feb 12 10:01:21.458: INFO: Got endpoints: latency-svc-hwpr8 [749.52378ms]
+Feb 12 10:01:21.473: INFO: Created: latency-svc-g7kg4
+Feb 12 10:01:21.507: INFO: Got endpoints: latency-svc-87vcn [743.390616ms]
+Feb 12 10:01:21.522: INFO: Created: latency-svc-kh9c2
+Feb 12 10:01:21.556: INFO: Got endpoints: latency-svc-9cwzs [750.177731ms]
+Feb 12 10:01:21.575: INFO: Created: latency-svc-475kn
+Feb 12 10:01:21.605: INFO: Got endpoints: latency-svc-wj2gf [748.230181ms]
+Feb 12 10:01:21.622: INFO: Created: latency-svc-lvlnn
+Feb 12 10:01:21.660: INFO: Got endpoints: latency-svc-7t8br [754.401359ms]
+Feb 12 10:01:21.677: INFO: Created: latency-svc-t5l5p
+Feb 12 10:01:21.704: INFO: Got endpoints: latency-svc-ql4w5 [749.392468ms]
+Feb 12 10:01:21.720: INFO: Created: latency-svc-k9w9b
+Feb 12 10:01:21.757: INFO: Got endpoints: latency-svc-t8482 [749.414118ms]
+Feb 12 10:01:21.774: INFO: Created: latency-svc-6hb8q
+Feb 12 10:01:21.807: INFO: Got endpoints: latency-svc-gkvtv [745.843193ms]
+Feb 12 10:01:21.820: INFO: Created: latency-svc-lj8dd
+Feb 12 10:01:21.879: INFO: Got endpoints: latency-svc-ng4kl [773.868164ms]
+Feb 12 10:01:21.893: INFO: Created: latency-svc-7gg57
+Feb 12 10:01:21.905: INFO: Got endpoints: latency-svc-5vl7z [732.109785ms]
+Feb 12 10:01:21.920: INFO: Created: latency-svc-nd2gp
+Feb 12 10:01:21.960: INFO: Got endpoints: latency-svc-vbpv6 [753.866388ms]
+Feb 12 10:01:21.988: INFO: Created: latency-svc-x7ftb
+Feb 12 10:01:22.005: INFO: Got endpoints: latency-svc-vnl2w [723.112751ms]
+Feb 12 10:01:22.024: INFO: Created: latency-svc-pbv95
+Feb 12 10:01:22.063: INFO: Got endpoints: latency-svc-9ccq8 [758.340866ms]
+Feb 12 10:01:22.082: INFO: Created: latency-svc-jl5cf
+Feb 12 10:01:22.105: INFO: Got endpoints: latency-svc-4x97q [740.608468ms]
+Feb 12 10:01:22.120: INFO: Created: latency-svc-h6k2h
+Feb 12 10:01:22.160: INFO: Got endpoints: latency-svc-6b67z [755.115936ms]
+Feb 12 10:01:22.181: INFO: Created: latency-svc-45kh9
+Feb 12 10:01:22.208: INFO: Got endpoints: latency-svc-g7kg4 [748.476449ms]
+Feb 12 10:01:22.233: INFO: Created: latency-svc-v4rvm
+Feb 12 10:01:22.257: INFO: Got endpoints: latency-svc-kh9c2 [749.466078ms]
+Feb 12 10:01:22.272: INFO: Created: latency-svc-djf4n
+Feb 12 10:01:22.305: INFO: Got endpoints: latency-svc-475kn [747.856952ms]
+Feb 12 10:01:22.320: INFO: Created: latency-svc-m2pmp
+Feb 12 10:01:22.354: INFO: Got endpoints: latency-svc-lvlnn [748.900051ms]
+Feb 12 10:01:22.370: INFO: Created: latency-svc-wk2fr
+Feb 12 10:01:22.410: INFO: Got endpoints: latency-svc-t5l5p [749.340395ms]
+Feb 12 10:01:22.447: INFO: Created: latency-svc-wcr76
+Feb 12 10:01:22.460: INFO: Got endpoints: latency-svc-k9w9b [754.942307ms]
+Feb 12 10:01:22.474: INFO: Created: latency-svc-zzzsh
+Feb 12 10:01:22.504: INFO: Got endpoints: latency-svc-6hb8q [746.926838ms]
+Feb 12 10:01:22.518: INFO: Created: latency-svc-mp28h
+Feb 12 10:01:22.563: INFO: Got endpoints: latency-svc-lj8dd [756.050904ms]
+Feb 12 10:01:22.582: INFO: Created: latency-svc-74pb7
+Feb 12 10:01:22.607: INFO: Got endpoints: latency-svc-7gg57 [727.733809ms]
+Feb 12 10:01:22.620: INFO: Created: latency-svc-rpwd7
+Feb 12 10:01:22.663: INFO: Got endpoints: latency-svc-nd2gp [757.700032ms]
+Feb 12 10:01:22.683: INFO: Created: latency-svc-s9lqk
+Feb 12 10:01:22.705: INFO: Got endpoints: latency-svc-x7ftb [742.502802ms]
+Feb 12 10:01:22.722: INFO: Created: latency-svc-z9w22
+Feb 12 10:01:22.760: INFO: Got endpoints: latency-svc-pbv95 [755.580621ms]
+Feb 12 10:01:22.780: INFO: Created: latency-svc-d2trw
+Feb 12 10:01:22.806: INFO: Got endpoints: latency-svc-jl5cf [743.097702ms]
+Feb 12 10:01:22.829: INFO: Created: latency-svc-vqlf2
+Feb 12 10:01:22.860: INFO: Got endpoints: latency-svc-h6k2h [755.665572ms]
+Feb 12 10:01:22.887: INFO: Created: latency-svc-tmlt7
+Feb 12 10:01:22.905: INFO: Got endpoints: latency-svc-45kh9 [744.461503ms]
+Feb 12 10:01:22.917: INFO: Created: latency-svc-wxxpm
+Feb 12 10:01:22.958: INFO: Got endpoints: latency-svc-v4rvm [750.800945ms]
+Feb 12 10:01:22.979: INFO: Created: latency-svc-qhjnl
+Feb 12 10:01:23.003: INFO: Got endpoints: latency-svc-djf4n [746.400422ms]
+Feb 12 10:01:23.017: INFO: Created: latency-svc-d9bl2
+Feb 12 10:01:23.061: INFO: Got endpoints: latency-svc-m2pmp [756.190537ms]
+Feb 12 10:01:23.080: INFO: Created: latency-svc-wfwf2
+Feb 12 10:01:23.106: INFO: Got endpoints: latency-svc-wk2fr [751.431573ms]
+Feb 12 10:01:23.120: INFO: Created: latency-svc-4dkjg
+Feb 12 10:01:23.157: INFO: Got endpoints: latency-svc-wcr76 [747.314991ms]
+Feb 12 10:01:23.187: INFO: Created: latency-svc-sf9cp
+Feb 12 10:01:23.205: INFO: Got endpoints: latency-svc-zzzsh [745.735993ms]
+Feb 12 10:01:23.221: INFO: Created: latency-svc-729jk
+Feb 12 10:01:23.278: INFO: Got endpoints: latency-svc-mp28h [773.850486ms]
+Feb 12 10:01:23.297: INFO: Created: latency-svc-xhp2t
+Feb 12 10:01:23.303: INFO: Got endpoints: latency-svc-74pb7 [740.422472ms]
+Feb 12 10:01:23.318: INFO: Created: latency-svc-678bd
+Feb 12 10:01:23.361: INFO: Got endpoints: latency-svc-rpwd7 [754.127657ms]
+Feb 12 10:01:23.382: INFO: Created: latency-svc-w4tls
+Feb 12 10:01:23.403: INFO: Got endpoints: latency-svc-s9lqk [740.889573ms]
+Feb 12 10:01:23.420: INFO: Created: latency-svc-wrf7r
+Feb 12 10:01:23.461: INFO: Got endpoints: latency-svc-z9w22 [756.460688ms]
+Feb 12 10:01:23.502: INFO: Created: latency-svc-jz94t
+Feb 12 10:01:23.504: INFO: Got endpoints: latency-svc-d2trw [743.621996ms]
+Feb 12 10:01:23.517: INFO: Created: latency-svc-mvs2w
+Feb 12 10:01:23.561: INFO: Got endpoints: latency-svc-vqlf2 [754.449202ms]
+Feb 12 10:01:23.594: INFO: Created: latency-svc-9sh5z
+Feb 12 10:01:23.603: INFO: Got endpoints: latency-svc-tmlt7 [742.883423ms]
+Feb 12 10:01:23.619: INFO: Created: latency-svc-7wc5h
+Feb 12 10:01:23.655: INFO: Got endpoints: latency-svc-wxxpm [750.279278ms]
+Feb 12 10:01:23.669: INFO: Created: latency-svc-dsmwt
+Feb 12 10:01:23.705: INFO: Got endpoints: latency-svc-qhjnl [746.189002ms]
+Feb 12 10:01:23.720: INFO: Created: latency-svc-d2d99
+Feb 12 10:01:23.780: INFO: Got endpoints: latency-svc-d9bl2 [776.245239ms]
+Feb 12 10:01:23.812: INFO: Got endpoints: latency-svc-wfwf2 [750.927771ms]
+Feb 12 10:01:23.813: INFO: Created: latency-svc-c6kjz
+Feb 12 10:01:23.828: INFO: Created: latency-svc-xthrp
+Feb 12 10:01:23.890: INFO: Got endpoints: latency-svc-4dkjg [784.515832ms]
+Feb 12 10:01:23.905: INFO: Got endpoints: latency-svc-sf9cp [747.985472ms]
+Feb 12 10:01:23.907: INFO: Created: latency-svc-wfpkb
+Feb 12 10:01:23.921: INFO: Created: latency-svc-bc4cz
+Feb 12 10:01:23.965: INFO: Got endpoints: latency-svc-729jk [759.09196ms]
+Feb 12 10:01:24.004: INFO: Created: latency-svc-hrnqq
+Feb 12 10:01:24.006: INFO: Got endpoints: latency-svc-xhp2t [727.577985ms]
+Feb 12 10:01:24.028: INFO: Created: latency-svc-7786x
+Feb 12 10:01:24.065: INFO: Got endpoints: latency-svc-678bd [761.400291ms]
+Feb 12 10:01:24.086: INFO: Created: latency-svc-86x2l
+Feb 12 10:01:24.103: INFO: Got endpoints: latency-svc-w4tls [741.994891ms]
+Feb 12 10:01:24.118: INFO: Created: latency-svc-fprhm
+Feb 12 10:01:24.183: INFO: Got endpoints: latency-svc-wrf7r [779.014381ms]
+Feb 12 10:01:24.205: INFO: Got endpoints: latency-svc-jz94t [743.804533ms]
+Feb 12 10:01:24.206: INFO: Created: latency-svc-b5hth
+Feb 12 10:01:24.221: INFO: Created: latency-svc-tx77d
+Feb 12 10:01:24.263: INFO: Got endpoints: latency-svc-mvs2w [758.725321ms]
+Feb 12 10:01:24.304: INFO: Got endpoints: latency-svc-9sh5z [743.4011ms]
+Feb 12 10:01:24.358: INFO: Got endpoints: latency-svc-7wc5h [754.244473ms]
+Feb 12 10:01:24.403: INFO: Got endpoints: latency-svc-dsmwt [748.131562ms]
+Feb 12 10:01:24.458: INFO: Got endpoints: latency-svc-d2d99 [752.926578ms]
+Feb 12 10:01:24.506: INFO: Got endpoints: latency-svc-c6kjz [726.330541ms]
+Feb 12 10:01:24.558: INFO: Got endpoints: latency-svc-xthrp [745.85341ms]
+Feb 12 10:01:24.605: INFO: Got endpoints: latency-svc-wfpkb [714.859889ms]
+Feb 12 10:01:24.654: INFO: Got endpoints: latency-svc-bc4cz [749.337441ms]
+Feb 12 10:01:24.703: INFO: Got endpoints: latency-svc-hrnqq [738.115036ms]
+Feb 12 10:01:24.753: INFO: Got endpoints: latency-svc-7786x [747.241034ms]
+Feb 12 10:01:24.804: INFO: Got endpoints: latency-svc-86x2l [739.103438ms]
+Feb 12 10:01:24.854: INFO: Got endpoints: latency-svc-fprhm [750.601406ms]
+Feb 12 10:01:24.904: INFO: Got endpoints: latency-svc-b5hth [721.316609ms]
+Feb 12 10:01:24.953: INFO: Got endpoints: latency-svc-tx77d [747.477642ms]
+Feb 12 10:01:24.953: INFO: Latencies: [34.262356ms 56.340677ms 65.706771ms 71.284207ms 80.336984ms 95.200899ms 101.40848ms 115.156862ms 138.07277ms 154.673879ms 159.765072ms 159.85863ms 165.545937ms 165.891996ms 167.627687ms 171.816617ms 173.877715ms 174.216326ms 177.488871ms 179.406789ms 179.846986ms 179.887899ms 181.409711ms 181.841958ms 183.196079ms 187.444769ms 188.361973ms 191.190222ms 193.655372ms 194.988195ms 195.604563ms 195.928223ms 196.396308ms 197.143884ms 198.763389ms 199.134311ms 209.894152ms 218.129055ms 221.050463ms 235.76627ms 248.761994ms 286.004ms 323.941366ms 362.780374ms 401.133673ms 444.76335ms 464.444243ms 511.205968ms 542.486863ms 580.504494ms 611.926688ms 645.707222ms 685.586515ms 714.859889ms 721.316609ms 723.112751ms 724.419229ms 724.566477ms 726.330541ms 727.577985ms 727.733809ms 732.109785ms 734.260134ms 735.881376ms 737.719909ms 738.115036ms 739.103438ms 739.301164ms 739.372816ms 740.270919ms 740.422472ms 740.608468ms 740.889573ms 741.286619ms 741.994891ms 742.011439ms 742.502802ms 742.883423ms 743.097702ms 743.390616ms 743.4011ms 743.621996ms 743.743849ms 743.804533ms 743.816116ms 743.96967ms 744.428009ms 744.461503ms 744.77742ms 745.735993ms 745.780573ms 745.843193ms 745.85341ms 745.886725ms 746.189002ms 746.400422ms 746.602472ms 746.654595ms 746.795715ms 746.926838ms 747.241034ms 747.273706ms 747.283631ms 747.300153ms 747.314991ms 747.471107ms 747.477642ms 747.758671ms 747.856952ms 747.914796ms 747.985472ms 748.068551ms 748.124775ms 748.131562ms 748.230181ms 748.405067ms 748.476449ms 748.558501ms 748.639412ms 748.644845ms 748.667905ms 748.765592ms 748.900051ms 749.132592ms 749.235726ms 749.337441ms 749.340395ms 749.392468ms 749.414118ms 749.429354ms 749.466078ms 749.478463ms 749.491395ms 749.52378ms 749.526151ms 749.620357ms 749.642105ms 749.649589ms 749.875783ms 749.944619ms 750.054819ms 750.177731ms 750.2248ms 750.279278ms 750.348502ms 750.433428ms 750.469081ms 750.601406ms 750.681828ms 750.800945ms 750.920833ms 750.927771ms 750.969601ms 751.286034ms 751.431573ms 751.577877ms 751.636725ms 751.721359ms 752.049985ms 752.067173ms 752.639311ms 752.926578ms 753.217309ms 753.583203ms 753.866388ms 754.127657ms 754.220729ms 754.244473ms 754.401359ms 754.449202ms 754.942307ms 755.115936ms 755.507746ms 755.580621ms 755.665572ms 755.953456ms 756.050904ms 756.190537ms 756.460688ms 757.359004ms 757.505812ms 757.700032ms 758.103389ms 758.117611ms 758.340866ms 758.575063ms 758.725321ms 759.09196ms 759.522244ms 760.61899ms 761.400291ms 764.786866ms 765.434339ms 773.504893ms 773.850486ms 773.868164ms 775.169428ms 776.245239ms 779.014381ms 784.515832ms]
+Feb 12 10:01:24.953: INFO: 50 %ile: 747.241034ms
+Feb 12 10:01:24.953: INFO: 90 %ile: 757.505812ms
+Feb 12 10:01:24.953: INFO: 99 %ile: 779.014381ms
+Feb 12 10:01:24.953: INFO: Total sample count: 200
+[AfterEach] [sig-network] Service endpoints latency
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:01:24.953: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "svc-latency-2063" for this suite.
 
-• [SLOW TEST:95.480 seconds]
-[sig-storage] ConfigMap
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:33
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:11.933 seconds]
+[sig-network] Service endpoints latency
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should not be very high  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":35,"skipped":675,"failed":0}
-SSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-network] Service endpoints latency should not be very high  [Conformance]","total":280,"completed":33,"skipped":457,"failed":0}
+SSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Projected downwardAPI 
+  should update labels on modification [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:48:27.369: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-6694
+Feb 12 10:01:24.974: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7417
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir volume type on tmpfs
-Aug 30 16:48:27.548: INFO: Waiting up to 5m0s for pod "pod-d441700d-ec20-4ffb-933a-0a927c2b57bf" in namespace "emptydir-6694" to be "success or failure"
-Aug 30 16:48:27.559: INFO: Pod "pod-d441700d-ec20-4ffb-933a-0a927c2b57bf": Phase="Pending", Reason="", readiness=false. Elapsed: 10.626048ms
-Aug 30 16:48:29.565: INFO: Pod "pod-d441700d-ec20-4ffb-933a-0a927c2b57bf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016472409s
-STEP: Saw pod success
-Aug 30 16:48:29.565: INFO: Pod "pod-d441700d-ec20-4ffb-933a-0a927c2b57bf" satisfied condition "success or failure"
-Aug 30 16:48:29.569: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod pod-d441700d-ec20-4ffb-933a-0a927c2b57bf container test-container: 
-STEP: delete the pod
-Aug 30 16:48:29.679: INFO: Waiting for pod pod-d441700d-ec20-4ffb-933a-0a927c2b57bf to disappear
-Aug 30 16:48:29.683: INFO: Pod pod-d441700d-ec20-4ffb-933a-0a927c2b57bf no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:48:29.683: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-6694" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":36,"skipped":696,"failed":0}
-SSSSSSSSSSSSSSSSSSS
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should update labels on modification [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating the pod
+Feb 12 10:01:27.715: INFO: Successfully updated pod "labelsupdateb5340d67-292c-412d-aa3c-9857c1e41aa4"
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:01:29.753: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-7417" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance]","total":280,"completed":34,"skipped":465,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-node] Downward API 
-  should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] Downward API
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] ResourceQuota 
+  should create a ResourceQuota and capture the life of a pod. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:48:29.699: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-634
+Feb 12 10:01:29.771: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename resourcequota
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-1932
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward api env vars
-Aug 30 16:48:29.868: INFO: Waiting up to 5m0s for pod "downward-api-543e9512-0410-4819-997b-3036c613f168" in namespace "downward-api-634" to be "success or failure"
-Aug 30 16:48:29.873: INFO: Pod "downward-api-543e9512-0410-4819-997b-3036c613f168": Phase="Pending", Reason="", readiness=false. Elapsed: 4.650776ms
-Aug 30 16:48:31.878: INFO: Pod "downward-api-543e9512-0410-4819-997b-3036c613f168": Phase="Pending", Reason="", readiness=false. Elapsed: 2.009927127s
-Aug 30 16:48:33.885: INFO: Pod "downward-api-543e9512-0410-4819-997b-3036c613f168": Phase="Pending", Reason="", readiness=false. Elapsed: 4.016495012s
-Aug 30 16:48:35.890: INFO: Pod "downward-api-543e9512-0410-4819-997b-3036c613f168": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.022259619s
-STEP: Saw pod success
-Aug 30 16:48:35.890: INFO: Pod "downward-api-543e9512-0410-4819-997b-3036c613f168" satisfied condition "success or failure"
-Aug 30 16:48:35.895: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod downward-api-543e9512-0410-4819-997b-3036c613f168 container dapi-container: 
-STEP: delete the pod
-Aug 30 16:48:35.978: INFO: Waiting for pod downward-api-543e9512-0410-4819-997b-3036c613f168 to disappear
-Aug 30 16:48:35.988: INFO: Pod downward-api-543e9512-0410-4819-997b-3036c613f168 no longer exists
-[AfterEach] [sig-node] Downward API
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:48:35.988: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-634" for this suite.
+[It] should create a ResourceQuota and capture the life of a pod. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Counting existing ResourceQuota
+STEP: Creating a ResourceQuota
+STEP: Ensuring resource quota status is calculated
+STEP: Creating a Pod that fits quota
+STEP: Ensuring ResourceQuota status captures the pod usage
+STEP: Not allowing a pod to be created that exceeds remaining quota
+STEP: Not allowing a pod to be created that exceeds remaining quota(validation on extended resources)
+STEP: Ensuring a pod cannot update its resource requirements
+STEP: Ensuring attempts to update pod resource requirements did not change quota usage
+STEP: Deleting the pod
+STEP: Ensuring resource quota status released the pod usage
+[AfterEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:01:43.049: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-1932" for this suite.
 
-• [SLOW TEST:6.303 seconds]
-[sig-node] Downward API
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:33
-  should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]","total":280,"completed":37,"skipped":715,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSS
+• [SLOW TEST:13.295 seconds]
+[sig-api-machinery] ResourceQuota
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should create a ResourceQuota and capture the life of a pod. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance]","total":280,"completed":35,"skipped":489,"failed":0}
+[sig-network] Services 
+  should be able to change the type from ExternalName to NodePort [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:48:36.002: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-7824
+Feb 12 10:01:43.067: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename services
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-3859
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Aug 30 16:48:36.544: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Aug 30 16:48:38.560: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402916, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402916, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402916, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402916, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Aug 30 16:48:41.580: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering a validating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API
-Aug 30 16:48:41.610: INFO: Waiting for webhook configuration to be ready...
-STEP: Registering a mutating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API
-STEP: Creating a dummy validating-webhook-configuration object
-STEP: Deleting the validating-webhook-configuration, which should be possible to remove
-STEP: Creating a dummy mutating-webhook-configuration object
-STEP: Deleting the mutating-webhook-configuration, which should be possible to remove
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:48:41.980: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-7824" for this suite.
-STEP: Destroying namespace "webhook-7824-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
+[It] should be able to change the type from ExternalName to NodePort [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating a service externalname-service with the type=ExternalName in namespace services-3859
+STEP: changing the ExternalName service to type=NodePort
+STEP: creating replication controller externalname-service in namespace services-3859
+I0212 10:01:43.351100      21 runners.go:189] Created replication controller with name: externalname-service, namespace: services-3859, replica count: 2
+I0212 10:01:46.401506      21 runners.go:189] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+Feb 12 10:01:46.401: INFO: Creating new exec pod
+Feb 12 10:01:49.431: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=services-3859 execpod6hq4q -- /bin/sh -x -c nc -zv -t -w 2 externalname-service 80'
+Feb 12 10:01:49.999: INFO: stderr: "+ nc -zv -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n"
+Feb 12 10:01:49.999: INFO: stdout: ""
+Feb 12 10:01:50.002: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=services-3859 execpod6hq4q -- /bin/sh -x -c nc -zv -t -w 2 10.240.19.89 80'
+Feb 12 10:01:50.574: INFO: stderr: "+ nc -zv -t -w 2 10.240.19.89 80\nConnection to 10.240.19.89 80 port [tcp/http] succeeded!\n"
+Feb 12 10:01:50.574: INFO: stdout: ""
+Feb 12 10:01:50.574: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=services-3859 execpod6hq4q -- /bin/sh -x -c nc -zv -t -w 2 139.59.138.127 32301'
+Feb 12 10:01:51.133: INFO: stderr: "+ nc -zv -t -w 2 139.59.138.127 32301\nConnection to 139.59.138.127 32301 port [tcp/32301] succeeded!\n"
+Feb 12 10:01:51.133: INFO: stdout: ""
+Feb 12 10:01:51.133: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=services-3859 execpod6hq4q -- /bin/sh -x -c nc -zv -t -w 2 104.248.134.247 32301'
+Feb 12 10:01:51.654: INFO: stderr: "+ nc -zv -t -w 2 104.248.134.247 32301\nConnection to 104.248.134.247 32301 port [tcp/32301] succeeded!\n"
+Feb 12 10:01:51.654: INFO: stdout: ""
+Feb 12 10:01:51.654: INFO: Cleaning up the ExternalName to NodePort test service
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:01:51.695: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-3859" for this suite.
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
 
-• [SLOW TEST:6.075 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:8.647 seconds]
+[sig-network] Services
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should be able to change the type from ExternalName to NodePort [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]","total":280,"completed":38,"skipped":739,"failed":0}
-SS
+{"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance]","total":280,"completed":36,"skipped":489,"failed":0}
+SSS
 ------------------------------
-[sig-apps] Deployment 
-  deployment should delete old replica sets [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] Watchers 
+  should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Watchers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:48:42.080: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename deployment
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-4729
+Feb 12 10:01:51.714: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename watch
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-7976
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
-[It] deployment should delete old replica sets [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 16:48:42.251: INFO: Pod name cleanup-pod: Found 0 pods out of 1
-Aug 30 16:48:47.257: INFO: Pod name cleanup-pod: Found 1 pods out of 1
-STEP: ensuring each pod is running
-Aug 30 16:48:47.257: INFO: Creating deployment test-cleanup-deployment
-STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
-Aug 30 16:48:47.288: INFO: Deployment "test-cleanup-deployment":
-&Deployment{ObjectMeta:{test-cleanup-deployment  deployment-4729 /apis/apps/v1/namespaces/deployment-4729/deployments/test-cleanup-deployment fe6bc564-b116-474e-8ace-b9bb3ccd4d28 7750 1 2020-08-30 16:48:47 +0000 UTC   map[name:cleanup-pod] map[] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:cleanup-pod] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc002827678  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[]DeploymentCondition{},ReadyReplicas:0,CollisionCount:nil,},}
-
-Aug 30 16:48:47.294: INFO: New ReplicaSet "test-cleanup-deployment-55ffc6b7b6" of Deployment "test-cleanup-deployment":
-&ReplicaSet{ObjectMeta:{test-cleanup-deployment-55ffc6b7b6  deployment-4729 /apis/apps/v1/namespaces/deployment-4729/replicasets/test-cleanup-deployment-55ffc6b7b6 97db64fd-daeb-42d8-87c2-f835b7a9bcfa 7752 1 2020-08-30 16:48:47 +0000 UTC   map[name:cleanup-pod pod-template-hash:55ffc6b7b6] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-cleanup-deployment fe6bc564-b116-474e-8ace-b9bb3ccd4d28 0xc002827a97 0xc002827a98}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod-template-hash: 55ffc6b7b6,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:cleanup-pod pod-template-hash:55ffc6b7b6] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc002827b18  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:0,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
-Aug 30 16:48:47.294: INFO: All old ReplicaSets of Deployment "test-cleanup-deployment":
-Aug 30 16:48:47.294: INFO: &ReplicaSet{ObjectMeta:{test-cleanup-controller  deployment-4729 /apis/apps/v1/namespaces/deployment-4729/replicasets/test-cleanup-controller 69c539d3-9a11-4dda-a8a4-2d14706b2924 7751 1 2020-08-30 16:48:42 +0000 UTC   map[name:cleanup-pod pod:httpd] map[] [{apps/v1 Deployment test-cleanup-deployment fe6bc564-b116-474e-8ace-b9bb3ccd4d28 0xc0028279d7 0xc0028279d8}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:cleanup-pod pod:httpd] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc002827a38  ClusterFirst map[]     false false false  PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},}
-Aug 30 16:48:47.302: INFO: Pod "test-cleanup-controller-h74rt" is available:
-&Pod{ObjectMeta:{test-cleanup-controller-h74rt test-cleanup-controller- deployment-4729 /api/v1/namespaces/deployment-4729/pods/test-cleanup-controller-h74rt 883cfb7c-577d-4eff-94d0-6d249565be11 7721 0 2020-08-30 16:48:42 +0000 UTC   map[name:cleanup-pod pod:httpd] map[cni.projectcalico.org/podIP:172.25.0.46/32] [{apps/v1 ReplicaSet test-cleanup-controller 69c539d3-9a11-4dda-a8a4-2d14706b2924 0xc002ab61cf 0xc002ab61f0}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-x26qx,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-x26qx,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-x26qx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:48:42 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:48:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:48:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 16:48:42 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:172.25.0.46,StartTime:2020-08-30 16:48:42 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-08-30 16:48:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://6f88b8c52d91ed05a581934074df975718614b90410e86b8f3c0d3ba7cf5d373,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.0.46,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 16:48:47.302: INFO: Pod "test-cleanup-deployment-55ffc6b7b6-zn4r8" is not available:
-&Pod{ObjectMeta:{test-cleanup-deployment-55ffc6b7b6-zn4r8 test-cleanup-deployment-55ffc6b7b6- deployment-4729 /api/v1/namespaces/deployment-4729/pods/test-cleanup-deployment-55ffc6b7b6-zn4r8 d05e5e8e-86be-44e3-a827-3a806f3a90c8 7754 0 2020-08-30 16:48:47 +0000 UTC   map[name:cleanup-pod pod-template-hash:55ffc6b7b6] map[] [{apps/v1 ReplicaSet test-cleanup-deployment-55ffc6b7b6 97db64fd-daeb-42d8-87c2-f835b7a9bcfa 0xc002ab6387 0xc002ab6388}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-x26qx,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-x26qx,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-x26qx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:48:47.302: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "deployment-4729" for this suite.
+[It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating a watch on configmaps with a certain label
+STEP: creating a new configmap
+STEP: modifying the configmap once
+STEP: changing the label value of the configmap
+STEP: Expecting to observe a delete notification for the watched object
+Feb 12 10:01:51.997: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-7976 /api/v1/namespaces/watch-7976/configmaps/e2e-watch-test-label-changed e1f0a0c4-a24d-4b64-b6ab-367b135f7175 8654 0 2021-02-12 10:01:51 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Feb 12 10:01:51.997: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-7976 /api/v1/namespaces/watch-7976/configmaps/e2e-watch-test-label-changed e1f0a0c4-a24d-4b64-b6ab-367b135f7175 8655 0 2021-02-12 10:01:51 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+Feb 12 10:01:51.997: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-7976 /api/v1/namespaces/watch-7976/configmaps/e2e-watch-test-label-changed e1f0a0c4-a24d-4b64-b6ab-367b135f7175 8656 0 2021-02-12 10:01:51 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+STEP: modifying the configmap a second time
+STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements
+STEP: changing the label value of the configmap back
+STEP: modifying the configmap a third time
+STEP: deleting the configmap
+STEP: Expecting to observe an add notification for the watched object when the label value was restored
+Feb 12 10:02:02.083: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-7976 /api/v1/namespaces/watch-7976/configmaps/e2e-watch-test-label-changed e1f0a0c4-a24d-4b64-b6ab-367b135f7175 8735 0 2021-02-12 10:01:51 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Feb 12 10:02:02.083: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-7976 /api/v1/namespaces/watch-7976/configmaps/e2e-watch-test-label-changed e1f0a0c4-a24d-4b64-b6ab-367b135f7175 8736 0 2021-02-12 10:01:51 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
+Feb 12 10:02:02.083: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-7976 /api/v1/namespaces/watch-7976/configmaps/e2e-watch-test-label-changed e1f0a0c4-a24d-4b64-b6ab-367b135f7175 8737 0 2021-02-12 10:01:51 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:02:02.084: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-7976" for this suite.
 
-• [SLOW TEST:5.245 seconds]
-[sig-apps] Deployment
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  deployment should delete old replica sets [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:10.388 seconds]
+[sig-api-machinery] Watchers
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] Deployment deployment should delete old replica sets [Conformance]","total":280,"completed":39,"skipped":741,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance]","total":280,"completed":37,"skipped":492,"failed":0}
+SSSS
 ------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Docker Containers 
+  should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Docker Containers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:48:47.327: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7104
+Feb 12 10:02:02.103: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename containers
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-2853
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating projection with secret that has name projected-secret-test-e174f66e-dc2d-4809-b01d-87718e87c79a
-STEP: Creating a pod to test consume secrets
-Aug 30 16:48:47.513: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-48126f7d-46f3-4b8f-9c56-4321a8ea34f1" in namespace "projected-7104" to be "success or failure"
-Aug 30 16:48:47.518: INFO: Pod "pod-projected-secrets-48126f7d-46f3-4b8f-9c56-4321a8ea34f1": Phase="Pending", Reason="", readiness=false. Elapsed: 4.213617ms
-Aug 30 16:48:49.524: INFO: Pod "pod-projected-secrets-48126f7d-46f3-4b8f-9c56-4321a8ea34f1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010572817s
+[It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test override command
+Feb 12 10:02:02.293: INFO: Waiting up to 5m0s for pod "client-containers-7f8f1923-6ee9-4e65-9c90-743faf1d1ab0" in namespace "containers-2853" to be "success or failure"
+Feb 12 10:02:02.302: INFO: Pod "client-containers-7f8f1923-6ee9-4e65-9c90-743faf1d1ab0": Phase="Pending", Reason="", readiness=false. Elapsed: 9.066149ms
+Feb 12 10:02:04.310: INFO: Pod "client-containers-7f8f1923-6ee9-4e65-9c90-743faf1d1ab0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016733945s
+Feb 12 10:02:06.317: INFO: Pod "client-containers-7f8f1923-6ee9-4e65-9c90-743faf1d1ab0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024203643s
 STEP: Saw pod success
-Aug 30 16:48:49.524: INFO: Pod "pod-projected-secrets-48126f7d-46f3-4b8f-9c56-4321a8ea34f1" satisfied condition "success or failure"
-Aug 30 16:48:49.528: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod pod-projected-secrets-48126f7d-46f3-4b8f-9c56-4321a8ea34f1 container projected-secret-volume-test: 
+Feb 12 10:02:06.317: INFO: Pod "client-containers-7f8f1923-6ee9-4e65-9c90-743faf1d1ab0" satisfied condition "success or failure"
+Feb 12 10:02:06.328: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod client-containers-7f8f1923-6ee9-4e65-9c90-743faf1d1ab0 container test-container: 
 STEP: delete the pod
-Aug 30 16:48:49.596: INFO: Waiting for pod pod-projected-secrets-48126f7d-46f3-4b8f-9c56-4321a8ea34f1 to disappear
-Aug 30 16:48:49.601: INFO: Pod pod-projected-secrets-48126f7d-46f3-4b8f-9c56-4321a8ea34f1 no longer exists
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:48:49.601: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-7104" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":40,"skipped":763,"failed":0}
-SSSSSSS
+Feb 12 10:02:06.450: INFO: Waiting for pod client-containers-7f8f1923-6ee9-4e65-9c90-743faf1d1ab0 to disappear
+Feb 12 10:02:06.468: INFO: Pod client-containers-7f8f1923-6ee9-4e65-9c90-743faf1d1ab0 no longer exists
+[AfterEach] [k8s.io] Docker Containers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:02:06.468: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "containers-2853" for this suite.
+•{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]","total":280,"completed":38,"skipped":496,"failed":0}
+SSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-cli] Kubectl client Kubectl run rc 
+  should create an rc from an image [Deprecated] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:48:49.620: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename sched-pred
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-9339
+Feb 12 10:02:06.487: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-1162
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
-Aug 30 16:48:49.778: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-Aug 30 16:48:49.793: INFO: Waiting for terminating namespaces to be deleted...
-Aug 30 16:48:49.798: INFO: 
-Logging pods the kubelet thinks is on node adoring-wozniak-54dcfd79fc-6rshr before test
-Aug 30 16:48:49.855: INFO: kube-proxy-wxdxv from kube-system started at 2020-08-30 16:28:14 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.855: INFO: 	Container kube-proxy ready: true, restart count 0
-Aug 30 16:48:49.855: INFO: node-local-dns-cvxnf from kube-system started at 2020-08-30 16:28:34 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.855: INFO: 	Container node-cache ready: true, restart count 0
-Aug 30 16:48:49.855: INFO: logrotate-57bmz from kube-system started at 2020-08-30 16:28:34 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.855: INFO: 	Container logrotate ready: true, restart count 0
-Aug 30 16:48:49.855: INFO: sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-nd5nc from sonobuoy started at 2020-08-30 16:29:24 +0000 UTC (2 container statuses recorded)
-Aug 30 16:48:49.855: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 16:48:49.855: INFO: 	Container systemd-logs ready: true, restart count 0
-Aug 30 16:48:49.856: INFO: coredns-54457d966b-fbnz4 from kube-system started at 2020-08-30 16:28:39 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.856: INFO: 	Container coredns ready: true, restart count 0
-Aug 30 16:48:49.856: INFO: sonobuoy-e2e-job-cf49606f646f4c8a from sonobuoy started at 2020-08-30 16:29:23 +0000 UTC (2 container statuses recorded)
-Aug 30 16:48:49.856: INFO: 	Container e2e ready: true, restart count 0
-Aug 30 16:48:49.856: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 16:48:49.856: INFO: canal-54glj from kube-system started at 2020-08-30 16:28:14 +0000 UTC (2 container statuses recorded)
-Aug 30 16:48:49.856: INFO: 	Container calico-node ready: true, restart count 0
-Aug 30 16:48:49.856: INFO: 	Container kube-flannel ready: true, restart count 0
-Aug 30 16:48:49.856: INFO: user-ssh-keys-agent-xwrzj from kube-system started at 2020-08-30 16:28:14 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.856: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
-Aug 30 16:48:49.856: INFO: sonobuoy from sonobuoy started at 2020-08-30 16:29:18 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.856: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-Aug 30 16:48:49.856: INFO: 
-Logging pods the kubelet thinks is on node adoring-wozniak-54dcfd79fc-948mf before test
-Aug 30 16:48:49.916: INFO: node-local-dns-wrs2k from kube-system started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.916: INFO: 	Container node-cache ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: test-cleanup-deployment-55ffc6b7b6-zn4r8 from deployment-4729 started at 2020-08-30 16:48:47 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.916: INFO: 	Container agnhost ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: dashboard-metrics-scraper-59bfc65dc9-94mm2 from kubernetes-dashboard started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.916: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: kube-proxy-v88gx from kube-system started at 2020-08-30 16:28:09 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.916: INFO: 	Container kube-proxy ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: user-ssh-keys-agent-vkbs9 from kube-system started at 2020-08-30 16:28:09 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.916: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: openvpn-client-78d595f58b-pht8r from kube-system started at 2020-08-30 16:28:29 +0000 UTC (2 container statuses recorded)
-Aug 30 16:48:49.916: INFO: 	Container dnat-controller ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: 	Container openvpn-client ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: dashboard-metrics-scraper-59bfc65dc9-fhfrl from kubernetes-dashboard started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.916: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: logrotate-pw2ff from kube-system started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.916: INFO: 	Container logrotate ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-cdm8v from sonobuoy started at 2020-08-30 16:29:23 +0000 UTC (2 container statuses recorded)
-Aug 30 16:48:49.916: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: 	Container systemd-logs ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: canal-lg5hn from kube-system started at 2020-08-30 16:28:09 +0000 UTC (2 container statuses recorded)
-Aug 30 16:48:49.916: INFO: 	Container calico-node ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: 	Container kube-flannel ready: true, restart count 0
-Aug 30 16:48:49.916: INFO: coredns-54457d966b-vr27r from kube-system started at 2020-08-30 16:28:39 +0000 UTC (1 container statuses recorded)
-Aug 30 16:48:49.916: INFO: 	Container coredns ready: true, restart count 0
-[It] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Trying to launch a pod without a label to get a node which can launch it.
-STEP: Explicitly delete pod here to free the resource it takes.
-STEP: Trying to apply a random label on the found node.
-STEP: verifying the node has the label kubernetes.io/e2e-aca8055d-bfed-4b9b-ae16-87b75752b18c 90
-STEP: Trying to create a pod(pod1) with hostport 54321 and hostIP 127.0.0.1 and expect scheduled
-STEP: Trying to create another pod(pod2) with hostport 54321 but hostIP 127.0.0.2 on the node which pod1 resides and expect scheduled
-STEP: Trying to create a third pod(pod3) with hostport 54321, hostIP 127.0.0.2 but use UDP protocol on the node which pod2 resides
-STEP: removing the label kubernetes.io/e2e-aca8055d-bfed-4b9b-ae16-87b75752b18c off the node adoring-wozniak-54dcfd79fc-948mf
-STEP: verifying the node doesn't have the label kubernetes.io/e2e-aca8055d-bfed-4b9b-ae16-87b75752b18c
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:49:02.059: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "sched-pred-9339" for this suite.
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[BeforeEach] Kubectl run rc
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1526
+[It] should create an rc from an image [Deprecated] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: running the image docker.io/library/httpd:2.4.38-alpine
+Feb 12 10:02:06.677: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 run e2e-test-httpd-rc --image=docker.io/library/httpd:2.4.38-alpine --generator=run/v1 --namespace=kubectl-1162'
+Feb 12 10:02:06.771: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Feb 12 10:02:06.771: INFO: stdout: "replicationcontroller/e2e-test-httpd-rc created\n"
+STEP: verifying the rc e2e-test-httpd-rc was created
+STEP: verifying the pod controlled by rc e2e-test-httpd-rc was created
+STEP: confirm that you can get logs from an rc
+Feb 12 10:02:06.788: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [e2e-test-httpd-rc-lldmv]
+Feb 12 10:02:06.788: INFO: Waiting up to 5m0s for pod "e2e-test-httpd-rc-lldmv" in namespace "kubectl-1162" to be "running and ready"
+Feb 12 10:02:06.795: INFO: Pod "e2e-test-httpd-rc-lldmv": Phase="Pending", Reason="", readiness=false. Elapsed: 7.052342ms
+Feb 12 10:02:08.802: INFO: Pod "e2e-test-httpd-rc-lldmv": Phase="Running", Reason="", readiness=true. Elapsed: 2.013597277s
+Feb 12 10:02:08.802: INFO: Pod "e2e-test-httpd-rc-lldmv" satisfied condition "running and ready"
+Feb 12 10:02:08.802: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [e2e-test-httpd-rc-lldmv]
+Feb 12 10:02:08.802: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 logs rc/e2e-test-httpd-rc --namespace=kubectl-1162'
+Feb 12 10:02:09.041: INFO: stderr: ""
+Feb 12 10:02:09.041: INFO: stdout: "AH00558: httpd: Could not reliably determine the server's fully qualified domain name, using 172.25.1.39. Set the 'ServerName' directive globally to suppress this message\nAH00558: httpd: Could not reliably determine the server's fully qualified domain name, using 172.25.1.39. Set the 'ServerName' directive globally to suppress this message\n[Fri Feb 12 10:02:08.022938 2021] [mpm_event:notice] [pid 1:tid 139738318953320] AH00489: Apache/2.4.38 (Unix) configured -- resuming normal operations\n[Fri Feb 12 10:02:08.022991 2021] [core:notice] [pid 1:tid 139738318953320] AH00094: Command line: 'httpd -D FOREGROUND'\n"
+[AfterEach] Kubectl run rc
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1531
+Feb 12 10:02:09.041: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete rc e2e-test-httpd-rc --namespace=kubectl-1162'
+Feb 12 10:02:09.135: INFO: stderr: ""
+Feb 12 10:02:09.135: INFO: stdout: "replicationcontroller \"e2e-test-httpd-rc\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:02:09.135: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-1162" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run rc should create an rc from an image [Deprecated] [Conformance]","total":280,"completed":39,"skipped":515,"failed":0}
+SSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:02:09.153: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-1777
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
+STEP: Creating service test in namespace statefulset-1777
+[It] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating stateful set ss in namespace statefulset-1777
+STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-1777
+Feb 12 10:02:09.368: INFO: Found 0 stateful pods, waiting for 1
+Feb 12 10:02:19.376: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod
+Feb 12 10:02:19.382: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-1777 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Feb 12 10:02:19.968: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Feb 12 10:02:19.968: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Feb 12 10:02:19.968: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Feb 12 10:02:19.975: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
+Feb 12 10:02:29.981: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Feb 12 10:02:29.981: INFO: Waiting for statefulset status.replicas updated to 0
+Feb 12 10:02:30.034: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
+Feb 12 10:02:30.034: INFO: ss-0  nostalgic-fermat-cdc5d8777-dbhjj  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:19 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  }]
+Feb 12 10:02:30.034: INFO: ss-1                                    Pending         []
+Feb 12 10:02:30.034: INFO: 
+Feb 12 10:02:30.034: INFO: StatefulSet ss has not reached scale 3, at 2
+Feb 12 10:02:31.040: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.970171855s
+Feb 12 10:02:32.048: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.964316507s
+Feb 12 10:02:33.054: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.956574552s
+Feb 12 10:02:34.061: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.949737543s
+Feb 12 10:02:35.068: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.943252794s
+Feb 12 10:02:36.074: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.936386304s
+Feb 12 10:02:37.080: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.93058483s
+Feb 12 10:02:38.088: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.923920311s
+Feb 12 10:02:39.142: INFO: Verifying statefulset ss doesn't scale past 3 for another 916.474652ms
+STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-1777
+Feb 12 10:02:40.151: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-1777 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:02:40.780: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
+Feb 12 10:02:40.780: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
+Feb 12 10:02:40.780: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
+
+Feb 12 10:02:40.780: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-1777 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:02:41.348: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n"
+Feb 12 10:02:41.348: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
+Feb 12 10:02:41.348: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
+
+Feb 12 10:02:41.348: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-1777 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:02:41.919: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n"
+Feb 12 10:02:41.919: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
+Feb 12 10:02:41.919: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
+
+Feb 12 10:02:41.927: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+Feb 12 10:02:41.927: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
+Feb 12 10:02:41.927: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Scale down will not halt with unhealthy stateful pod
+Feb 12 10:02:41.934: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-1777 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Feb 12 10:02:42.487: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Feb 12 10:02:42.487: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Feb 12 10:02:42.487: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Feb 12 10:02:42.487: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-1777 ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Feb 12 10:02:43.100: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Feb 12 10:02:43.100: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Feb 12 10:02:43.100: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Feb 12 10:02:43.100: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-1777 ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Feb 12 10:02:43.651: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Feb 12 10:02:43.651: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Feb 12 10:02:43.651: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Feb 12 10:02:43.651: INFO: Waiting for statefulset status.replicas updated to 0
+Feb 12 10:02:43.655: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1
+Feb 12 10:02:53.672: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Feb 12 10:02:53.672: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
+Feb 12 10:02:53.672: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
+Feb 12 10:02:53.733: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
+Feb 12 10:02:53.734: INFO: ss-0  nostalgic-fermat-cdc5d8777-dbhjj  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  }]
+Feb 12 10:02:53.734: INFO: ss-1  nostalgic-fermat-cdc5d8777-xg74t  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:53.734: INFO: ss-2  nostalgic-fermat-cdc5d8777-r6snm  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:53.734: INFO: 
+Feb 12 10:02:53.734: INFO: StatefulSet ss has not reached scale 0, at 3
+Feb 12 10:02:54.748: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
+Feb 12 10:02:54.748: INFO: ss-0  nostalgic-fermat-cdc5d8777-dbhjj  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  }]
+Feb 12 10:02:54.748: INFO: ss-1  nostalgic-fermat-cdc5d8777-xg74t  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:54.748: INFO: ss-2  nostalgic-fermat-cdc5d8777-r6snm  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:54.748: INFO: 
+Feb 12 10:02:54.748: INFO: StatefulSet ss has not reached scale 0, at 3
+Feb 12 10:02:55.756: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
+Feb 12 10:02:55.756: INFO: ss-0  nostalgic-fermat-cdc5d8777-dbhjj  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  }]
+Feb 12 10:02:55.756: INFO: ss-1  nostalgic-fermat-cdc5d8777-xg74t  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:55.756: INFO: ss-2  nostalgic-fermat-cdc5d8777-r6snm  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:55.756: INFO: 
+Feb 12 10:02:55.756: INFO: StatefulSet ss has not reached scale 0, at 3
+Feb 12 10:02:56.763: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
+Feb 12 10:02:56.763: INFO: ss-0  nostalgic-fermat-cdc5d8777-dbhjj  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  }]
+Feb 12 10:02:56.763: INFO: ss-1  nostalgic-fermat-cdc5d8777-xg74t  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:56.763: INFO: ss-2  nostalgic-fermat-cdc5d8777-r6snm  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:56.763: INFO: 
+Feb 12 10:02:56.763: INFO: StatefulSet ss has not reached scale 0, at 3
+Feb 12 10:02:57.770: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
+Feb 12 10:02:57.770: INFO: ss-0  nostalgic-fermat-cdc5d8777-dbhjj  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  }]
+Feb 12 10:02:57.771: INFO: ss-1  nostalgic-fermat-cdc5d8777-xg74t  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:57.771: INFO: ss-2  nostalgic-fermat-cdc5d8777-r6snm  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:57.771: INFO: 
+Feb 12 10:02:57.771: INFO: StatefulSet ss has not reached scale 0, at 3
+Feb 12 10:02:58.780: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
+Feb 12 10:02:58.780: INFO: ss-0  nostalgic-fermat-cdc5d8777-dbhjj  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  }]
+Feb 12 10:02:58.780: INFO: ss-1  nostalgic-fermat-cdc5d8777-xg74t  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:58.780: INFO: ss-2  nostalgic-fermat-cdc5d8777-r6snm  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:58.780: INFO: 
+Feb 12 10:02:58.780: INFO: StatefulSet ss has not reached scale 0, at 3
+Feb 12 10:02:59.802: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
+Feb 12 10:02:59.802: INFO: ss-0  nostalgic-fermat-cdc5d8777-dbhjj  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  }]
+Feb 12 10:02:59.802: INFO: ss-1  nostalgic-fermat-cdc5d8777-xg74t  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:59.802: INFO: ss-2  nostalgic-fermat-cdc5d8777-r6snm  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:02:59.802: INFO: 
+Feb 12 10:02:59.802: INFO: StatefulSet ss has not reached scale 0, at 3
+Feb 12 10:03:00.830: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
+Feb 12 10:03:00.830: INFO: ss-0  nostalgic-fermat-cdc5d8777-dbhjj  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  }]
+Feb 12 10:03:00.830: INFO: ss-1  nostalgic-fermat-cdc5d8777-xg74t  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:03:00.830: INFO: ss-2  nostalgic-fermat-cdc5d8777-r6snm  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:03:00.830: INFO: 
+Feb 12 10:03:00.830: INFO: StatefulSet ss has not reached scale 0, at 3
+Feb 12 10:03:01.838: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
+Feb 12 10:03:01.838: INFO: ss-0  nostalgic-fermat-cdc5d8777-dbhjj  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:42 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:09 +0000 UTC  }]
+Feb 12 10:03:01.838: INFO: ss-1  nostalgic-fermat-cdc5d8777-xg74t  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:43 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:03:01.838: INFO: ss-2  nostalgic-fermat-cdc5d8777-r6snm  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:44 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2021-02-12 10:02:29 +0000 UTC  }]
+Feb 12 10:03:01.838: INFO: 
+Feb 12 10:03:01.838: INFO: StatefulSet ss has not reached scale 0, at 3
+Feb 12 10:03:02.845: INFO: Verifying statefulset ss doesn't scale past 0 for another 848.633516ms
+STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-1777
+Feb 12 10:03:03.854: INFO: Scaling statefulset ss to 0
+Feb 12 10:03:03.878: INFO: Waiting for statefulset status.replicas updated to 0
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
+Feb 12 10:03:03.892: INFO: Deleting all statefulset in ns statefulset-1777
+Feb 12 10:03:03.913: INFO: Scaling statefulset ss to 0
+Feb 12 10:03:03.930: INFO: Waiting for statefulset status.replicas updated to 0
+Feb 12 10:03:03.935: INFO: Deleting statefulset ss
+[AfterEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:03:03.958: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-1777" for this suite.
 
-• [SLOW TEST:12.453 seconds]
-[sig-scheduling] SchedulerPredicates [Serial]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
-  validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:54.821 seconds]
+[sig-apps] StatefulSet
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+    Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]","total":280,"completed":41,"skipped":770,"failed":0}
+{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]","total":280,"completed":40,"skipped":535,"failed":0}
+[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
+  should be possible to delete [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:03:03.976: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-5002
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[BeforeEach] when scheduling a busybox command that always fails in a pod
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
+[It] should be possible to delete [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[AfterEach] [k8s.io] Kubelet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:03:04.180: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-5002" for this suite.
+•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance]","total":280,"completed":41,"skipped":535,"failed":0}
 SS
 ------------------------------
-[k8s.io] Variable Expansion 
-  should allow substituting values in a container's args [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Variable Expansion
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] Deployment 
+  deployment should support rollover [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:49:02.074: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename var-expansion
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-393
+Feb 12 10:03:04.229: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename deployment
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-5651
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should allow substituting values in a container's args [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test substitution in container's args
-Aug 30 16:49:02.243: INFO: Waiting up to 5m0s for pod "var-expansion-27cca060-9c17-4f7d-9379-19f1e5fe5dc9" in namespace "var-expansion-393" to be "success or failure"
-Aug 30 16:49:02.247: INFO: Pod "var-expansion-27cca060-9c17-4f7d-9379-19f1e5fe5dc9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.162286ms
-Aug 30 16:49:04.252: INFO: Pod "var-expansion-27cca060-9c17-4f7d-9379-19f1e5fe5dc9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.009390887s
-Aug 30 16:49:06.258: INFO: Pod "var-expansion-27cca060-9c17-4f7d-9379-19f1e5fe5dc9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01570666s
+[BeforeEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
+[It] deployment should support rollover [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:03:04.434: INFO: Pod name rollover-pod: Found 0 pods out of 1
+Feb 12 10:03:09.442: INFO: Pod name rollover-pod: Found 1 pods out of 1
+STEP: ensuring each pod is running
+Feb 12 10:03:09.442: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready
+Feb 12 10:03:11.448: INFO: Creating deployment "test-rollover-deployment"
+Feb 12 10:03:11.461: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations
+Feb 12 10:03:13.472: INFO: Check revision of new replica set for deployment "test-rollover-deployment"
+Feb 12 10:03:13.488: INFO: Ensure that both replica sets have 1 created replica
+Feb 12 10:03:13.500: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update
+Feb 12 10:03:13.520: INFO: Updating deployment test-rollover-deployment
+Feb 12 10:03:13.520: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller
+Feb 12 10:03:15.532: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2
+Feb 12 10:03:15.542: INFO: Make sure deployment "test-rollover-deployment" is complete
+Feb 12 10:03:15.552: INFO: all replica sets need to contain the pod-template-hash label
+Feb 12 10:03:15.552: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720993, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:03:17.563: INFO: all replica sets need to contain the pod-template-hash label
+Feb 12 10:03:17.564: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720995, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:03:19.563: INFO: all replica sets need to contain the pod-template-hash label
+Feb 12 10:03:19.563: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720995, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:03:21.564: INFO: all replica sets need to contain the pod-template-hash label
+Feb 12 10:03:21.564: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720995, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:03:23.566: INFO: all replica sets need to contain the pod-template-hash label
+Feb 12 10:03:23.566: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720995, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:03:25.564: INFO: all replica sets need to contain the pod-template-hash label
+Feb 12 10:03:25.564: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720995, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748720991, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-574d6dfbff\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:03:27.564: INFO: 
+Feb 12 10:03:27.564: INFO: Ensure that both old replica sets have no replicas
+[AfterEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
+Feb 12 10:03:27.583: INFO: Deployment "test-rollover-deployment":
+&Deployment{ObjectMeta:{test-rollover-deployment  deployment-5651 /apis/apps/v1/namespaces/deployment-5651/deployments/test-rollover-deployment e0f6cd79-c910-4c09-a84f-6a35117d7a75 9481 2 2021-02-12 10:03:11 +0000 UTC   map[name:rollover-pod] map[deployment.kubernetes.io/revision:2] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc0029597a8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2021-02-12 10:03:11 +0000 UTC,LastTransitionTime:2021-02-12 10:03:11 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rollover-deployment-574d6dfbff" has successfully progressed.,LastUpdateTime:2021-02-12 10:03:25 +0000 UTC,LastTransitionTime:2021-02-12 10:03:11 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},}
+
+Feb 12 10:03:27.588: INFO: New ReplicaSet "test-rollover-deployment-574d6dfbff" of Deployment "test-rollover-deployment":
+&ReplicaSet{ObjectMeta:{test-rollover-deployment-574d6dfbff  deployment-5651 /apis/apps/v1/namespaces/deployment-5651/replicasets/test-rollover-deployment-574d6dfbff a2398453-3538-4762-93d9-38866ef9c660 9470 2 2021-02-12 10:03:13 +0000 UTC   map[name:rollover-pod pod-template-hash:574d6dfbff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-rollover-deployment e0f6cd79-c910-4c09-a84f-6a35117d7a75 0xc002959dc7 0xc002959dc8}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 574d6dfbff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod pod-template-hash:574d6dfbff] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc002959e78  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},}
+Feb 12 10:03:27.588: INFO: All old ReplicaSets of Deployment "test-rollover-deployment":
+Feb 12 10:03:27.589: INFO: &ReplicaSet{ObjectMeta:{test-rollover-controller  deployment-5651 /apis/apps/v1/namespaces/deployment-5651/replicasets/test-rollover-controller 7b3499aa-7340-4e82-8438-99dd0e33790f 9480 2 2021-02-12 10:03:04 +0000 UTC   map[name:rollover-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2] [{apps/v1 Deployment test-rollover-deployment e0f6cd79-c910-4c09-a84f-6a35117d7a75 0xc002959c4f 0xc002959c90}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod pod:httpd] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc002959cf8  ClusterFirst map[]     false false false  PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
+Feb 12 10:03:27.589: INFO: &ReplicaSet{ObjectMeta:{test-rollover-deployment-f6c94f66c  deployment-5651 /apis/apps/v1/namespaces/deployment-5651/replicasets/test-rollover-deployment-f6c94f66c 9c5c672c-4dc9-4288-a6f5-a47b8df9971e 9400 2 2021-02-12 10:03:11 +0000 UTC   map[name:rollover-pod pod-template-hash:f6c94f66c] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-rollover-deployment e0f6cd79-c910-4c09-a84f-6a35117d7a75 0xc002959ef0 0xc002959ef1}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: f6c94f66c,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:rollover-pod pod-template-hash:f6c94f66c] map[] [] []  []} {[] [] [{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc002959f68  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
+Feb 12 10:03:27.593: INFO: Pod "test-rollover-deployment-574d6dfbff-qv9v5" is available:
+&Pod{ObjectMeta:{test-rollover-deployment-574d6dfbff-qv9v5 test-rollover-deployment-574d6dfbff- deployment-5651 /api/v1/namespaces/deployment-5651/pods/test-rollover-deployment-574d6dfbff-qv9v5 7a819623-9fc8-4b9d-93e1-b34fde7d9f3e 9422 0 2021-02-12 10:03:13 +0000 UTC   map[name:rollover-pod pod-template-hash:574d6dfbff] map[cni.projectcalico.org/podIP:172.25.1.42/32] [{apps/v1 ReplicaSet test-rollover-deployment-574d6dfbff a2398453-3538-4762-93d9-38866ef9c660 0xc0007c9127 0xc0007c9128}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-stkvz,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-stkvz,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-stkvz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-xg74t,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:03:13 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:03:15 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:03:15 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:03:13 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:138.68.97.197,PodIP:172.25.1.42,StartTime:2021-02-12 10:03:13 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-02-12 10:03:14 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/agnhost@sha256:daf5332100521b1256d0e3c56d697a238eaec3af48897ed9167cbadd426773b5,ContainerID:docker://245351ff035c837cc18ef0a12329e048e49ad97de7c6ce76a90fafb4a83fc913,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.42,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+[AfterEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:03:27.593: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-5651" for this suite.
+
+• [SLOW TEST:23.382 seconds]
+[sig-apps] Deployment
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  deployment should support rollover [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-apps] Deployment deployment should support rollover [Conformance]","total":280,"completed":42,"skipped":537,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected secret 
+  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:03:27.611: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3610
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating projection with secret that has name projected-secret-test-map-dd5f8a45-c7a6-477b-b144-9806c5f6cbf1
+STEP: Creating a pod to test consume secrets
+Feb 12 10:03:27.824: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-d559391b-6fb1-4667-b4dd-0b9f0d661700" in namespace "projected-3610" to be "success or failure"
+Feb 12 10:03:27.832: INFO: Pod "pod-projected-secrets-d559391b-6fb1-4667-b4dd-0b9f0d661700": Phase="Pending", Reason="", readiness=false. Elapsed: 7.696576ms
+Feb 12 10:03:29.843: INFO: Pod "pod-projected-secrets-d559391b-6fb1-4667-b4dd-0b9f0d661700": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.018327214s
 STEP: Saw pod success
-Aug 30 16:49:06.258: INFO: Pod "var-expansion-27cca060-9c17-4f7d-9379-19f1e5fe5dc9" satisfied condition "success or failure"
-Aug 30 16:49:06.263: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod var-expansion-27cca060-9c17-4f7d-9379-19f1e5fe5dc9 container dapi-container: 
+Feb 12 10:03:29.843: INFO: Pod "pod-projected-secrets-d559391b-6fb1-4667-b4dd-0b9f0d661700" satisfied condition "success or failure"
+Feb 12 10:03:29.850: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-dbhjj pod pod-projected-secrets-d559391b-6fb1-4667-b4dd-0b9f0d661700 container projected-secret-volume-test: 
 STEP: delete the pod
-Aug 30 16:49:06.333: INFO: Waiting for pod var-expansion-27cca060-9c17-4f7d-9379-19f1e5fe5dc9 to disappear
-Aug 30 16:49:06.338: INFO: Pod var-expansion-27cca060-9c17-4f7d-9379-19f1e5fe5dc9 no longer exists
-[AfterEach] [k8s.io] Variable Expansion
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:49:06.338: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "var-expansion-393" for this suite.
-•{"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance]","total":280,"completed":42,"skipped":772,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Feb 12 10:03:29.958: INFO: Waiting for pod pod-projected-secrets-d559391b-6fb1-4667-b4dd-0b9f0d661700 to disappear
+Feb 12 10:03:29.964: INFO: Pod pod-projected-secrets-d559391b-6fb1-4667-b4dd-0b9f0d661700 no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:03:29.964: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-3610" for this suite.
+•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":43,"skipped":576,"failed":0}
+SSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition 
+  creating/deleting custom resource definition objects works  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:49:06.356: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-1616
+Feb 12 10:03:30.017: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename custom-resource-definition
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-7539
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name configmap-test-volume-09ba011e-d647-4e57-9544-1e0510178fac
-STEP: Creating a pod to test consume configMaps
-Aug 30 16:49:06.533: INFO: Waiting up to 5m0s for pod "pod-configmaps-fcd7409b-b213-4b80-aa89-2bde8b7f62b9" in namespace "configmap-1616" to be "success or failure"
-Aug 30 16:49:06.537: INFO: Pod "pod-configmaps-fcd7409b-b213-4b80-aa89-2bde8b7f62b9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.189363ms
-Aug 30 16:49:08.543: INFO: Pod "pod-configmaps-fcd7409b-b213-4b80-aa89-2bde8b7f62b9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010092674s
-Aug 30 16:49:10.548: INFO: Pod "pod-configmaps-fcd7409b-b213-4b80-aa89-2bde8b7f62b9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015152772s
+[It] creating/deleting custom resource definition objects works  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:03:30.180: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:03:31.248: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "custom-resource-definition-7539" for this suite.
+•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works  [Conformance]","total":280,"completed":44,"skipped":581,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:03:31.267: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-5192
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir 0777 on node default medium
+Feb 12 10:03:31.449: INFO: Waiting up to 5m0s for pod "pod-442067a6-c48a-4eee-aa7f-d36ca6f44276" in namespace "emptydir-5192" to be "success or failure"
+Feb 12 10:03:31.455: INFO: Pod "pod-442067a6-c48a-4eee-aa7f-d36ca6f44276": Phase="Pending", Reason="", readiness=false. Elapsed: 6.248547ms
+Feb 12 10:03:33.462: INFO: Pod "pod-442067a6-c48a-4eee-aa7f-d36ca6f44276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013394552s
+Feb 12 10:03:35.472: INFO: Pod "pod-442067a6-c48a-4eee-aa7f-d36ca6f44276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022957469s
 STEP: Saw pod success
-Aug 30 16:49:10.548: INFO: Pod "pod-configmaps-fcd7409b-b213-4b80-aa89-2bde8b7f62b9" satisfied condition "success or failure"
-Aug 30 16:49:10.555: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod pod-configmaps-fcd7409b-b213-4b80-aa89-2bde8b7f62b9 container configmap-volume-test: 
+Feb 12 10:03:35.472: INFO: Pod "pod-442067a6-c48a-4eee-aa7f-d36ca6f44276" satisfied condition "success or failure"
+Feb 12 10:03:35.496: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-442067a6-c48a-4eee-aa7f-d36ca6f44276 container test-container: 
 STEP: delete the pod
-Aug 30 16:49:10.620: INFO: Waiting for pod pod-configmaps-fcd7409b-b213-4b80-aa89-2bde8b7f62b9 to disappear
-Aug 30 16:49:10.625: INFO: Pod pod-configmaps-fcd7409b-b213-4b80-aa89-2bde8b7f62b9 no longer exists
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:49:10.626: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-1616" for this suite.
-•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":43,"skipped":831,"failed":0}
-SSSSSSSSSSSSSSS
+Feb 12 10:03:35.615: INFO: Waiting for pod pod-442067a6-c48a-4eee-aa7f-d36ca6f44276 to disappear
+Feb 12 10:03:35.623: INFO: Pod pod-442067a6-c48a-4eee-aa7f-d36ca6f44276 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:03:35.623: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-5192" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":45,"skipped":619,"failed":0}
+SSSSSSS
 ------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates that NodeSelector is respected if not matching  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Downward API volume 
+  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:49:10.643: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename sched-pred
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-3398
+Feb 12 10:03:35.642: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-6360
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
-Aug 30 16:49:10.827: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-Aug 30 16:49:10.841: INFO: Waiting for terminating namespaces to be deleted...
-Aug 30 16:49:10.846: INFO: 
-Logging pods the kubelet thinks is on node adoring-wozniak-54dcfd79fc-6rshr before test
-Aug 30 16:49:10.908: INFO: kube-proxy-wxdxv from kube-system started at 2020-08-30 16:28:14 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.908: INFO: 	Container kube-proxy ready: true, restart count 0
-Aug 30 16:49:10.908: INFO: node-local-dns-cvxnf from kube-system started at 2020-08-30 16:28:34 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.908: INFO: 	Container node-cache ready: true, restart count 0
-Aug 30 16:49:10.908: INFO: logrotate-57bmz from kube-system started at 2020-08-30 16:28:34 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.908: INFO: 	Container logrotate ready: true, restart count 0
-Aug 30 16:49:10.908: INFO: sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-nd5nc from sonobuoy started at 2020-08-30 16:29:24 +0000 UTC (2 container statuses recorded)
-Aug 30 16:49:10.908: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 16:49:10.908: INFO: 	Container systemd-logs ready: true, restart count 0
-Aug 30 16:49:10.908: INFO: coredns-54457d966b-fbnz4 from kube-system started at 2020-08-30 16:28:39 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.908: INFO: 	Container coredns ready: true, restart count 0
-Aug 30 16:49:10.908: INFO: sonobuoy-e2e-job-cf49606f646f4c8a from sonobuoy started at 2020-08-30 16:29:23 +0000 UTC (2 container statuses recorded)
-Aug 30 16:49:10.908: INFO: 	Container e2e ready: true, restart count 0
-Aug 30 16:49:10.908: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 16:49:10.908: INFO: canal-54glj from kube-system started at 2020-08-30 16:28:14 +0000 UTC (2 container statuses recorded)
-Aug 30 16:49:10.908: INFO: 	Container calico-node ready: true, restart count 0
-Aug 30 16:49:10.908: INFO: 	Container kube-flannel ready: true, restart count 0
-Aug 30 16:49:10.908: INFO: user-ssh-keys-agent-xwrzj from kube-system started at 2020-08-30 16:28:14 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.908: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
-Aug 30 16:49:10.908: INFO: sonobuoy from sonobuoy started at 2020-08-30 16:29:18 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.908: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-Aug 30 16:49:10.908: INFO: 
-Logging pods the kubelet thinks is on node adoring-wozniak-54dcfd79fc-948mf before test
-Aug 30 16:49:10.995: INFO: dashboard-metrics-scraper-59bfc65dc9-94mm2 from kubernetes-dashboard started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.995: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
-Aug 30 16:49:10.995: INFO: pod1 from sched-pred-9339 started at 2020-08-30 16:48:51 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.995: INFO: 	Container pod1 ready: false, restart count 0
-Aug 30 16:49:10.995: INFO: dashboard-metrics-scraper-59bfc65dc9-fhfrl from kubernetes-dashboard started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.995: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
-Aug 30 16:49:10.995: INFO: logrotate-pw2ff from kube-system started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.995: INFO: 	Container logrotate ready: true, restart count 0
-Aug 30 16:49:10.995: INFO: sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-cdm8v from sonobuoy started at 2020-08-30 16:29:23 +0000 UTC (2 container statuses recorded)
-Aug 30 16:49:10.995: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 16:49:10.995: INFO: 	Container systemd-logs ready: true, restart count 0
-Aug 30 16:49:10.995: INFO: kube-proxy-v88gx from kube-system started at 2020-08-30 16:28:09 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.995: INFO: 	Container kube-proxy ready: true, restart count 0
-Aug 30 16:49:10.995: INFO: user-ssh-keys-agent-vkbs9 from kube-system started at 2020-08-30 16:28:09 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.995: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
-Aug 30 16:49:10.995: INFO: openvpn-client-78d595f58b-pht8r from kube-system started at 2020-08-30 16:28:29 +0000 UTC (2 container statuses recorded)
-Aug 30 16:49:10.995: INFO: 	Container dnat-controller ready: true, restart count 0
-Aug 30 16:49:10.995: INFO: 	Container openvpn-client ready: true, restart count 0
-Aug 30 16:49:10.995: INFO: pod3 from sched-pred-9339 started at 2020-08-30 16:48:58 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.996: INFO: 	Container pod3 ready: false, restart count 0
-Aug 30 16:49:10.996: INFO: canal-lg5hn from kube-system started at 2020-08-30 16:28:09 +0000 UTC (2 container statuses recorded)
-Aug 30 16:49:10.996: INFO: 	Container calico-node ready: true, restart count 0
-Aug 30 16:49:10.996: INFO: 	Container kube-flannel ready: true, restart count 0
-Aug 30 16:49:10.996: INFO: coredns-54457d966b-vr27r from kube-system started at 2020-08-30 16:28:39 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.996: INFO: 	Container coredns ready: true, restart count 0
-Aug 30 16:49:10.996: INFO: pod2 from sched-pred-9339 started at 2020-08-30 16:48:53 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.996: INFO: 	Container pod2 ready: false, restart count 0
-Aug 30 16:49:10.996: INFO: node-local-dns-wrs2k from kube-system started at 2020-08-30 16:28:29 +0000 UTC (1 container statuses recorded)
-Aug 30 16:49:10.996: INFO: 	Container node-cache ready: true, restart count 0
-[It] validates that NodeSelector is respected if not matching  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Trying to schedule Pod with nonempty NodeSelector.
-STEP: Considering event: 
-Type = [Warning], Name = [restricted-pod.163019b937d891cc], Reason = [FailedScheduling], Message = [0/2 nodes are available: 2 node(s) didn't match node selector.]
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:49:12.030: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "sched-pred-3398" for this suite.
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
-•{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching  [Conformance]","total":280,"completed":44,"skipped":846,"failed":0}
-SSSSSSSSSSSSSSSSSS
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Feb 12 10:03:35.853: INFO: Waiting up to 5m0s for pod "downwardapi-volume-718c9eee-4d2c-4d1d-a8e9-b9e41ff71892" in namespace "downward-api-6360" to be "success or failure"
+Feb 12 10:03:35.861: INFO: Pod "downwardapi-volume-718c9eee-4d2c-4d1d-a8e9-b9e41ff71892": Phase="Pending", Reason="", readiness=false. Elapsed: 8.356381ms
+Feb 12 10:03:37.870: INFO: Pod "downwardapi-volume-718c9eee-4d2c-4d1d-a8e9-b9e41ff71892": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017026479s
+Feb 12 10:03:39.879: INFO: Pod "downwardapi-volume-718c9eee-4d2c-4d1d-a8e9-b9e41ff71892": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.026447194s
+STEP: Saw pod success
+Feb 12 10:03:39.879: INFO: Pod "downwardapi-volume-718c9eee-4d2c-4d1d-a8e9-b9e41ff71892" satisfied condition "success or failure"
+Feb 12 10:03:39.908: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-718c9eee-4d2c-4d1d-a8e9-b9e41ff71892 container client-container: 
+STEP: delete the pod
+Feb 12 10:03:40.020: INFO: Waiting for pod downwardapi-volume-718c9eee-4d2c-4d1d-a8e9-b9e41ff71892 to disappear
+Feb 12 10:03:40.026: INFO: Pod downwardapi-volume-718c9eee-4d2c-4d1d-a8e9-b9e41ff71892 no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:03:40.026: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-6360" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":46,"skipped":626,"failed":0}
+SSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:03:40.044: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-3193
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir 0666 on tmpfs
+Feb 12 10:03:40.229: INFO: Waiting up to 5m0s for pod "pod-85f1282a-e7dd-4332-99fa-f313ae5d9940" in namespace "emptydir-3193" to be "success or failure"
+Feb 12 10:03:40.236: INFO: Pod "pod-85f1282a-e7dd-4332-99fa-f313ae5d9940": Phase="Pending", Reason="", readiness=false. Elapsed: 6.142479ms
+Feb 12 10:03:42.241: INFO: Pod "pod-85f1282a-e7dd-4332-99fa-f313ae5d9940": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011595446s
+STEP: Saw pod success
+Feb 12 10:03:42.241: INFO: Pod "pod-85f1282a-e7dd-4332-99fa-f313ae5d9940" satisfied condition "success or failure"
+Feb 12 10:03:42.246: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-85f1282a-e7dd-4332-99fa-f313ae5d9940 container test-container: 
+STEP: delete the pod
+Feb 12 10:03:42.325: INFO: Waiting for pod pod-85f1282a-e7dd-4332-99fa-f313ae5d9940 to disappear
+Feb 12 10:03:42.331: INFO: Pod pod-85f1282a-e7dd-4332-99fa-f313ae5d9940 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:03:42.331: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-3193" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":47,"skipped":639,"failed":0}
+SSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected downwardAPI 
+  should provide container's cpu request [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:03:42.348: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1369
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should provide container's cpu request [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Feb 12 10:03:42.536: INFO: Waiting up to 5m0s for pod "downwardapi-volume-a24e309f-e398-4f50-b81f-2ddbf680befb" in namespace "projected-1369" to be "success or failure"
+Feb 12 10:03:42.542: INFO: Pod "downwardapi-volume-a24e309f-e398-4f50-b81f-2ddbf680befb": Phase="Pending", Reason="", readiness=false. Elapsed: 6.118046ms
+Feb 12 10:03:44.549: INFO: Pod "downwardapi-volume-a24e309f-e398-4f50-b81f-2ddbf680befb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012449342s
+STEP: Saw pod success
+Feb 12 10:03:44.549: INFO: Pod "downwardapi-volume-a24e309f-e398-4f50-b81f-2ddbf680befb" satisfied condition "success or failure"
+Feb 12 10:03:44.554: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-a24e309f-e398-4f50-b81f-2ddbf680befb container client-container: 
+STEP: delete the pod
+Feb 12 10:03:44.620: INFO: Waiting for pod downwardapi-volume-a24e309f-e398-4f50-b81f-2ddbf680befb to disappear
+Feb 12 10:03:44.626: INFO: Pod downwardapi-volume-a24e309f-e398-4f50-b81f-2ddbf680befb no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:03:44.626: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-1369" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance]","total":280,"completed":48,"skipped":652,"failed":0}
+SSS
 ------------------------------
 [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
   updates the published spec when one version gets renamed [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:49:12.046: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:03:44.645: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename crd-publish-openapi
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-3042
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-698
 STEP: Waiting for a default service account to be provisioned in namespace
 [It] updates the published spec when one version gets renamed [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: set up a multi version CRD
-Aug 30 16:49:12.203: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:03:44.806: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: rename a version
 STEP: check the new version name is served
 STEP: check the old version name is removed
 STEP: check the other version is not changed
 [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:49:29.791: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-3042" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:04:01.120: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-698" for this suite.
 
-• [SLOW TEST:17.759 seconds]
+• [SLOW TEST:16.492 seconds]
 [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
   updates the published spec when one version gets renamed [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance]","total":280,"completed":45,"skipped":864,"failed":0}
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance]","total":280,"completed":49,"skipped":655,"failed":0}
 SSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Proxy server 
-  should support --unix-socket=/path  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] DNS 
+  should provide DNS for pods for Hostname [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:49:29.805: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-9346
+Feb 12 10:04:01.140: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename dns
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-2758
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[It] should support --unix-socket=/path  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Starting the proxy
-Aug 30 16:49:29.962: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-962893509 proxy --unix-socket=/tmp/kubectl-proxy-unix984500412/test'
-STEP: retrieving proxy /api/ output
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:49:30.008: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-9346" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support --unix-socket=/path  [Conformance]","total":280,"completed":46,"skipped":874,"failed":0}
-SSSSSSSSSSSS
+[It] should provide DNS for pods for Hostname [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a test headless service
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-2758.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-2.dns-test-service-2.dns-2758.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/wheezy_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-2758.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
+
+STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-2758.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-2.dns-test-service-2.dns-2758.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/jessie_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-2758.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+
+STEP: creating a pod to probe DNS
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Feb 12 10:04:06.150: INFO: DNS probes using dns-2758/dns-test-e348b30d-8cb7-4bd6-ac7d-90f3a3ef3830 succeeded
+
+STEP: deleting the pod
+STEP: deleting the test headless service
+[AfterEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:04:06.255: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-2758" for this suite.
+
+• [SLOW TEST:5.139 seconds]
+[sig-network] DNS
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should provide DNS for pods for Hostname [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] Pods 
-  should support retrieving logs from the container over websockets [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Hostname [LinuxOnly] [Conformance]","total":280,"completed":50,"skipped":665,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Docker Containers 
+  should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Docker Containers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:49:30.023: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-6855
+Feb 12 10:04:06.281: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename containers
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-5263
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
-[It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 16:49:30.189: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: creating the pod
-STEP: submitting the pod to kubernetes
-[AfterEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:49:32.241: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-6855" for this suite.
-•{"msg":"PASSED [k8s.io] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]","total":280,"completed":47,"skipped":886,"failed":0}
-SSSSSSSSSSSSSSSSSSSS
+[It] should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test override arguments
+Feb 12 10:04:06.517: INFO: Waiting up to 5m0s for pod "client-containers-4c917468-46ab-4784-9580-6da1efd41a9e" in namespace "containers-5263" to be "success or failure"
+Feb 12 10:04:06.525: INFO: Pod "client-containers-4c917468-46ab-4784-9580-6da1efd41a9e": Phase="Pending", Reason="", readiness=false. Elapsed: 7.536505ms
+Feb 12 10:04:08.532: INFO: Pod "client-containers-4c917468-46ab-4784-9580-6da1efd41a9e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014531136s
+Feb 12 10:04:10.538: INFO: Pod "client-containers-4c917468-46ab-4784-9580-6da1efd41a9e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020265743s
+STEP: Saw pod success
+Feb 12 10:04:10.538: INFO: Pod "client-containers-4c917468-46ab-4784-9580-6da1efd41a9e" satisfied condition "success or failure"
+Feb 12 10:04:10.543: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod client-containers-4c917468-46ab-4784-9580-6da1efd41a9e container test-container: 
+STEP: delete the pod
+Feb 12 10:04:10.650: INFO: Waiting for pod client-containers-4c917468-46ab-4784-9580-6da1efd41a9e to disappear
+Feb 12 10:04:10.656: INFO: Pod client-containers-4c917468-46ab-4784-9580-6da1efd41a9e no longer exists
+[AfterEach] [k8s.io] Docker Containers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:04:10.656: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "containers-5263" for this suite.
+•{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default arguments (docker cmd) [NodeConformance] [Conformance]","total":280,"completed":51,"skipped":707,"failed":0}
+SSSSSSS
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should unconditionally reject operations on fail closed webhook [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Probing container 
+  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:49:32.258: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-2052
+Feb 12 10:04:10.677: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-6342
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Aug 30 16:49:32.922: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Aug 30 16:49:34.936: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402972, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402972, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402972, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734402972, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Aug 30 16:49:37.961: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should unconditionally reject operations on fail closed webhook [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering a webhook that server cannot talk to, with fail closed policy, via the AdmissionRegistration API
-STEP: create a namespace for the webhook
-STEP: create a configmap should be unconditionally rejected by the webhook
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:49:38.315: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-2052" for this suite.
-STEP: Destroying namespace "webhook-2052-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+[BeforeEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[AfterEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:05:10.864: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-6342" for this suite.
 
-• [SLOW TEST:6.150 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should unconditionally reject operations on fail closed webhook [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:60.221 seconds]
+[k8s.io] Probing container
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]","total":280,"completed":48,"skipped":906,"failed":0}
-SSSSSSSS
+{"msg":"PASSED [k8s.io] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]","total":280,"completed":52,"skipped":714,"failed":0}
+S
 ------------------------------
-[sig-api-machinery] ResourceQuota 
-  should create a ResourceQuota and capture the life of a configMap. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute prestop exec hook properly [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Container Lifecycle Hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:49:38.408: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename resourcequota
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-5625
+Feb 12 10:05:10.899: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-lifecycle-hook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-6310
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should create a ResourceQuota and capture the life of a configMap. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Counting existing ResourceQuota
-STEP: Creating a ResourceQuota
-STEP: Ensuring resource quota status is calculated
-STEP: Creating a ConfigMap
-STEP: Ensuring resource quota status captures configMap creation
-STEP: Deleting a ConfigMap
-STEP: Ensuring resource quota status released usage
-[AfterEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:49:54.630: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-5625" for this suite.
-
-• [SLOW TEST:16.237 seconds]
-[sig-api-machinery] ResourceQuota
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should create a ResourceQuota and capture the life of a configMap. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute prestop exec hook properly [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: create the pod with lifecycle hook
+STEP: delete the pod with lifecycle hook
+Feb 12 10:05:15.175: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Feb 12 10:05:15.181: INFO: Pod pod-with-prestop-exec-hook still exists
+Feb 12 10:05:17.181: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Feb 12 10:05:17.204: INFO: Pod pod-with-prestop-exec-hook still exists
+Feb 12 10:05:19.182: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear
+Feb 12 10:05:19.194: INFO: Pod pod-with-prestop-exec-hook no longer exists
+STEP: check prestop hook
+[AfterEach] [k8s.io] Container Lifecycle Hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:05:19.249: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-6310" for this suite.
+
+• [SLOW TEST:8.378 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  when create a pod with lifecycle hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
+    should execute prestop exec hook properly [NodeConformance] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance]","total":280,"completed":49,"skipped":914,"failed":0}
-SSSS
+{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance]","total":280,"completed":53,"skipped":715,"failed":0}
+SSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Secrets 
+  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:49:54.645: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-8225
+Feb 12 10:05:19.278: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-64
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name configmap-test-volume-map-bf5659d9-be44-4a12-86fc-b7bc3c1a9464
-STEP: Creating a pod to test consume configMaps
-Aug 30 16:49:54.823: INFO: Waiting up to 5m0s for pod "pod-configmaps-2a726753-e88b-405a-b5aa-a4a72bd355c9" in namespace "configmap-8225" to be "success or failure"
-Aug 30 16:49:54.830: INFO: Pod "pod-configmaps-2a726753-e88b-405a-b5aa-a4a72bd355c9": Phase="Pending", Reason="", readiness=false. Elapsed: 7.148696ms
-Aug 30 16:49:56.837: INFO: Pod "pod-configmaps-2a726753-e88b-405a-b5aa-a4a72bd355c9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.013710561s
+[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating secret with name secret-test-map-4cef49e5-7f69-4e76-b4bf-9760f56ea4ec
+STEP: Creating a pod to test consume secrets
+Feb 12 10:05:19.456: INFO: Waiting up to 5m0s for pod "pod-secrets-41a6283d-69dd-4f51-a2fe-e2a9078cdb6f" in namespace "secrets-64" to be "success or failure"
+Feb 12 10:05:19.475: INFO: Pod "pod-secrets-41a6283d-69dd-4f51-a2fe-e2a9078cdb6f": Phase="Pending", Reason="", readiness=false. Elapsed: 18.131747ms
+Feb 12 10:05:21.491: INFO: Pod "pod-secrets-41a6283d-69dd-4f51-a2fe-e2a9078cdb6f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.034975675s
 STEP: Saw pod success
-Aug 30 16:49:56.837: INFO: Pod "pod-configmaps-2a726753-e88b-405a-b5aa-a4a72bd355c9" satisfied condition "success or failure"
-Aug 30 16:49:56.841: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-configmaps-2a726753-e88b-405a-b5aa-a4a72bd355c9 container configmap-volume-test: 
+Feb 12 10:05:21.492: INFO: Pod "pod-secrets-41a6283d-69dd-4f51-a2fe-e2a9078cdb6f" satisfied condition "success or failure"
+Feb 12 10:05:21.497: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-dbhjj pod pod-secrets-41a6283d-69dd-4f51-a2fe-e2a9078cdb6f container secret-volume-test: 
 STEP: delete the pod
-Aug 30 16:49:56.911: INFO: Waiting for pod pod-configmaps-2a726753-e88b-405a-b5aa-a4a72bd355c9 to disappear
-Aug 30 16:49:56.915: INFO: Pod pod-configmaps-2a726753-e88b-405a-b5aa-a4a72bd355c9 no longer exists
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:49:56.915: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-8225" for this suite.
-•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":50,"skipped":918,"failed":0}
-SSSSS
-------------------------------
-[sig-cli] Kubectl client Proxy server 
-  should support proxy with --port 0  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:49:56.930: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7092
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[It] should support proxy with --port 0  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: starting the proxy server
-Aug 30 16:49:57.091: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-962893509 proxy -p 0 --disable-filter'
-STEP: curling proxy /api/ output
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:49:57.145: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-7092" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support proxy with --port 0  [Conformance]","total":280,"completed":51,"skipped":923,"failed":0}
-SSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  works for CRD without validation schema [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:49:57.162: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename crd-publish-openapi
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-8920
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] works for CRD without validation schema [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 16:49:57.328: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: client-side validation (kubectl create and apply) allows request with any unknown properties
-Aug 30 16:50:00.707: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-8920 create -f -'
-Aug 30 16:50:01.059: INFO: stderr: ""
-Aug 30 16:50:01.059: INFO: stdout: "e2e-test-crd-publish-openapi-8397-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n"
-Aug 30 16:50:01.059: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-8920 delete e2e-test-crd-publish-openapi-8397-crds test-cr'
-Aug 30 16:50:01.186: INFO: stderr: ""
-Aug 30 16:50:01.186: INFO: stdout: "e2e-test-crd-publish-openapi-8397-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n"
-Aug 30 16:50:01.186: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-8920 apply -f -'
-Aug 30 16:50:01.343: INFO: stderr: ""
-Aug 30 16:50:01.343: INFO: stdout: "e2e-test-crd-publish-openapi-8397-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n"
-Aug 30 16:50:01.343: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-8920 delete e2e-test-crd-publish-openapi-8397-crds test-cr'
-Aug 30 16:50:01.418: INFO: stderr: ""
-Aug 30 16:50:01.418: INFO: stdout: "e2e-test-crd-publish-openapi-8397-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n"
-STEP: kubectl explain works to explain CR without validation schema
-Aug 30 16:50:01.418: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 explain e2e-test-crd-publish-openapi-8397-crds'
-Aug 30 16:50:01.643: INFO: stderr: ""
-Aug 30 16:50:01.643: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-8397-crd\nVERSION:  crd-publish-openapi-test-empty.example.com/v1\n\nDESCRIPTION:\n     \n"
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:50:04.485: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-8920" for this suite.
-
-• [SLOW TEST:7.339 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for CRD without validation schema [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance]","total":280,"completed":52,"skipped":944,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Feb 12 10:05:21.607: INFO: Waiting for pod pod-secrets-41a6283d-69dd-4f51-a2fe-e2a9078cdb6f to disappear
+Feb 12 10:05:21.625: INFO: Pod pod-secrets-41a6283d-69dd-4f51-a2fe-e2a9078cdb6f no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:05:21.625: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-64" for this suite.
+•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":54,"skipped":729,"failed":0}
+SSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Kubelet when scheduling a busybox Pod with hostAliases 
-  should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
+  should have an terminated reason [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:50:04.511: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:05:21.643: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename kubelet-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-984
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-9491
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
-[It] should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[BeforeEach] when scheduling a busybox command that always fails in a pod
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
+[It] should have an terminated reason [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [AfterEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:50:08.717: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubelet-test-984" for this suite.
-•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox Pod with hostAliases should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":53,"skipped":974,"failed":0}
-SSSSSSS
-------------------------------
-[sig-api-machinery] Secrets 
-  should be consumable from pods in env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:50:08.732: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-2846
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-2c7b1eb4-84fb-478f-9827-1801e51a849c
-STEP: Creating a pod to test consume secrets
-Aug 30 16:50:08.923: INFO: Waiting up to 5m0s for pod "pod-secrets-fe951ddf-90c7-4a4f-aa9c-b0b0207bd45f" in namespace "secrets-2846" to be "success or failure"
-Aug 30 16:50:08.930: INFO: Pod "pod-secrets-fe951ddf-90c7-4a4f-aa9c-b0b0207bd45f": Phase="Pending", Reason="", readiness=false. Elapsed: 6.833296ms
-Aug 30 16:50:10.936: INFO: Pod "pod-secrets-fe951ddf-90c7-4a4f-aa9c-b0b0207bd45f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012307109s
-Aug 30 16:50:12.941: INFO: Pod "pod-secrets-fe951ddf-90c7-4a4f-aa9c-b0b0207bd45f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017855898s
-STEP: Saw pod success
-Aug 30 16:50:12.941: INFO: Pod "pod-secrets-fe951ddf-90c7-4a4f-aa9c-b0b0207bd45f" satisfied condition "success or failure"
-Aug 30 16:50:12.946: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod pod-secrets-fe951ddf-90c7-4a4f-aa9c-b0b0207bd45f container secret-env-test: 
-STEP: delete the pod
-Aug 30 16:50:13.012: INFO: Waiting for pod pod-secrets-fe951ddf-90c7-4a4f-aa9c-b0b0207bd45f to disappear
-Aug 30 16:50:13.017: INFO: Pod pod-secrets-fe951ddf-90c7-4a4f-aa9c-b0b0207bd45f no longer exists
-[AfterEach] [sig-api-machinery] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:50:13.018: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-2846" for this suite.
-•{"msg":"PASSED [sig-api-machinery] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance]","total":280,"completed":54,"skipped":981,"failed":0}
-SSSSSSSSSSSSSSSS
-------------------------------
-[sig-network] DNS 
-  should support configurable pod DNS nameservers [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:50:13.033: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename dns
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-7306
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support configurable pod DNS nameservers [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod with dnsPolicy=None and customized dnsConfig...
-Aug 30 16:50:13.203: INFO: Created pod &Pod{ObjectMeta:{dns-7306  dns-7306 /api/v1/namespaces/dns-7306/pods/dns-7306 aa2c0257-2bcc-427b-903b-35c2c5e6a861 8644 0 2020-08-30 16:50:13 +0000 UTC   map[] map[] [] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-wtr57,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-wtr57,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[pause],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-wtr57,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:None,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:&PodDNSConfig{Nameservers:[1.1.1.1],Searches:[resolv.conf.local],Options:[]PodDNSConfigOption{},},ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-STEP: Verifying customized DNS suffix list is configured on pod...
-Aug 30 16:50:15.215: INFO: ExecWithOptions {Command:[/agnhost dns-suffix] Namespace:dns-7306 PodName:dns-7306 ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:50:15.215: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Verifying customized DNS server is configured on pod...
-Aug 30 16:50:15.729: INFO: ExecWithOptions {Command:[/agnhost dns-server-list] Namespace:dns-7306 PodName:dns-7306 ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:50:15.729: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:50:16.219: INFO: Deleting pod dns-7306...
-[AfterEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:50:16.234: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-7306" for this suite.
-•{"msg":"PASSED [sig-network] DNS should support configurable pod DNS nameservers [Conformance]","total":280,"completed":55,"skipped":997,"failed":0}
-SSSSSSSSSSSSSSS
-------------------------------
-[k8s.io] Pods 
-  should get a host IP [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:50:16.248: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-7014
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
-[It] should get a host IP [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating pod
-Aug 30 16:50:18.438: INFO: Pod pod-hostip-56164592-8487-4f96-8d59-771151151a01 has hostIP: 46.101.135.210
-[AfterEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:50:18.438: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-7014" for this suite.
-•{"msg":"PASSED [k8s.io] Pods should get a host IP [NodeConformance] [Conformance]","total":280,"completed":56,"skipped":1012,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[k8s.io] Probing container 
-  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:50:18.458: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-120
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod liveness-2a0b2d9e-7701-4c95-a3b8-80b14889492e in namespace container-probe-120
-Aug 30 16:50:20.633: INFO: Started pod liveness-2a0b2d9e-7701-4c95-a3b8-80b14889492e in namespace container-probe-120
-STEP: checking the pod's current state and verifying that restartCount is present
-Aug 30 16:50:20.639: INFO: Initial restart count of pod liveness-2a0b2d9e-7701-4c95-a3b8-80b14889492e is 0
-Aug 30 16:50:42.712: INFO: Restart count of pod container-probe-120/liveness-2a0b2d9e-7701-4c95-a3b8-80b14889492e is now 1 (22.073414768s elapsed)
-STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:50:42.728: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-120" for this suite.
-
-• [SLOW TEST:24.300 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [k8s.io] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":280,"completed":57,"skipped":1034,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:05:25.876: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-9491" for this suite.
+•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance]","total":280,"completed":55,"skipped":743,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] [sig-node] PreStop 
-  should call prestop when killing a pod  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] [sig-node] PreStop
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  listing validating webhooks should work [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:50:42.759: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename prestop
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in prestop-3260
+Feb 12 10:05:25.919: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-8413
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] [sig-node] PreStop
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pre_stop.go:172
-[It] should call prestop when killing a pod  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating server pod server in namespace prestop-3260
-STEP: Waiting for pods to come up.
-STEP: Creating tester pod tester in namespace prestop-3260
-STEP: Deleting pre-stop pod
-Aug 30 16:50:52.073: INFO: Saw: {
-	"Hostname": "server",
-	"Sent": null,
-	"Received": {
-		"prestop": 1
-	},
-	"Errors": null,
-	"Log": [
-		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.",
-		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up."
-	],
-	"StillContactingPeers": true
-}
-STEP: Deleting the server pod
-[AfterEach] [k8s.io] [sig-node] PreStop
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:50:52.082: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "prestop-3260" for this suite.
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Feb 12 10:05:26.539: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:05:28.555: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721126, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721126, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721126, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721126, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Feb 12 10:05:31.580: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] listing validating webhooks should work [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Listing all of the created validation webhooks
+STEP: Creating a configMap that does not comply to the validation webhook rules
+STEP: Deleting the collection of validation webhooks
+STEP: Creating a configMap that does not comply to the validation webhook rules
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:05:32.023: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-8413" for this suite.
+STEP: Destroying namespace "webhook-8413-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:9.340 seconds]
-[k8s.io] [sig-node] PreStop
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should call prestop when killing a pod  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:6.219 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  listing validating webhooks should work [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] [sig-node] PreStop should call prestop when killing a pod  [Conformance]","total":280,"completed":58,"skipped":1059,"failed":0}
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance]","total":280,"completed":56,"skipped":782,"failed":0}
 S
 ------------------------------
-[sig-network] DNS 
-  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:50:52.099: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename dns
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-6093
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-6093.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-6093.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6093.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
-
-STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-6093.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-6093.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6093.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
-
-STEP: creating a pod to probe /etc/hosts
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Aug 30 16:51:05.036: INFO: DNS probes using dns-6093/dns-test-841d7211-0b87-45a8-9d4c-344d51dbecd2 succeeded
-
-STEP: deleting the pod
-[AfterEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:51:05.056: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-6093" for this suite.
-
-• [SLOW TEST:12.973 seconds]
-[sig-network] DNS
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] DNS should provide /etc/hosts entries for the cluster [LinuxOnly] [Conformance]","total":280,"completed":59,"skipped":1060,"failed":0}
-SSSSSSSSSSS
-------------------------------
-[sig-apps] ReplicationController 
-  should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:51:05.072: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename replication-controller
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-9123
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating replication controller my-hostname-basic-a7c7870a-1c5d-41cd-815e-4f9b603f1e73
-Aug 30 16:51:05.247: INFO: Pod name my-hostname-basic-a7c7870a-1c5d-41cd-815e-4f9b603f1e73: Found 0 pods out of 1
-Aug 30 16:51:10.254: INFO: Pod name my-hostname-basic-a7c7870a-1c5d-41cd-815e-4f9b603f1e73: Found 1 pods out of 1
-Aug 30 16:51:10.254: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-a7c7870a-1c5d-41cd-815e-4f9b603f1e73" are running
-Aug 30 16:51:10.260: INFO: Pod "my-hostname-basic-a7c7870a-1c5d-41cd-815e-4f9b603f1e73-2b4c6" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-08-30 16:51:05 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-08-30 16:51:07 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-08-30 16:51:07 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-08-30 16:51:05 +0000 UTC Reason: Message:}])
-Aug 30 16:51:10.260: INFO: Trying to dial the pod
-Aug 30 16:51:15.405: INFO: Controller my-hostname-basic-a7c7870a-1c5d-41cd-815e-4f9b603f1e73: Got expected result from replica 1 [my-hostname-basic-a7c7870a-1c5d-41cd-815e-4f9b603f1e73-2b4c6]: "my-hostname-basic-a7c7870a-1c5d-41cd-815e-4f9b603f1e73-2b4c6", 1 of 1 required successes so far
-[AfterEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:51:15.405: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replication-controller-9123" for this suite.
-
-• [SLOW TEST:10.347 seconds]
-[sig-apps] ReplicationController
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] ReplicationController should serve a basic image on each replica with a public image  [Conformance]","total":280,"completed":60,"skipped":1071,"failed":0}
-SSSS
-------------------------------
-[sig-network] Services 
-  should be able to change the type from NodePort to ExternalName [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:51:15.420: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename services
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-6837
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should be able to change the type from NodePort to ExternalName [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a service nodeport-service with the type=NodePort in namespace services-6837
-STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service
-STEP: creating service externalsvc in namespace services-6837
-STEP: creating replication controller externalsvc in namespace services-6837
-I0830 16:51:15.615884      23 runners.go:189] Created replication controller with name: externalsvc, namespace: services-6837, replica count: 2
-I0830 16:51:18.668516      23 runners.go:189] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-STEP: changing the NodePort service to type=ExternalName
-Aug 30 16:51:18.700: INFO: Creating new exec pod
-Aug 30 16:51:20.722: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=services-6837 execpodnnq5g -- /bin/sh -x -c nslookup nodeport-service'
-Aug 30 16:51:21.260: INFO: stderr: "+ nslookup nodeport-service\n"
-Aug 30 16:51:21.260: INFO: stdout: "Server:\t\t10.240.16.10\nAddress:\t10.240.16.10#53\n\nnodeport-service.services-6837.svc.cluster.local\tcanonical name = externalsvc.services-6837.svc.cluster.local.\nName:\texternalsvc.services-6837.svc.cluster.local\nAddress: 10.240.30.4\n\n"
-STEP: deleting ReplicationController externalsvc in namespace services-6837, will wait for the garbage collector to delete the pods
-Aug 30 16:51:21.326: INFO: Deleting ReplicationController externalsvc took: 10.301369ms
-Aug 30 16:51:21.826: INFO: Terminating ReplicationController externalsvc pods took: 500.159597ms
-Aug 30 16:51:34.348: INFO: Cleaning up the NodePort to ExternalName test service
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:51:34.372: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-6837" for this suite.
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
-
-• [SLOW TEST:18.966 seconds]
-[sig-network] Services
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should be able to change the type from NodePort to ExternalName [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance]","total":280,"completed":61,"skipped":1075,"failed":0}
-SSSSSSSSSSSS
-------------------------------
 [sig-storage] Projected downwardAPI 
-  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:51:34.386: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:05:32.143: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1005
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6971
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Creating a pod to test downward API volume plugin
-Aug 30 16:51:34.563: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e6fff232-6368-4e2f-89eb-fb26c6670adf" in namespace "projected-1005" to be "success or failure"
-Aug 30 16:51:34.572: INFO: Pod "downwardapi-volume-e6fff232-6368-4e2f-89eb-fb26c6670adf": Phase="Pending", Reason="", readiness=false. Elapsed: 9.228356ms
-Aug 30 16:51:36.577: INFO: Pod "downwardapi-volume-e6fff232-6368-4e2f-89eb-fb26c6670adf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014290644s
+Feb 12 10:05:32.317: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d721deda-0a1b-4dde-ad38-416af32261ef" in namespace "projected-6971" to be "success or failure"
+Feb 12 10:05:32.324: INFO: Pod "downwardapi-volume-d721deda-0a1b-4dde-ad38-416af32261ef": Phase="Pending", Reason="", readiness=false. Elapsed: 7.153983ms
+Feb 12 10:05:34.331: INFO: Pod "downwardapi-volume-d721deda-0a1b-4dde-ad38-416af32261ef": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014494384s
 STEP: Saw pod success
-Aug 30 16:51:36.577: INFO: Pod "downwardapi-volume-e6fff232-6368-4e2f-89eb-fb26c6670adf" satisfied condition "success or failure"
-Aug 30 16:51:36.581: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-e6fff232-6368-4e2f-89eb-fb26c6670adf container client-container: 
+Feb 12 10:05:34.331: INFO: Pod "downwardapi-volume-d721deda-0a1b-4dde-ad38-416af32261ef" satisfied condition "success or failure"
+Feb 12 10:05:34.337: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-d721deda-0a1b-4dde-ad38-416af32261ef container client-container: 
 STEP: delete the pod
-Aug 30 16:51:36.651: INFO: Waiting for pod downwardapi-volume-e6fff232-6368-4e2f-89eb-fb26c6670adf to disappear
-Aug 30 16:51:36.657: INFO: Pod downwardapi-volume-e6fff232-6368-4e2f-89eb-fb26c6670adf no longer exists
+Feb 12 10:05:34.410: INFO: Waiting for pod downwardapi-volume-d721deda-0a1b-4dde-ad38-416af32261ef to disappear
+Feb 12 10:05:34.418: INFO: Pod downwardapi-volume-d721deda-0a1b-4dde-ad38-416af32261ef no longer exists
 [AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:51:36.657: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-1005" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":62,"skipped":1087,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:05:34.418: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-6971" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":57,"skipped":783,"failed":0}
+SSSSS
 ------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  should have a working scale subresource [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Projected downwardAPI 
+  should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:51:36.673: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename statefulset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-8075
+Feb 12 10:05:34.439: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3394
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
-STEP: Creating service test in namespace statefulset-8075
-[It] should have a working scale subresource [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating statefulset ss in namespace statefulset-8075
-Aug 30 16:51:36.873: INFO: Found 0 stateful pods, waiting for 1
-Aug 30 16:51:46.880: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-STEP: getting scale subresource
-STEP: updating a scale subresource
-STEP: verifying the statefulset Spec.Replicas was modified
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
-Aug 30 16:51:46.909: INFO: Deleting all statefulset in ns statefulset-8075
-Aug 30 16:51:46.914: INFO: Scaling statefulset ss to 0
-Aug 30 16:51:56.980: INFO: Waiting for statefulset status.replicas updated to 0
-Aug 30 16:51:56.986: INFO: Deleting statefulset ss
-[AfterEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:51:57.004: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-8075" for this suite.
-
-• [SLOW TEST:20.345 seconds]
-[sig-apps] StatefulSet
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-    should have a working scale subresource [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance]","total":280,"completed":63,"skipped":1125,"failed":0}
-SSSSSSSSS
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should provide container's memory request [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Feb 12 10:05:34.624: INFO: Waiting up to 5m0s for pod "downwardapi-volume-da3e909c-5edf-48ef-86be-a222a1d79df0" in namespace "projected-3394" to be "success or failure"
+Feb 12 10:05:34.632: INFO: Pod "downwardapi-volume-da3e909c-5edf-48ef-86be-a222a1d79df0": Phase="Pending", Reason="", readiness=false. Elapsed: 7.785942ms
+Feb 12 10:05:36.640: INFO: Pod "downwardapi-volume-da3e909c-5edf-48ef-86be-a222a1d79df0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016032161s
+STEP: Saw pod success
+Feb 12 10:05:36.640: INFO: Pod "downwardapi-volume-da3e909c-5edf-48ef-86be-a222a1d79df0" satisfied condition "success or failure"
+Feb 12 10:05:36.646: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-da3e909c-5edf-48ef-86be-a222a1d79df0 container client-container: 
+STEP: delete the pod
+Feb 12 10:05:36.780: INFO: Waiting for pod downwardapi-volume-da3e909c-5edf-48ef-86be-a222a1d79df0 to disappear
+Feb 12 10:05:36.788: INFO: Pod downwardapi-volume-da3e909c-5edf-48ef-86be-a222a1d79df0 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:05:36.788: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-3394" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance]","total":280,"completed":58,"skipped":788,"failed":0}
+SSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [k8s.io] KubeletManagedEtcHosts 
   should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [k8s.io] KubeletManagedEtcHosts
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:51:57.018: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:05:36.809: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-kubelet-etc-hosts-5565
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-kubelet-etc-hosts-4724
 STEP: Waiting for a default service account to be provisioned in namespace
 [It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Setting up the test
 STEP: Creating hostNetwork=false pod
 STEP: Creating hostNetwork=true pod
 STEP: Running the test
 STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false
-Aug 30 16:52:03.232: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-5565 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:52:03.232: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:52:03.750: INFO: Exec stderr: ""
-Aug 30 16:52:03.750: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-5565 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:52:03.750: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:52:04.279: INFO: Exec stderr: ""
-Aug 30 16:52:04.279: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-5565 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:52:04.280: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:52:04.782: INFO: Exec stderr: ""
-Aug 30 16:52:04.782: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-5565 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:52:04.782: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:52:05.319: INFO: Exec stderr: ""
+Feb 12 10:05:43.049: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4724 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:05:43.049: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:05:43.589: INFO: Exec stderr: ""
+Feb 12 10:05:43.589: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4724 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:05:43.589: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:05:44.197: INFO: Exec stderr: ""
+Feb 12 10:05:44.197: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4724 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:05:44.197: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:05:44.706: INFO: Exec stderr: ""
+Feb 12 10:05:44.706: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4724 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:05:44.706: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:05:45.272: INFO: Exec stderr: ""
 STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount
-Aug 30 16:52:05.319: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-5565 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:52:05.319: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:52:05.848: INFO: Exec stderr: ""
-Aug 30 16:52:05.848: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-5565 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:52:05.848: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:52:06.428: INFO: Exec stderr: ""
+Feb 12 10:05:45.273: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4724 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:05:45.273: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:05:45.871: INFO: Exec stderr: ""
+Feb 12 10:05:45.871: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4724 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:05:45.871: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:05:46.420: INFO: Exec stderr: ""
 STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true
-Aug 30 16:52:06.428: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-5565 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:52:06.428: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:52:06.983: INFO: Exec stderr: ""
-Aug 30 16:52:06.983: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-5565 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:52:06.983: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:52:07.499: INFO: Exec stderr: ""
-Aug 30 16:52:07.499: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-5565 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:52:07.499: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:52:08.003: INFO: Exec stderr: ""
-Aug 30 16:52:08.003: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-5565 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:52:08.003: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:52:08.513: INFO: Exec stderr: ""
+Feb 12 10:05:46.420: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4724 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:05:46.420: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:05:46.935: INFO: Exec stderr: ""
+Feb 12 10:05:46.935: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4724 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:05:46.935: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:05:47.430: INFO: Exec stderr: ""
+Feb 12 10:05:47.430: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-4724 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:05:47.430: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:05:47.997: INFO: Exec stderr: ""
+Feb 12 10:05:47.997: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-4724 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:05:47.997: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:05:48.526: INFO: Exec stderr: ""
 [AfterEach] [k8s.io] KubeletManagedEtcHosts
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:52:08.513: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "e2e-kubelet-etc-hosts-5565" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:05:48.526: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "e2e-kubelet-etc-hosts-4724" for this suite.
 
-• [SLOW TEST:11.509 seconds]
+• [SLOW TEST:11.744 seconds]
 [k8s.io] KubeletManagedEtcHosts
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
   should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":64,"skipped":1134,"failed":0}
-[k8s.io] Docker Containers 
-  should be able to override the image's default command and arguments [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Docker Containers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:52:08.528: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename containers
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-6565
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to override the image's default command and arguments [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test override all
-Aug 30 16:52:08.703: INFO: Waiting up to 5m0s for pod "client-containers-b9dbd7eb-84bd-4cba-91e7-954a85c5c62e" in namespace "containers-6565" to be "success or failure"
-Aug 30 16:52:08.710: INFO: Pod "client-containers-b9dbd7eb-84bd-4cba-91e7-954a85c5c62e": Phase="Pending", Reason="", readiness=false. Elapsed: 7.243744ms
-Aug 30 16:52:10.716: INFO: Pod "client-containers-b9dbd7eb-84bd-4cba-91e7-954a85c5c62e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012844861s
-STEP: Saw pod success
-Aug 30 16:52:10.721: INFO: Pod "client-containers-b9dbd7eb-84bd-4cba-91e7-954a85c5c62e" satisfied condition "success or failure"
-Aug 30 16:52:10.729: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod client-containers-b9dbd7eb-84bd-4cba-91e7-954a85c5c62e container test-container: 
-STEP: delete the pod
-Aug 30 16:52:10.801: INFO: Waiting for pod client-containers-b9dbd7eb-84bd-4cba-91e7-954a85c5c62e to disappear
-Aug 30 16:52:10.806: INFO: Pod client-containers-b9dbd7eb-84bd-4cba-91e7-954a85c5c62e no longer exists
-[AfterEach] [k8s.io] Docker Containers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:52:10.806: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "containers-6565" for this suite.
-•{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance]","total":280,"completed":65,"skipped":1134,"failed":0}
-SSSSSSS
+{"msg":"PASSED [k8s.io] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":59,"skipped":809,"failed":0}
+SSS
 ------------------------------
-[k8s.io] Kubelet when scheduling a read only busybox container 
-  should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-cli] Kubectl client Kubectl rolling-update 
+  should support rolling-update to same image [Deprecated] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:52:10.818: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubelet-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-814
+Feb 12 10:05:48.554: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7225
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
-[It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:52:13.028: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubelet-test-814" for this suite.
-•{"msg":"PASSED [k8s.io] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":66,"skipped":1141,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[BeforeEach] Kubectl rolling-update
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1587
+[It] should support rolling-update to same image [Deprecated] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: running the image docker.io/library/httpd:2.4.38-alpine
+Feb 12 10:05:48.743: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 run e2e-test-httpd-rc --image=docker.io/library/httpd:2.4.38-alpine --generator=run/v1 --namespace=kubectl-7225'
+Feb 12 10:05:48.854: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Feb 12 10:05:48.854: INFO: stdout: "replicationcontroller/e2e-test-httpd-rc created\n"
+STEP: verifying the rc e2e-test-httpd-rc was created
+Feb 12 10:05:48.863: INFO: Waiting for rc e2e-test-httpd-rc to stabilize, generation 1 observed generation 0 spec.replicas 1 status.replicas 0
+Feb 12 10:05:48.866: INFO: Waiting for rc e2e-test-httpd-rc to stabilize, generation 1 observed generation 1 spec.replicas 1 status.replicas 0
+STEP: rolling-update to same image controller
+Feb 12 10:05:48.885: INFO: scanned /root for discovery docs: 
+Feb 12 10:05:48.896: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 rolling-update e2e-test-httpd-rc --update-period=1s --image=docker.io/library/httpd:2.4.38-alpine --image-pull-policy=IfNotPresent --namespace=kubectl-7225'
+Feb 12 10:06:04.761: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n"
+Feb 12 10:06:04.761: INFO: stdout: "Created e2e-test-httpd-rc-8ade455c344d0f0ddf519331ddadfee1\nScaling up e2e-test-httpd-rc-8ade455c344d0f0ddf519331ddadfee1 from 0 to 1, scaling down e2e-test-httpd-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-httpd-rc-8ade455c344d0f0ddf519331ddadfee1 up to 1\nScaling e2e-test-httpd-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-httpd-rc\nRenaming e2e-test-httpd-rc-8ade455c344d0f0ddf519331ddadfee1 to e2e-test-httpd-rc\nreplicationcontroller/e2e-test-httpd-rc rolling updated\n"
+Feb 12 10:06:04.761: INFO: stdout: "Created e2e-test-httpd-rc-8ade455c344d0f0ddf519331ddadfee1\nScaling up e2e-test-httpd-rc-8ade455c344d0f0ddf519331ddadfee1 from 0 to 1, scaling down e2e-test-httpd-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-httpd-rc-8ade455c344d0f0ddf519331ddadfee1 up to 1\nScaling e2e-test-httpd-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-httpd-rc\nRenaming e2e-test-httpd-rc-8ade455c344d0f0ddf519331ddadfee1 to e2e-test-httpd-rc\nreplicationcontroller/e2e-test-httpd-rc rolling updated\n"
+STEP: waiting for all containers in run=e2e-test-httpd-rc pods to come up.
+Feb 12 10:06:04.761: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-httpd-rc --namespace=kubectl-7225'
+Feb 12 10:06:04.847: INFO: stderr: ""
+Feb 12 10:06:04.847: INFO: stdout: "e2e-test-httpd-rc-8ade455c344d0f0ddf519331ddadfee1-g9kcj "
+Feb 12 10:06:04.847: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods e2e-test-httpd-rc-8ade455c344d0f0ddf519331ddadfee1-g9kcj -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "e2e-test-httpd-rc") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-7225'
+Feb 12 10:06:04.926: INFO: stderr: ""
+Feb 12 10:06:04.926: INFO: stdout: "true"
+Feb 12 10:06:04.926: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods e2e-test-httpd-rc-8ade455c344d0f0ddf519331ddadfee1-g9kcj -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "e2e-test-httpd-rc"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-7225'
+Feb 12 10:06:05.014: INFO: stderr: ""
+Feb 12 10:06:05.014: INFO: stdout: "docker.io/library/httpd:2.4.38-alpine"
+Feb 12 10:06:05.014: INFO: e2e-test-httpd-rc-8ade455c344d0f0ddf519331ddadfee1-g9kcj is verified up and running
+[AfterEach] Kubectl rolling-update
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1593
+Feb 12 10:06:05.014: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete rc e2e-test-httpd-rc --namespace=kubectl-7225'
+Feb 12 10:06:05.100: INFO: stderr: ""
+Feb 12 10:06:05.100: INFO: stdout: "replicationcontroller \"e2e-test-httpd-rc\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:06:05.100: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-7225" for this suite.
+
+• [SLOW TEST:16.573 seconds]
+[sig-cli] Kubectl client
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Kubectl rolling-update
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1582
+    should support rolling-update to same image [Deprecated] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-node] Downward API 
-  should provide pod UID as env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] Downward API
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-cli] Kubectl client Kubectl rolling-update should support rolling-update to same image [Deprecated] [Conformance]","total":280,"completed":60,"skipped":812,"failed":0}
+SSSSSS
+------------------------------
+[sig-network] Networking Granular Checks: Pods 
+  should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Networking
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:52:13.044: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-8421
+Feb 12 10:06:05.127: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename pod-network-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-6621
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide pod UID as env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward api env vars
-Aug 30 16:52:13.207: INFO: Waiting up to 5m0s for pod "downward-api-2e855c8a-2429-4f62-a059-9de181b518ca" in namespace "downward-api-8421" to be "success or failure"
-Aug 30 16:52:13.211: INFO: Pod "downward-api-2e855c8a-2429-4f62-a059-9de181b518ca": Phase="Pending", Reason="", readiness=false. Elapsed: 3.9841ms
-Aug 30 16:52:15.219: INFO: Pod "downward-api-2e855c8a-2429-4f62-a059-9de181b518ca": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011110875s
-STEP: Saw pod success
-Aug 30 16:52:15.219: INFO: Pod "downward-api-2e855c8a-2429-4f62-a059-9de181b518ca" satisfied condition "success or failure"
-Aug 30 16:52:15.224: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod downward-api-2e855c8a-2429-4f62-a059-9de181b518ca container dapi-container: 
-STEP: delete the pod
-Aug 30 16:52:15.293: INFO: Waiting for pod downward-api-2e855c8a-2429-4f62-a059-9de181b518ca to disappear
-Aug 30 16:52:15.297: INFO: Pod downward-api-2e855c8a-2429-4f62-a059-9de181b518ca no longer exists
-[AfterEach] [sig-node] Downward API
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:52:15.297: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-8421" for this suite.
-•{"msg":"PASSED [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance]","total":280,"completed":67,"skipped":1171,"failed":0}
-SSSSSSSSSSSSSSSSSSSSS
+[It] should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Performing setup for networking test in namespace pod-network-test-6621
+STEP: creating a selector
+STEP: Creating the service pods in kubernetes
+Feb 12 10:06:05.289: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+STEP: Creating test pods
+Feb 12 10:06:27.437: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.25.0.16:8080/dial?request=hostname&protocol=udp&host=172.25.2.24&port=8081&tries=1'] Namespace:pod-network-test-6621 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:06:27.437: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:06:27.950: INFO: Waiting for responses: map[]
+Feb 12 10:06:27.956: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.25.0.16:8080/dial?request=hostname&protocol=udp&host=172.25.0.15&port=8081&tries=1'] Namespace:pod-network-test-6621 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:06:27.956: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:06:28.494: INFO: Waiting for responses: map[]
+Feb 12 10:06:28.500: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.25.0.16:8080/dial?request=hostname&protocol=udp&host=172.25.1.54&port=8081&tries=1'] Namespace:pod-network-test-6621 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:06:28.500: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:06:29.067: INFO: Waiting for responses: map[]
+[AfterEach] [sig-network] Networking
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:06:29.067: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pod-network-test-6621" for this suite.
+
+• [SLOW TEST:23.961 seconds]
+[sig-network] Networking
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
+  Granular Checks: Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
+    should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class 
-  should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] [sig-node] Pods Extended
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":61,"skipped":818,"failed":0}
+S
+------------------------------
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute poststart http hook properly [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Container Lifecycle Hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:52:15.312: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-5489
+Feb 12 10:06:29.088: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-lifecycle-hook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-9514
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods Set QOS Class
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:178
-[It] should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-STEP: submitting the pod to kubernetes
-STEP: verifying QOS class is set on the pod
-[AfterEach] [k8s.io] [sig-node] Pods Extended
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:52:15.497: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-5489" for this suite.
-•{"msg":"PASSED [k8s.io] [sig-node] Pods Extended [k8s.io] Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]","total":280,"completed":68,"skipped":1192,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute poststart http hook properly [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: create the pod with lifecycle hook
+STEP: check poststart hook
+STEP: delete the pod with lifecycle hook
+Feb 12 10:06:33.367: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Feb 12 10:06:33.374: INFO: Pod pod-with-poststart-http-hook still exists
+Feb 12 10:06:35.374: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Feb 12 10:06:35.380: INFO: Pod pod-with-poststart-http-hook still exists
+Feb 12 10:06:37.374: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Feb 12 10:06:37.393: INFO: Pod pod-with-poststart-http-hook still exists
+Feb 12 10:06:39.374: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Feb 12 10:06:39.380: INFO: Pod pod-with-poststart-http-hook still exists
+Feb 12 10:06:41.374: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Feb 12 10:06:41.380: INFO: Pod pod-with-poststart-http-hook still exists
+Feb 12 10:06:43.374: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
+Feb 12 10:06:43.382: INFO: Pod pod-with-poststart-http-hook no longer exists
+[AfterEach] [k8s.io] Container Lifecycle Hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:06:43.382: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-9514" for this suite.
+
+• [SLOW TEST:14.315 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  when create a pod with lifecycle hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
+    should execute poststart http hook properly [NodeConformance] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Downward API volume 
-  should update labels on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance]","total":280,"completed":62,"skipped":819,"failed":0}
+SSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Garbage collector 
+  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:52:15.513: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7724
+Feb 12 10:06:43.404: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-4220
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should update labels on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating the pod
-Aug 30 16:52:18.225: INFO: Successfully updated pod "labelsupdate57159e48-ae0c-4f9e-bfda-e8de55c6ce6e"
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:52:22.277: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-7724" for this suite.
+[It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: create the deployment
+STEP: Wait for the Deployment to create new ReplicaSet
+STEP: delete the deployment
+STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the rs
+STEP: Gathering metrics
+W0212 10:07:14.141289      21 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Feb 12 10:07:14.141: INFO: For apiserver_request_total:
+For apiserver_request_latency_seconds:
+For apiserver_init_events_total:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
 
-• [SLOW TEST:6.779 seconds]
-[sig-storage] Downward API volume
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:35
-  should update labels on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:07:14.141: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-4220" for this suite.
+
+• [SLOW TEST:30.752 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance]","total":280,"completed":69,"skipped":1215,"failed":0}
-SSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]","total":280,"completed":63,"skipped":833,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] 
-  should be able to convert from CR v1 to CR v2 [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:52:22.293: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename crd-webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-webhook-3986
+Feb 12 10:07:14.156: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-210
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:125
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
 STEP: Setting up server cert
-STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication
-STEP: Deploying the custom resource conversion webhook pod
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Aug 30 16:52:22.920: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set
-Aug 30 16:52:24.937: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403142, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403142, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403142, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403142, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-78dcf5dd84\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:07:14.759: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:07:16.777: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721234, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721234, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721234, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721234, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Aug 30 16:52:27.958: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1
-[It] should be able to convert from CR v1 to CR v2 [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 16:52:27.964: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Creating a v1 custom resource
-STEP: v2 custom resource should be converted
-[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:52:29.570: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-webhook-3986" for this suite.
-[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:136
-
-• [SLOW TEST:7.365 seconds]
-[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should be able to convert from CR v1 to CR v2 [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance]","total":280,"completed":70,"skipped":1230,"failed":0}
-SSSS
-------------------------------
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  works for CRD preserving unknown fields in an embedded object [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:52:29.658: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename crd-publish-openapi
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-1356
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] works for CRD preserving unknown fields in an embedded object [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 16:52:29.816: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: client-side validation (kubectl create and apply) allows request with any unknown properties
-Aug 30 16:52:32.672: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-1356 create -f -'
-Aug 30 16:52:33.059: INFO: stderr: ""
-Aug 30 16:52:33.059: INFO: stdout: "e2e-test-crd-publish-openapi-5414-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n"
-Aug 30 16:52:33.059: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-1356 delete e2e-test-crd-publish-openapi-5414-crds test-cr'
-Aug 30 16:52:33.136: INFO: stderr: ""
-Aug 30 16:52:33.136: INFO: stdout: "e2e-test-crd-publish-openapi-5414-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n"
-Aug 30 16:52:33.136: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-1356 apply -f -'
-Aug 30 16:52:33.453: INFO: stderr: ""
-Aug 30 16:52:33.453: INFO: stdout: "e2e-test-crd-publish-openapi-5414-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n"
-Aug 30 16:52:33.453: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-1356 delete e2e-test-crd-publish-openapi-5414-crds test-cr'
-Aug 30 16:52:33.528: INFO: stderr: ""
-Aug 30 16:52:33.528: INFO: stdout: "e2e-test-crd-publish-openapi-5414-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n"
-STEP: kubectl explain works to explain CR
-Aug 30 16:52:33.528: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 explain e2e-test-crd-publish-openapi-5414-crds'
-Aug 30 16:52:33.664: INFO: stderr: ""
-Aug 30 16:52:33.664: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-5414-crd\nVERSION:  crd-publish-openapi-test-unknown-in-nested.example.com/v1\n\nDESCRIPTION:\n     preserve-unknown-properties in nested field for Testing\n\nFIELDS:\n   apiVersion\t\n     APIVersion defines the versioned schema of this representation of an\n     object. Servers should convert recognized schemas to the latest internal\n     value, and may reject unrecognized values. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n   kind\t\n     Kind is a string value representing the REST resource this object\n     represents. Servers may infer this from the endpoint the client submits\n     requests to. Cannot be updated. In CamelCase. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n   metadata\t\n     Standard object's metadata. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   spec\t\n     Specification of Waldo\n\n   status\t\n     Status of Waldo\n\n"
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:52:36.999: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-1356" for this suite.
+Feb 12 10:07:19.800: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Registering a validating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API
+STEP: Registering a mutating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API
+STEP: Creating a dummy validating-webhook-configuration object
+STEP: Deleting the validating-webhook-configuration, which should be possible to remove
+STEP: Creating a dummy mutating-webhook-configuration object
+STEP: Deleting the mutating-webhook-configuration, which should be possible to remove
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:07:20.405: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-210" for this suite.
+STEP: Destroying namespace "webhook-210-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:7.355 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for CRD preserving unknown fields in an embedded object [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:6.353 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance]","total":280,"completed":71,"skipped":1234,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]","total":280,"completed":64,"skipped":859,"failed":0}
+SSSSSSSSS
 ------------------------------
-[sig-apps] Daemon set [Serial] 
-  should run and stop simple daemon [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] Job 
+  should adopt matching orphans and release non-matching pods [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] Job
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:52:37.013: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename daemonsets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-4831
+Feb 12 10:07:20.512: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename job
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in job-6903
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
-[It] should run and stop simple daemon [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating simple DaemonSet "daemon-set"
-STEP: Check that daemon pods launch on every node of the cluster.
-Aug 30 16:52:37.224: INFO: Number of nodes with available pods: 0
-Aug 30 16:52:37.224: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:38.235: INFO: Number of nodes with available pods: 0
-Aug 30 16:52:38.235: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:39.234: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:39.234: INFO: Node adoring-wozniak-54dcfd79fc-948mf is running more than one daemon pod
-Aug 30 16:52:40.236: INFO: Number of nodes with available pods: 2
-Aug 30 16:52:40.236: INFO: Number of running nodes: 2, number of available pods: 2
-STEP: Stop a daemon pod, check that the daemon pod is revived.
-Aug 30 16:52:40.267: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:40.267: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:41.278: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:41.278: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:42.276: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:42.276: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:43.277: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:43.277: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:44.278: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:44.279: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:45.279: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:45.279: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:46.277: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:46.277: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:47.280: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:47.280: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:48.278: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:48.278: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:49.279: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:49.279: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:50.279: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:50.279: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:51.278: INFO: Number of nodes with available pods: 1
-Aug 30 16:52:51.278: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 16:52:52.279: INFO: Number of nodes with available pods: 2
-Aug 30 16:52:52.279: INFO: Number of running nodes: 2, number of available pods: 2
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-4831, will wait for the garbage collector to delete the pods
-Aug 30 16:52:52.359: INFO: Deleting DaemonSet.extensions daemon-set took: 16.777252ms
-Aug 30 16:52:52.859: INFO: Terminating DaemonSet.extensions daemon-set pods took: 500.225636ms
-Aug 30 16:52:59.967: INFO: Number of nodes with available pods: 0
-Aug 30 16:52:59.967: INFO: Number of running nodes: 0, number of available pods: 0
-Aug 30 16:52:59.971: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-4831/daemonsets","resourceVersion":"10136"},"items":null}
-
-Aug 30 16:52:59.978: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-4831/pods","resourceVersion":"10136"},"items":null}
-
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:53:00.000: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "daemonsets-4831" for this suite.
+[It] should adopt matching orphans and release non-matching pods [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a job
+STEP: Ensuring active pods == parallelism
+STEP: Orphaning one of the Job's Pods
+Feb 12 10:07:25.206: INFO: Successfully updated pod "adopt-release-js5xn"
+STEP: Checking that the Job readopts the Pod
+Feb 12 10:07:25.207: INFO: Waiting up to 15m0s for pod "adopt-release-js5xn" in namespace "job-6903" to be "adopted"
+Feb 12 10:07:25.211: INFO: Pod "adopt-release-js5xn": Phase="Running", Reason="", readiness=true. Elapsed: 4.619389ms
+Feb 12 10:07:27.221: INFO: Pod "adopt-release-js5xn": Phase="Running", Reason="", readiness=true. Elapsed: 2.014211516s
+Feb 12 10:07:27.221: INFO: Pod "adopt-release-js5xn" satisfied condition "adopted"
+STEP: Removing the labels from the Job's Pod
+Feb 12 10:07:27.735: INFO: Successfully updated pod "adopt-release-js5xn"
+STEP: Checking that the Job releases the Pod
+Feb 12 10:07:27.735: INFO: Waiting up to 15m0s for pod "adopt-release-js5xn" in namespace "job-6903" to be "released"
+Feb 12 10:07:27.741: INFO: Pod "adopt-release-js5xn": Phase="Running", Reason="", readiness=true. Elapsed: 5.042083ms
+Feb 12 10:07:29.747: INFO: Pod "adopt-release-js5xn": Phase="Running", Reason="", readiness=true. Elapsed: 2.011095514s
+Feb 12 10:07:29.747: INFO: Pod "adopt-release-js5xn" satisfied condition "released"
+[AfterEach] [sig-apps] Job
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:07:29.747: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "job-6903" for this suite.
 
-• [SLOW TEST:23.000 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should run and stop simple daemon [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:9.270 seconds]
+[sig-apps] Job
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should adopt matching orphans and release non-matching pods [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance]","total":280,"completed":72,"skipped":1288,"failed":0}
-SSS
+{"msg":"PASSED [sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance]","total":280,"completed":65,"skipped":868,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-node] Downward API 
+  should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-node] Downward API
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:53:00.013: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-3569
+Feb 12 10:07:29.782: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-9483
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name configmap-test-volume-41312b37-d091-43bc-b555-b93585e2603f
-STEP: Creating a pod to test consume configMaps
-Aug 30 16:53:00.201: INFO: Waiting up to 5m0s for pod "pod-configmaps-b4b12971-c9a1-4ee2-b6d5-ce8637435789" in namespace "configmap-3569" to be "success or failure"
-Aug 30 16:53:00.212: INFO: Pod "pod-configmaps-b4b12971-c9a1-4ee2-b6d5-ce8637435789": Phase="Pending", Reason="", readiness=false. Elapsed: 10.465585ms
-Aug 30 16:53:02.218: INFO: Pod "pod-configmaps-b4b12971-c9a1-4ee2-b6d5-ce8637435789": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016764774s
+[It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward api env vars
+Feb 12 10:07:29.966: INFO: Waiting up to 5m0s for pod "downward-api-7f59bb49-f88e-4c7a-9c7e-8d9944e0fd25" in namespace "downward-api-9483" to be "success or failure"
+Feb 12 10:07:29.994: INFO: Pod "downward-api-7f59bb49-f88e-4c7a-9c7e-8d9944e0fd25": Phase="Pending", Reason="", readiness=false. Elapsed: 28.165428ms
+Feb 12 10:07:32.000: INFO: Pod "downward-api-7f59bb49-f88e-4c7a-9c7e-8d9944e0fd25": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033916434s
+Feb 12 10:07:34.011: INFO: Pod "downward-api-7f59bb49-f88e-4c7a-9c7e-8d9944e0fd25": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.044922845s
 STEP: Saw pod success
-Aug 30 16:53:02.218: INFO: Pod "pod-configmaps-b4b12971-c9a1-4ee2-b6d5-ce8637435789" satisfied condition "success or failure"
-Aug 30 16:53:02.222: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-configmaps-b4b12971-c9a1-4ee2-b6d5-ce8637435789 container configmap-volume-test: 
+Feb 12 10:07:34.011: INFO: Pod "downward-api-7f59bb49-f88e-4c7a-9c7e-8d9944e0fd25" satisfied condition "success or failure"
+Feb 12 10:07:34.022: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-r6snm pod downward-api-7f59bb49-f88e-4c7a-9c7e-8d9944e0fd25 container dapi-container: 
 STEP: delete the pod
-Aug 30 16:53:02.291: INFO: Waiting for pod pod-configmaps-b4b12971-c9a1-4ee2-b6d5-ce8637435789 to disappear
-Aug 30 16:53:02.294: INFO: Pod pod-configmaps-b4b12971-c9a1-4ee2-b6d5-ce8637435789 no longer exists
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:53:02.294: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-3569" for this suite.
-•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":73,"skipped":1291,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Feb 12 10:07:34.109: INFO: Waiting for pod downward-api-7f59bb49-f88e-4c7a-9c7e-8d9944e0fd25 to disappear
+Feb 12 10:07:34.120: INFO: Pod downward-api-7f59bb49-f88e-4c7a-9c7e-8d9944e0fd25 no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:07:34.121: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-9483" for this suite.
+•{"msg":"PASSED [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]","total":280,"completed":66,"skipped":904,"failed":0}
+SSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  works for multiple CRDs of same group and version but different kinds [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] Aggregator 
+  Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Aggregator
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:53:02.309: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename crd-publish-openapi
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-7933
+Feb 12 10:07:34.141: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename aggregator
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in aggregator-9636
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] works for multiple CRDs of same group and version but different kinds [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: CRs in the same group and version but different kinds (two CRDs) show up in OpenAPI documentation
-Aug 30 16:53:02.475: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:53:05.323: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:53:16.512: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-7933" for this suite.
+[BeforeEach] [sig-api-machinery] Aggregator
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:76
+Feb 12 10:07:34.336: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+[It] Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Registering the sample API server.
+Feb 12 10:07:34.840: INFO: deployment "sample-apiserver-deployment" doesn't have the required revision set
+Feb 12 10:07:36.901: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:07:38.907: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:07:40.908: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:07:42.908: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:07:44.910: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721254, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:07:48.023: INFO: Waited 1.100847703s for the sample-apiserver to be ready to handle requests.
+[AfterEach] [sig-api-machinery] Aggregator
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:67
+[AfterEach] [sig-api-machinery] Aggregator
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:07:49.108: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "aggregator-9636" for this suite.
 
-• [SLOW TEST:14.217 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for multiple CRDs of same group and version but different kinds [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:14.995 seconds]
+[sig-api-machinery] Aggregator
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance]","total":280,"completed":74,"skipped":1322,"failed":0}
-SSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] Aggregator Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]","total":280,"completed":67,"skipped":920,"failed":0}
+SSSSSSSSSSS
 ------------------------------
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with downward pod [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Subpath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:53:16.526: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename subpath
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-8301
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
-STEP: Setting up data
-[It] should support subpaths with downward pod [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod pod-subpath-test-downwardapi-vhzl
-STEP: Creating a pod to test atomic-volume-subpath
-Aug 30 16:53:16.711: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-vhzl" in namespace "subpath-8301" to be "success or failure"
-Aug 30 16:53:16.717: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Pending", Reason="", readiness=false. Elapsed: 5.201288ms
-Aug 30 16:53:18.722: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010600931s
-Aug 30 16:53:20.728: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Running", Reason="", readiness=true. Elapsed: 4.016827865s
-Aug 30 16:53:22.734: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Running", Reason="", readiness=true. Elapsed: 6.02238565s
-Aug 30 16:53:24.741: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Running", Reason="", readiness=true. Elapsed: 8.029293134s
-Aug 30 16:53:26.747: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Running", Reason="", readiness=true. Elapsed: 10.035929937s
-Aug 30 16:53:28.753: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Running", Reason="", readiness=true. Elapsed: 12.041904235s
-Aug 30 16:53:30.759: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Running", Reason="", readiness=true. Elapsed: 14.047499912s
-Aug 30 16:53:32.765: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Running", Reason="", readiness=true. Elapsed: 16.053675131s
-Aug 30 16:53:34.771: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Running", Reason="", readiness=true. Elapsed: 18.059842839s
-Aug 30 16:53:36.776: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Running", Reason="", readiness=true. Elapsed: 20.06495855s
-Aug 30 16:53:38.787: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Running", Reason="", readiness=true. Elapsed: 22.075882696s
-Aug 30 16:53:40.794: INFO: Pod "pod-subpath-test-downwardapi-vhzl": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.083037824s
-STEP: Saw pod success
-Aug 30 16:53:40.794: INFO: Pod "pod-subpath-test-downwardapi-vhzl" satisfied condition "success or failure"
-Aug 30 16:53:40.799: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-subpath-test-downwardapi-vhzl container test-container-subpath-downwardapi-vhzl: 
-STEP: delete the pod
-Aug 30 16:53:40.870: INFO: Waiting for pod pod-subpath-test-downwardapi-vhzl to disappear
-Aug 30 16:53:40.876: INFO: Pod pod-subpath-test-downwardapi-vhzl no longer exists
-STEP: Deleting pod pod-subpath-test-downwardapi-vhzl
-Aug 30 16:53:40.876: INFO: Deleting pod "pod-subpath-test-downwardapi-vhzl" in namespace "subpath-8301"
-[AfterEach] [sig-storage] Subpath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:53:40.881: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "subpath-8301" for this suite.
-
-• [SLOW TEST:24.378 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
-  Atomic writer volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
-    should support subpaths with downward pod [LinuxOnly] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [LinuxOnly] [Conformance]","total":280,"completed":75,"skipped":1340,"failed":0}
-SSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:53:40.905: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-7904
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0777 on tmpfs
-Aug 30 16:53:41.089: INFO: Waiting up to 5m0s for pod "pod-3c8c1b16-2c11-4b59-ad67-ffde9ba8a61e" in namespace "emptydir-7904" to be "success or failure"
-Aug 30 16:53:41.098: INFO: Pod "pod-3c8c1b16-2c11-4b59-ad67-ffde9ba8a61e": Phase="Pending", Reason="", readiness=false. Elapsed: 8.896854ms
-Aug 30 16:53:43.103: INFO: Pod "pod-3c8c1b16-2c11-4b59-ad67-ffde9ba8a61e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014370293s
-Aug 30 16:53:45.110: INFO: Pod "pod-3c8c1b16-2c11-4b59-ad67-ffde9ba8a61e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021204517s
-STEP: Saw pod success
-Aug 30 16:53:45.110: INFO: Pod "pod-3c8c1b16-2c11-4b59-ad67-ffde9ba8a61e" satisfied condition "success or failure"
-Aug 30 16:53:45.115: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-3c8c1b16-2c11-4b59-ad67-ffde9ba8a61e container test-container: 
-STEP: delete the pod
-Aug 30 16:53:45.184: INFO: Waiting for pod pod-3c8c1b16-2c11-4b59-ad67-ffde9ba8a61e to disappear
-Aug 30 16:53:45.188: INFO: Pod pod-3c8c1b16-2c11-4b59-ad67-ffde9ba8a61e no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:53:45.188: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-7904" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":76,"skipped":1360,"failed":0}
-S
-------------------------------
-[k8s.io] Security Context When creating a container with runAsUser 
-  should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:53:45.207: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename security-context-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in security-context-test-6148
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39
-[It] should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 16:53:45.382: INFO: Waiting up to 5m0s for pod "busybox-user-65534-a2a3c968-1e32-48f6-b0ed-6abaa2eb856a" in namespace "security-context-test-6148" to be "success or failure"
-Aug 30 16:53:45.389: INFO: Pod "busybox-user-65534-a2a3c968-1e32-48f6-b0ed-6abaa2eb856a": Phase="Pending", Reason="", readiness=false. Elapsed: 6.911547ms
-Aug 30 16:53:47.394: INFO: Pod "busybox-user-65534-a2a3c968-1e32-48f6-b0ed-6abaa2eb856a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012328334s
-Aug 30 16:53:47.394: INFO: Pod "busybox-user-65534-a2a3c968-1e32-48f6-b0ed-6abaa2eb856a" satisfied condition "success or failure"
-[AfterEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:53:47.394: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "security-context-test-6148" for this suite.
-•{"msg":"PASSED [k8s.io] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":77,"skipped":1361,"failed":0}
-SS
-------------------------------
-[sig-network] Networking Granular Checks: Pods 
-  should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Networking
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:53:47.406: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename pod-network-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-7478
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Performing setup for networking test in namespace pod-network-test-7478
-STEP: creating a selector
-STEP: Creating the service pods in kubernetes
-Aug 30 16:53:47.569: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
-STEP: Creating test pods
-Aug 30 16:54:09.664: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.25.0.74:8080/dial?request=hostname&protocol=udp&host=172.25.1.37&port=8081&tries=1'] Namespace:pod-network-test-7478 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:54:09.664: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:54:10.149: INFO: Waiting for responses: map[]
-Aug 30 16:54:10.155: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.25.0.74:8080/dial?request=hostname&protocol=udp&host=172.25.0.73&port=8081&tries=1'] Namespace:pod-network-test-7478 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:54:10.155: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:54:10.667: INFO: Waiting for responses: map[]
-[AfterEach] [sig-network] Networking
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:54:10.667: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pod-network-test-7478" for this suite.
-
-• [SLOW TEST:23.277 seconds]
-[sig-network] Networking
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
-  Granular Checks: Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
-    should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":78,"skipped":1363,"failed":0}
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  removes definition from spec when one version gets changed to not be served [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:54:10.684: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename crd-publish-openapi
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-5806
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] removes definition from spec when one version gets changed to not be served [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: set up a multi version CRD
-Aug 30 16:54:10.842: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: mark a version not serverd
-STEP: check the unserved version gets removed
-STEP: check the other version is not changed
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:54:26.561: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-5806" for this suite.
-
-• [SLOW TEST:15.890 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  removes definition from spec when one version gets changed to not be served [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance]","total":280,"completed":79,"skipped":1363,"failed":0}
-[k8s.io] Probing container 
-  should have monotonically increasing restart count [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Pods 
+  should be submitted and removed [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:54:26.574: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-7936
+Feb 12 10:07:49.138: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-7645
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] should have monotonically increasing restart count [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod liveness-bad71537-b448-4b23-8a08-43f9759a6ea4 in namespace container-probe-7936
-Aug 30 16:54:28.763: INFO: Started pod liveness-bad71537-b448-4b23-8a08-43f9759a6ea4 in namespace container-probe-7936
-STEP: checking the pod's current state and verifying that restartCount is present
-Aug 30 16:54:28.768: INFO: Initial restart count of pod liveness-bad71537-b448-4b23-8a08-43f9759a6ea4 is 0
-Aug 30 16:54:38.809: INFO: Restart count of pod container-probe-7936/liveness-bad71537-b448-4b23-8a08-43f9759a6ea4 is now 1 (10.041506124s elapsed)
-Aug 30 16:55:00.874: INFO: Restart count of pod container-probe-7936/liveness-bad71537-b448-4b23-8a08-43f9759a6ea4 is now 2 (32.105789395s elapsed)
-Aug 30 16:55:20.938: INFO: Restart count of pod container-probe-7936/liveness-bad71537-b448-4b23-8a08-43f9759a6ea4 is now 3 (52.170517379s elapsed)
-Aug 30 16:55:40.998: INFO: Restart count of pod container-probe-7936/liveness-bad71537-b448-4b23-8a08-43f9759a6ea4 is now 4 (1m12.229727837s elapsed)
-Aug 30 16:56:39.186: INFO: Restart count of pod container-probe-7936/liveness-bad71537-b448-4b23-8a08-43f9759a6ea4 is now 5 (2m10.417823754s elapsed)
-STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:56:39.200: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-7936" for this suite.
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
+[It] should be submitted and removed [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating the pod
+STEP: setting up watch
+STEP: submitting the pod to kubernetes
+Feb 12 10:07:49.358: INFO: observed the pod list
+STEP: verifying the pod is in kubernetes
+STEP: verifying pod creation was observed
+STEP: deleting the pod gracefully
+STEP: verifying the kubelet observed the termination notice
+STEP: verifying pod deletion was observed
+[AfterEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:08:02.208: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-7645" for this suite.
 
-• [SLOW TEST:132.640 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should have monotonically increasing restart count [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [k8s.io] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance]","total":280,"completed":80,"skipped":1363,"failed":0}
-SSSSS
+• [SLOW TEST:13.086 seconds]
+[k8s.io] Pods
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should be submitted and removed [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should update labels on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:56:39.214: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3522
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should update labels on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating the pod
-Aug 30 16:56:41.932: INFO: Successfully updated pod "labelsupdate035d0125-60f0-4cf0-b1d0-576719b51a12"
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:56:43.963: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-3522" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance]","total":280,"completed":81,"skipped":1368,"failed":0}
-SSS
+{"msg":"PASSED [k8s.io] Pods should be submitted and removed [NodeConformance] [Conformance]","total":280,"completed":68,"skipped":931,"failed":0}
+SSSSSS
 ------------------------------
 [sig-storage] Downward API volume 
-  should provide podname only [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:56:43.979: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:08:02.224: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-6565
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-3808
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should provide podname only [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Creating a pod to test downward API volume plugin
-Aug 30 16:56:44.143: INFO: Waiting up to 5m0s for pod "downwardapi-volume-87f79ca0-3d98-4a27-b6b5-47bb123e0866" in namespace "downward-api-6565" to be "success or failure"
-Aug 30 16:56:44.149: INFO: Pod "downwardapi-volume-87f79ca0-3d98-4a27-b6b5-47bb123e0866": Phase="Pending", Reason="", readiness=false. Elapsed: 5.561741ms
-Aug 30 16:56:46.155: INFO: Pod "downwardapi-volume-87f79ca0-3d98-4a27-b6b5-47bb123e0866": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01125812s
-Aug 30 16:56:48.161: INFO: Pod "downwardapi-volume-87f79ca0-3d98-4a27-b6b5-47bb123e0866": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017792728s
+Feb 12 10:08:02.404: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9141b8e7-8f4c-418c-b926-9cb977da304b" in namespace "downward-api-3808" to be "success or failure"
+Feb 12 10:08:02.411: INFO: Pod "downwardapi-volume-9141b8e7-8f4c-418c-b926-9cb977da304b": Phase="Pending", Reason="", readiness=false. Elapsed: 7.479887ms
+Feb 12 10:08:04.417: INFO: Pod "downwardapi-volume-9141b8e7-8f4c-418c-b926-9cb977da304b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012995023s
+Feb 12 10:08:06.426: INFO: Pod "downwardapi-volume-9141b8e7-8f4c-418c-b926-9cb977da304b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022230034s
 STEP: Saw pod success
-Aug 30 16:56:48.161: INFO: Pod "downwardapi-volume-87f79ca0-3d98-4a27-b6b5-47bb123e0866" satisfied condition "success or failure"
-Aug 30 16:56:48.166: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-87f79ca0-3d98-4a27-b6b5-47bb123e0866 container client-container: 
+Feb 12 10:08:06.426: INFO: Pod "downwardapi-volume-9141b8e7-8f4c-418c-b926-9cb977da304b" satisfied condition "success or failure"
+Feb 12 10:08:06.433: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-r6snm pod downwardapi-volume-9141b8e7-8f4c-418c-b926-9cb977da304b container client-container: 
 STEP: delete the pod
-Aug 30 16:56:48.232: INFO: Waiting for pod downwardapi-volume-87f79ca0-3d98-4a27-b6b5-47bb123e0866 to disappear
-Aug 30 16:56:48.236: INFO: Pod downwardapi-volume-87f79ca0-3d98-4a27-b6b5-47bb123e0866 no longer exists
+Feb 12 10:08:06.507: INFO: Waiting for pod downwardapi-volume-9141b8e7-8f4c-418c-b926-9cb977da304b to disappear
+Feb 12 10:08:06.520: INFO: Pod downwardapi-volume-9141b8e7-8f4c-418c-b926-9cb977da304b no longer exists
 [AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:56:48.236: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-6565" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance]","total":280,"completed":82,"skipped":1371,"failed":0}
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:08:06.520: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-3808" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":69,"skipped":937,"failed":0}
+SSSSSSSSSSSSS
+------------------------------
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates that NodeSelector is respected if matching  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:08:06.539: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename sched-pred
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-4602
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
+Feb 12 10:08:06.770: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Feb 12 10:08:06.798: INFO: Waiting for terminating namespaces to be deleted...
+Feb 12 10:08:06.804: INFO: 
+Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-dbhjj before test
+Feb 12 10:08:06.861: INFO: node-local-dns-k2khb from kube-system started at 2021-02-12 09:48:36 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.861: INFO: 	Container node-cache ready: true, restart count 0
+Feb 12 10:08:06.861: INFO: logrotate-4f9dq from kube-system started at 2021-02-12 09:48:36 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.861: INFO: 	Container logrotate ready: true, restart count 0
+Feb 12 10:08:06.861: INFO: adopt-release-pvzhv from job-6903 started at 2021-02-12 10:07:20 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.861: INFO: 	Container c ready: false, restart count 0
+Feb 12 10:08:06.861: INFO: user-ssh-keys-agent-8lrpd from kube-system started at 2021-02-12 09:48:16 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.861: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
+Feb 12 10:08:06.861: INFO: kube-proxy-c88c9 from kube-system started at 2021-02-12 09:48:16 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.861: INFO: 	Container kube-proxy ready: true, restart count 0
+Feb 12 10:08:06.861: INFO: canal-k48gh from kube-system started at 2021-02-12 09:48:16 +0000 UTC (2 container statuses recorded)
+Feb 12 10:08:06.861: INFO: 	Container calico-node ready: true, restart count 0
+Feb 12 10:08:06.861: INFO: 	Container kube-flannel ready: true, restart count 0
+Feb 12 10:08:06.861: INFO: sonobuoy-e2e-job-690c418daa0f4403 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded)
+Feb 12 10:08:06.861: INFO: 	Container e2e ready: true, restart count 0
+Feb 12 10:08:06.861: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Feb 12 10:08:06.861: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-vkql5 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded)
+Feb 12 10:08:06.861: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Feb 12 10:08:06.861: INFO: 	Container systemd-logs ready: true, restart count 0
+Feb 12 10:08:06.861: INFO: 
+Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-r6snm before test
+Feb 12 10:08:06.937: INFO: node-local-dns-fkrhg from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.937: INFO: 	Container node-cache ready: true, restart count 0
+Feb 12 10:08:06.937: INFO: logrotate-7vwrm from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.937: INFO: 	Container logrotate ready: true, restart count 0
+Feb 12 10:08:06.937: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-zp8g7 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded)
+Feb 12 10:08:06.937: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Feb 12 10:08:06.937: INFO: 	Container systemd-logs ready: true, restart count 0
+Feb 12 10:08:06.937: INFO: kube-proxy-gfdtg from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.937: INFO: 	Container kube-proxy ready: true, restart count 0
+Feb 12 10:08:06.937: INFO: user-ssh-keys-agent-vxhth from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.937: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
+Feb 12 10:08:06.937: INFO: dashboard-metrics-scraper-59bfc65dc9-rhjf7 from kubernetes-dashboard started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.937: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
+Feb 12 10:08:06.937: INFO: dashboard-metrics-scraper-59bfc65dc9-d8l4b from kubernetes-dashboard started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.937: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
+Feb 12 10:08:06.938: INFO: canal-b9q6f from kube-system started at 2021-02-12 09:48:02 +0000 UTC (2 container statuses recorded)
+Feb 12 10:08:06.938: INFO: 	Container calico-node ready: true, restart count 0
+Feb 12 10:08:06.938: INFO: 	Container kube-flannel ready: true, restart count 0
+Feb 12 10:08:06.938: INFO: sonobuoy from sonobuoy started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.938: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Feb 12 10:08:06.938: INFO: coredns-6449d494cc-psw4s from kube-system started at 2021-02-12 09:48:21 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:06.938: INFO: 	Container coredns ready: true, restart count 0
+Feb 12 10:08:06.938: INFO: 
+Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-xg74t before test
+Feb 12 10:08:07.004: INFO: user-ssh-keys-agent-kc6zq from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:07.004: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
+Feb 12 10:08:07.004: INFO: logrotate-s9rq2 from kube-system started at 2021-02-12 09:59:42 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:07.004: INFO: 	Container logrotate ready: true, restart count 0
+Feb 12 10:08:07.004: INFO: coredns-6449d494cc-b78nb from kube-system started at 2021-02-12 09:59:35 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:07.004: INFO: 	Container coredns ready: true, restart count 0
+Feb 12 10:08:07.004: INFO: node-local-dns-rbbms from kube-system started at 2021-02-12 09:59:34 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:07.004: INFO: 	Container node-cache ready: true, restart count 0
+Feb 12 10:08:07.004: INFO: kube-proxy-4zfp6 from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:07.005: INFO: 	Container kube-proxy ready: true, restart count 0
+Feb 12 10:08:07.005: INFO: canal-ww2cz from kube-system started at 2021-02-12 09:48:02 +0000 UTC (2 container statuses recorded)
+Feb 12 10:08:07.005: INFO: 	Container calico-node ready: true, restart count 0
+Feb 12 10:08:07.005: INFO: 	Container kube-flannel ready: true, restart count 0
+Feb 12 10:08:07.005: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-wbbpd from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded)
+Feb 12 10:08:07.005: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Feb 12 10:08:07.005: INFO: 	Container systemd-logs ready: true, restart count 0
+Feb 12 10:08:07.005: INFO: adopt-release-js5xn from job-6903 started at 2021-02-12 10:07:20 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:07.005: INFO: 	Container c ready: false, restart count 0
+Feb 12 10:08:07.005: INFO: openvpn-client-78d595f58b-8b4r4 from kube-system started at 2021-02-12 09:59:34 +0000 UTC (2 container statuses recorded)
+Feb 12 10:08:07.005: INFO: 	Container dnat-controller ready: true, restart count 0
+Feb 12 10:08:07.005: INFO: 	Container openvpn-client ready: true, restart count 0
+Feb 12 10:08:07.005: INFO: adopt-release-kdt42 from job-6903 started at 2021-02-12 10:07:27 +0000 UTC (1 container statuses recorded)
+Feb 12 10:08:07.005: INFO: 	Container c ready: false, restart count 0
+[It] validates that NodeSelector is respected if matching  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Trying to launch a pod without a label to get a node which can launch it.
+STEP: Explicitly delete pod here to free the resource it takes.
+STEP: Trying to apply a random label on the found node.
+STEP: verifying the node has the label kubernetes.io/e2e-41925020-4d8e-4c8c-90d5-b446f55a5da3 42
+STEP: Trying to relaunch the pod, now with labels.
+STEP: removing the label kubernetes.io/e2e-41925020-4d8e-4c8c-90d5-b446f55a5da3 off the node nostalgic-fermat-cdc5d8777-xg74t
+STEP: verifying the node doesn't have the label kubernetes.io/e2e-41925020-4d8e-4c8c-90d5-b446f55a5da3
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:08:15.123: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-4602" for this suite.
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
 
+• [SLOW TEST:8.601 seconds]
+[sig-scheduling] SchedulerPredicates [Serial]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
+  validates that NodeSelector is respected if matching  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Projected configMap 
+{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching  [Conformance]","total":280,"completed":70,"skipped":950,"failed":0}
+[sig-storage] ConfigMap 
   should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:56:48.250: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8356
+Feb 12 10:08:15.140: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-881
 STEP: Waiting for a default service account to be provisioned in namespace
 [It] should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name projected-configmap-test-volume-ec4477c7-364b-4b83-b066-71a96b830c06
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name configmap-test-volume-f6b3488e-de5c-4a9f-b763-eeb2787a263c
 STEP: Creating a pod to test consume configMaps
-Aug 30 16:56:48.424: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-23666c77-0c71-4837-9d1b-abccaf248e7c" in namespace "projected-8356" to be "success or failure"
-Aug 30 16:56:48.429: INFO: Pod "pod-projected-configmaps-23666c77-0c71-4837-9d1b-abccaf248e7c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.91345ms
-Aug 30 16:56:50.434: INFO: Pod "pod-projected-configmaps-23666c77-0c71-4837-9d1b-abccaf248e7c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010011039s
-Aug 30 16:56:52.440: INFO: Pod "pod-projected-configmaps-23666c77-0c71-4837-9d1b-abccaf248e7c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015977144s
+Feb 12 10:08:15.348: INFO: Waiting up to 5m0s for pod "pod-configmaps-f3d1b080-cac7-4f41-ab8b-f833deadc3c4" in namespace "configmap-881" to be "success or failure"
+Feb 12 10:08:15.356: INFO: Pod "pod-configmaps-f3d1b080-cac7-4f41-ab8b-f833deadc3c4": Phase="Pending", Reason="", readiness=false. Elapsed: 7.342694ms
+Feb 12 10:08:17.367: INFO: Pod "pod-configmaps-f3d1b080-cac7-4f41-ab8b-f833deadc3c4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018439857s
+Feb 12 10:08:19.374: INFO: Pod "pod-configmaps-f3d1b080-cac7-4f41-ab8b-f833deadc3c4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025353159s
 STEP: Saw pod success
-Aug 30 16:56:52.440: INFO: Pod "pod-projected-configmaps-23666c77-0c71-4837-9d1b-abccaf248e7c" satisfied condition "success or failure"
-Aug 30 16:56:52.445: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-projected-configmaps-23666c77-0c71-4837-9d1b-abccaf248e7c container projected-configmap-volume-test: 
+Feb 12 10:08:19.374: INFO: Pod "pod-configmaps-f3d1b080-cac7-4f41-ab8b-f833deadc3c4" satisfied condition "success or failure"
+Feb 12 10:08:19.382: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-configmaps-f3d1b080-cac7-4f41-ab8b-f833deadc3c4 container configmap-volume-test: 
 STEP: delete the pod
-Aug 30 16:56:52.474: INFO: Waiting for pod pod-projected-configmaps-23666c77-0c71-4837-9d1b-abccaf248e7c to disappear
-Aug 30 16:56:52.478: INFO: Pod pod-projected-configmaps-23666c77-0c71-4837-9d1b-abccaf248e7c no longer exists
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:56:52.478: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-8356" for this suite.
-•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":83,"skipped":1371,"failed":0}
-SSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-storage] EmptyDir volumes 
-  pod should support shared volumes between containers [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:56:52.493: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-213
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] pod should support shared volumes between containers [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating Pod
-STEP: Waiting for the pod running
-STEP: Geting the pod
-STEP: Reading file content from the nginx-container
-Aug 30 16:56:54.690: INFO: ExecWithOptions {Command:[/bin/sh -c cat /usr/share/volumeshare/shareddata.txt] Namespace:emptydir-213 PodName:pod-sharedvolume-29076d9e-a2fa-414f-8d5e-a8cd1624f8eb ContainerName:busybox-main-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 16:56:54.690: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 16:56:55.192: INFO: Exec stderr: ""
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:56:55.192: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-213" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance]","total":280,"completed":84,"skipped":1391,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Feb 12 10:08:19.470: INFO: Waiting for pod pod-configmaps-f3d1b080-cac7-4f41-ab8b-f833deadc3c4 to disappear
+Feb 12 10:08:19.475: INFO: Pod pod-configmaps-f3d1b080-cac7-4f41-ab8b-f833deadc3c4 no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:08:19.475: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-881" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":71,"skipped":950,"failed":0}
+SSS
 ------------------------------
-[sig-network] DNS 
-  should provide DNS for services  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:56:55.210: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename dns
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-9437
+Feb 12 10:08:19.491: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-6914
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide DNS for services  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a test headless service
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-9437.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-9437.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-9437.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-9437.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-9437.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-9437.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-9437.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-9437.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-9437.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-9437.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-9437.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-9437.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-9437.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 26.23.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.23.26_udp@PTR;check="$$(dig +tcp +noall +answer +search 26.23.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.23.26_tcp@PTR;sleep 1; done
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
+STEP: Creating service test in namespace statefulset-6914
+[It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Initializing watcher for selector baz=blah,foo=bar
+STEP: Creating stateful set ss in namespace statefulset-6914
+STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-6914
+Feb 12 10:08:19.710: INFO: Found 0 stateful pods, waiting for 1
+Feb 12 10:08:29.720: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod
+Feb 12 10:08:29.727: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Feb 12 10:08:30.451: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Feb 12 10:08:30.451: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Feb 12 10:08:30.451: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Feb 12 10:08:30.474: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
+Feb 12 10:08:40.481: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Feb 12 10:08:40.481: INFO: Waiting for statefulset status.replicas updated to 0
+Feb 12 10:08:40.511: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999999717s
+Feb 12 10:08:41.517: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.99352583s
+Feb 12 10:08:42.522: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.987507159s
+Feb 12 10:08:43.533: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.981751788s
+Feb 12 10:08:44.541: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.970991779s
+Feb 12 10:08:45.547: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.963534458s
+Feb 12 10:08:46.554: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.957081818s
+Feb 12 10:08:47.561: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.950063255s
+Feb 12 10:08:48.566: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.943535048s
+Feb 12 10:08:49.576: INFO: Verifying statefulset ss doesn't scale past 1 for another 937.602586ms
+STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-6914
+Feb 12 10:08:50.583: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:08:51.149: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
+Feb 12 10:08:51.149: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
+Feb 12 10:08:51.149: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
+
+Feb 12 10:08:51.157: INFO: Found 1 stateful pods, waiting for 3
+Feb 12 10:09:01.165: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+Feb 12 10:09:01.165: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
+Feb 12 10:09:01.165: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Verifying that stateful set ss was scaled up in order
+STEP: Scale down will halt with unhealthy stateful pod
+Feb 12 10:09:01.174: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Feb 12 10:09:01.746: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Feb 12 10:09:01.746: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Feb 12 10:09:01.746: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Feb 12 10:09:01.746: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Feb 12 10:09:02.324: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Feb 12 10:09:02.324: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Feb 12 10:09:02.324: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Feb 12 10:09:02.324: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Feb 12 10:09:02.968: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Feb 12 10:09:02.968: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Feb 12 10:09:02.968: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Feb 12 10:09:02.968: INFO: Waiting for statefulset status.replicas updated to 0
+Feb 12 10:09:02.974: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 1
+Feb 12 10:09:12.990: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
+Feb 12 10:09:12.990: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
+Feb 12 10:09:12.990: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
+Feb 12 10:09:13.018: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999999798s
+Feb 12 10:09:14.030: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.985414853s
+Feb 12 10:09:15.038: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.972730959s
+Feb 12 10:09:16.047: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.965270509s
+Feb 12 10:09:17.054: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.956517452s
+Feb 12 10:09:18.063: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.948609115s
+Feb 12 10:09:19.070: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.940325163s
+Feb 12 10:09:20.077: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.933217497s
+Feb 12 10:09:21.085: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.925789197s
+Feb 12 10:09:22.091: INFO: Verifying statefulset ss doesn't scale past 3 for another 918.359492ms
+STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-6914
+Feb 12 10:09:23.098: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:09:23.644: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
+Feb 12 10:09:23.645: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
+Feb 12 10:09:23.645: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
+
+Feb 12 10:09:23.645: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:09:24.245: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
+Feb 12 10:09:24.245: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
+Feb 12 10:09:24.245: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
+
+Feb 12 10:09:24.245: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:09:24.746: INFO: rc: 1
+Feb 12 10:09:24.746: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
 
-STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-9437.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-9437.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-9437.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-9437.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-9437.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-9437.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-9437.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-9437.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-9437.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-9437.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-9437.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-9437.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-9437.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 26.23.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.23.26_udp@PTR;check="$$(dig +tcp +noall +answer +search 26.23.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.23.26_tcp@PTR;sleep 1; done
+stderr:
+error: Internal error occurred: error executing command in container: container not running (719349eaf1276fa9d99cef51ccf0f0296429d44ca432e4890ff46e03232c6263)
 
-STEP: creating a pod to probe DNS
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Aug 30 16:56:59.577: INFO: Unable to read wheezy_udp@dns-test-service.dns-9437.svc.cluster.local from pod dns-9437/dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3: the server could not find the requested resource (get pods dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3)
-Aug 30 16:56:59.623: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9437.svc.cluster.local from pod dns-9437/dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3: the server could not find the requested resource (get pods dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3)
-Aug 30 16:56:59.632: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9437.svc.cluster.local from pod dns-9437/dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3: the server could not find the requested resource (get pods dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3)
-Aug 30 16:56:59.641: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9437.svc.cluster.local from pod dns-9437/dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3: the server could not find the requested resource (get pods dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3)
-Aug 30 16:57:00.170: INFO: Unable to read jessie_udp@dns-test-service.dns-9437.svc.cluster.local from pod dns-9437/dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3: the server could not find the requested resource (get pods dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3)
-Aug 30 16:57:00.179: INFO: Unable to read jessie_tcp@dns-test-service.dns-9437.svc.cluster.local from pod dns-9437/dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3: the server could not find the requested resource (get pods dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3)
-Aug 30 16:57:00.188: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9437.svc.cluster.local from pod dns-9437/dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3: the server could not find the requested resource (get pods dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3)
-Aug 30 16:57:00.197: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9437.svc.cluster.local from pod dns-9437/dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3: the server could not find the requested resource (get pods dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3)
-Aug 30 16:57:00.630: INFO: Lookups using dns-9437/dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3 failed for: [wheezy_udp@dns-test-service.dns-9437.svc.cluster.local wheezy_tcp@dns-test-service.dns-9437.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-9437.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-9437.svc.cluster.local jessie_udp@dns-test-service.dns-9437.svc.cluster.local jessie_tcp@dns-test-service.dns-9437.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-9437.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-9437.svc.cluster.local]
-
-Aug 30 16:57:07.214: INFO: DNS probes using dns-9437/dns-test-2d85c08f-4123-4145-baff-a2f52667e8e3 succeeded
+error:
+exit status 1
+Feb 12 10:09:34.747: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:09:34.836: INFO: rc: 1
+Feb 12 10:09:34.836: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
 
-STEP: deleting the pod
-STEP: deleting the test service
-STEP: deleting the test headless service
-[AfterEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:57:07.295: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-9437" for this suite.
+stderr:
+Error from server (NotFound): pods "ss-2" not found
 
-• [SLOW TEST:12.099 seconds]
-[sig-network] DNS
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should provide DNS for services  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] DNS should provide DNS for services  [Conformance]","total":280,"completed":85,"skipped":1421,"failed":0}
-SSSSSSS
-------------------------------
-[sig-storage] EmptyDir volumes 
-  volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:57:07.310: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-8848
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir volume type on node default medium
-Aug 30 16:57:07.477: INFO: Waiting up to 5m0s for pod "pod-cbfd0dca-9d4f-4890-8f12-4ca438b14997" in namespace "emptydir-8848" to be "success or failure"
-Aug 30 16:57:07.485: INFO: Pod "pod-cbfd0dca-9d4f-4890-8f12-4ca438b14997": Phase="Pending", Reason="", readiness=false. Elapsed: 7.375665ms
-Aug 30 16:57:09.491: INFO: Pod "pod-cbfd0dca-9d4f-4890-8f12-4ca438b14997": Phase="Running", Reason="", readiness=true. Elapsed: 2.013221069s
-Aug 30 16:57:11.496: INFO: Pod "pod-cbfd0dca-9d4f-4890-8f12-4ca438b14997": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.019149205s
-STEP: Saw pod success
-Aug 30 16:57:11.496: INFO: Pod "pod-cbfd0dca-9d4f-4890-8f12-4ca438b14997" satisfied condition "success or failure"
-Aug 30 16:57:11.501: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-cbfd0dca-9d4f-4890-8f12-4ca438b14997 container test-container: 
-STEP: delete the pod
-Aug 30 16:57:11.571: INFO: Waiting for pod pod-cbfd0dca-9d4f-4890-8f12-4ca438b14997 to disappear
-Aug 30 16:57:11.575: INFO: Pod pod-cbfd0dca-9d4f-4890-8f12-4ca438b14997 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:57:11.576: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-8848" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":86,"skipped":1428,"failed":0}
-SSSSSSSS
-------------------------------
-[sig-cli] Kubectl client Kubectl run rc 
-  should create an rc from an image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:57:11.590: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2592
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[BeforeEach] Kubectl run rc
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1525
-[It] should create an rc from an image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: running the image docker.io/library/httpd:2.4.38-alpine
-Aug 30 16:57:11.750: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 run e2e-test-httpd-rc --image=docker.io/library/httpd:2.4.38-alpine --generator=run/v1 --namespace=kubectl-2592'
-Aug 30 16:57:11.825: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-Aug 30 16:57:11.825: INFO: stdout: "replicationcontroller/e2e-test-httpd-rc created\n"
-STEP: verifying the rc e2e-test-httpd-rc was created
-STEP: verifying the pod controlled by rc e2e-test-httpd-rc was created
-STEP: confirm that you can get logs from an rc
-Aug 30 16:57:11.839: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [e2e-test-httpd-rc-k2dhv]
-Aug 30 16:57:11.839: INFO: Waiting up to 5m0s for pod "e2e-test-httpd-rc-k2dhv" in namespace "kubectl-2592" to be "running and ready"
-Aug 30 16:57:11.843: INFO: Pod "e2e-test-httpd-rc-k2dhv": Phase="Pending", Reason="", readiness=false. Elapsed: 3.772084ms
-Aug 30 16:57:13.848: INFO: Pod "e2e-test-httpd-rc-k2dhv": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008437497s
-Aug 30 16:57:15.853: INFO: Pod "e2e-test-httpd-rc-k2dhv": Phase="Running", Reason="", readiness=true. Elapsed: 4.014136885s
-Aug 30 16:57:15.853: INFO: Pod "e2e-test-httpd-rc-k2dhv" satisfied condition "running and ready"
-Aug 30 16:57:15.853: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [e2e-test-httpd-rc-k2dhv]
-Aug 30 16:57:15.853: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 logs rc/e2e-test-httpd-rc --namespace=kubectl-2592'
-Aug 30 16:57:16.076: INFO: stderr: ""
-Aug 30 16:57:16.076: INFO: stdout: "AH00558: httpd: Could not reliably determine the server's fully qualified domain name, using 172.25.0.81. Set the 'ServerName' directive globally to suppress this message\nAH00558: httpd: Could not reliably determine the server's fully qualified domain name, using 172.25.0.81. Set the 'ServerName' directive globally to suppress this message\n[Sun Aug 30 16:57:13.068258 2020] [mpm_event:notice] [pid 1:tid 139877614336872] AH00489: Apache/2.4.38 (Unix) configured -- resuming normal operations\n[Sun Aug 30 16:57:13.068313 2020] [core:notice] [pid 1:tid 139877614336872] AH00094: Command line: 'httpd -D FOREGROUND'\n"
-[AfterEach] Kubectl run rc
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1530
-Aug 30 16:57:16.076: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete rc e2e-test-httpd-rc --namespace=kubectl-2592'
-Aug 30 16:57:16.147: INFO: stderr: ""
-Aug 30 16:57:16.147: INFO: stdout: "replicationcontroller \"e2e-test-httpd-rc\" deleted\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:57:16.147: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-2592" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run rc should create an rc from an image  [Conformance]","total":280,"completed":87,"skipped":1436,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-node] Downward API 
-  should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] Downward API
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:57:16.163: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-6562
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward api env vars
-Aug 30 16:57:16.331: INFO: Waiting up to 5m0s for pod "downward-api-cf7afb18-fe74-41b0-ae66-fff628d79318" in namespace "downward-api-6562" to be "success or failure"
-Aug 30 16:57:16.337: INFO: Pod "downward-api-cf7afb18-fe74-41b0-ae66-fff628d79318": Phase="Pending", Reason="", readiness=false. Elapsed: 5.778091ms
-Aug 30 16:57:18.342: INFO: Pod "downward-api-cf7afb18-fe74-41b0-ae66-fff628d79318": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.01134933s
-STEP: Saw pod success
-Aug 30 16:57:18.342: INFO: Pod "downward-api-cf7afb18-fe74-41b0-ae66-fff628d79318" satisfied condition "success or failure"
-Aug 30 16:57:18.347: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod downward-api-cf7afb18-fe74-41b0-ae66-fff628d79318 container dapi-container: 
-STEP: delete the pod
-Aug 30 16:57:18.419: INFO: Waiting for pod downward-api-cf7afb18-fe74-41b0-ae66-fff628d79318 to disappear
-Aug 30 16:57:18.422: INFO: Pod downward-api-cf7afb18-fe74-41b0-ae66-fff628d79318 no longer exists
-[AfterEach] [sig-node] Downward API
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:57:18.422: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-6562" for this suite.
-•{"msg":"PASSED [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]","total":280,"completed":88,"skipped":1463,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:57:18.434: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-7144
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-998083cf-6c9e-4b31-9220-2cdb3e446ff9
-STEP: Creating a pod to test consume secrets
-Aug 30 16:57:18.608: INFO: Waiting up to 5m0s for pod "pod-secrets-768e7354-39df-4083-b1ee-dde859a90c9b" in namespace "secrets-7144" to be "success or failure"
-Aug 30 16:57:18.616: INFO: Pod "pod-secrets-768e7354-39df-4083-b1ee-dde859a90c9b": Phase="Pending", Reason="", readiness=false. Elapsed: 7.644868ms
-Aug 30 16:57:20.621: INFO: Pod "pod-secrets-768e7354-39df-4083-b1ee-dde859a90c9b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012926395s
-STEP: Saw pod success
-Aug 30 16:57:20.621: INFO: Pod "pod-secrets-768e7354-39df-4083-b1ee-dde859a90c9b" satisfied condition "success or failure"
-Aug 30 16:57:20.626: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-secrets-768e7354-39df-4083-b1ee-dde859a90c9b container secret-volume-test: 
-STEP: delete the pod
-Aug 30 16:57:20.729: INFO: Waiting for pod pod-secrets-768e7354-39df-4083-b1ee-dde859a90c9b to disappear
-Aug 30 16:57:20.733: INFO: Pod pod-secrets-768e7354-39df-4083-b1ee-dde859a90c9b no longer exists
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:57:20.733: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-7144" for this suite.
-•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":89,"skipped":1496,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
-  should execute poststart exec hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:57:20.749: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-lifecycle-hook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-6365
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] when create a pod with lifecycle hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
-STEP: create the container to handle the HTTPGet hook request.
-[It] should execute poststart exec hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the pod with lifecycle hook
-STEP: check poststart hook
-STEP: delete the pod with lifecycle hook
-Aug 30 16:57:28.990: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Aug 30 16:57:28.995: INFO: Pod pod-with-poststart-exec-hook still exists
-Aug 30 16:57:30.995: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Aug 30 16:57:31.003: INFO: Pod pod-with-poststart-exec-hook still exists
-Aug 30 16:57:32.995: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
-Aug 30 16:57:33.000: INFO: Pod pod-with-poststart-exec-hook no longer exists
-[AfterEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:57:33.000: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-lifecycle-hook-6365" for this suite.
+error:
+exit status 1
+Feb 12 10:09:44.836: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:09:44.919: INFO: rc: 1
+Feb 12 10:09:44.919: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
 
-• [SLOW TEST:12.267 seconds]
-[k8s.io] Container Lifecycle Hook
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  when create a pod with lifecycle hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
-    should execute poststart exec hook properly [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance]","total":280,"completed":90,"skipped":1575,"failed":0}
-SSSSSSSSSSSSSSS
-------------------------------
-[sig-node] ConfigMap 
-  should fail to create ConfigMap with empty key [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:57:33.016: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-2060
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should fail to create ConfigMap with empty key [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap that has name configmap-test-emptyKey-c39561a7-f559-4b15-bd96-5be1619c61cb
-[AfterEach] [sig-node] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:57:33.187: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-2060" for this suite.
-•{"msg":"PASSED [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance]","total":280,"completed":91,"skipped":1590,"failed":0}
-SSSSSSSSS
-------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:57:33.202: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-803
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 16:57:33.380: INFO: Waiting up to 5m0s for pod "downwardapi-volume-b0cffc00-944b-4fd1-adb7-1d545e235695" in namespace "projected-803" to be "success or failure"
-Aug 30 16:57:33.386: INFO: Pod "downwardapi-volume-b0cffc00-944b-4fd1-adb7-1d545e235695": Phase="Pending", Reason="", readiness=false. Elapsed: 5.086596ms
-Aug 30 16:57:35.392: INFO: Pod "downwardapi-volume-b0cffc00-944b-4fd1-adb7-1d545e235695": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011795847s
-STEP: Saw pod success
-Aug 30 16:57:35.392: INFO: Pod "downwardapi-volume-b0cffc00-944b-4fd1-adb7-1d545e235695" satisfied condition "success or failure"
-Aug 30 16:57:35.397: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-b0cffc00-944b-4fd1-adb7-1d545e235695 container client-container: 
-STEP: delete the pod
-Aug 30 16:57:35.466: INFO: Waiting for pod downwardapi-volume-b0cffc00-944b-4fd1-adb7-1d545e235695 to disappear
-Aug 30 16:57:35.471: INFO: Pod downwardapi-volume-b0cffc00-944b-4fd1-adb7-1d545e235695 no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:57:35.471: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-803" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":92,"skipped":1599,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-network] Services 
-  should provide secure master service  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:57:35.490: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename services
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-7234
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should provide secure master service  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:57:35.654: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-7234" for this suite.
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
-•{"msg":"PASSED [sig-network] Services should provide secure master service  [Conformance]","total":280,"completed":93,"skipped":1625,"failed":0}
-SSSSSSSSSSSSS
-------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  Should recreate evicted statefulset [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:57:35.671: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename statefulset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-6493
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
-STEP: Creating service test in namespace statefulset-6493
-[It] Should recreate evicted statefulset [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Looking for a node to schedule stateful set and pod
-STEP: Creating pod with conflicting port in namespace statefulset-6493
-STEP: Creating statefulset with conflicting port in namespace statefulset-6493
-STEP: Waiting until pod test-pod will start running in namespace statefulset-6493
-STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-6493
-Aug 30 16:57:39.926: INFO: Observed stateful pod in namespace: statefulset-6493, name: ss-0, uid: 461803bf-b314-4878-a95c-22533646ca30, status phase: Pending. Waiting for statefulset controller to delete.
-Aug 30 16:57:40.115: INFO: Observed stateful pod in namespace: statefulset-6493, name: ss-0, uid: 461803bf-b314-4878-a95c-22533646ca30, status phase: Failed. Waiting for statefulset controller to delete.
-Aug 30 16:57:40.125: INFO: Observed stateful pod in namespace: statefulset-6493, name: ss-0, uid: 461803bf-b314-4878-a95c-22533646ca30, status phase: Failed. Waiting for statefulset controller to delete.
-Aug 30 16:57:40.132: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-6493
-STEP: Removing pod with conflicting port in namespace statefulset-6493
-STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-6493 and will be in running state
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:09:54.919: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:09:55.006: INFO: rc: 1
+Feb 12 10:09:55.006: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:10:05.007: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:10:05.104: INFO: rc: 1
+Feb 12 10:10:05.104: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:10:15.104: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:10:15.182: INFO: rc: 1
+Feb 12 10:10:15.182: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:10:25.182: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:10:25.264: INFO: rc: 1
+Feb 12 10:10:25.264: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:10:35.264: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:10:35.342: INFO: rc: 1
+Feb 12 10:10:35.342: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:10:45.342: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:10:45.424: INFO: rc: 1
+Feb 12 10:10:45.424: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:10:55.424: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:10:55.504: INFO: rc: 1
+Feb 12 10:10:55.504: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:11:05.504: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:11:05.590: INFO: rc: 1
+Feb 12 10:11:05.590: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:11:15.590: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:11:45.648: INFO: rc: 1
+Feb 12 10:11:45.648: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Unable to connect to the server: dial tcp 10.240.16.1:443: i/o timeout
+
+error:
+exit status 1
+Feb 12 10:11:55.648: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:11:55.732: INFO: rc: 1
+Feb 12 10:11:55.732: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:12:05.732: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:12:05.823: INFO: rc: 1
+Feb 12 10:12:05.823: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:12:15.823: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:12:15.920: INFO: rc: 1
+Feb 12 10:12:15.920: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:12:25.920: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:12:26.002: INFO: rc: 1
+Feb 12 10:12:26.002: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:12:36.002: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:12:36.100: INFO: rc: 1
+Feb 12 10:12:36.100: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:12:46.100: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:12:46.198: INFO: rc: 1
+Feb 12 10:12:46.198: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:12:56.198: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:12:56.296: INFO: rc: 1
+Feb 12 10:12:56.296: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:13:06.296: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:13:06.409: INFO: rc: 1
+Feb 12 10:13:06.409: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:13:16.409: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:13:16.497: INFO: rc: 1
+Feb 12 10:13:16.497: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:13:26.497: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:13:26.576: INFO: rc: 1
+Feb 12 10:13:26.576: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:13:36.576: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:13:36.655: INFO: rc: 1
+Feb 12 10:13:36.655: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:13:46.656: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:13:46.746: INFO: rc: 1
+Feb 12 10:13:46.746: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:13:56.746: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:13:56.841: INFO: rc: 1
+Feb 12 10:13:56.841: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:14:06.842: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:14:06.921: INFO: rc: 1
+Feb 12 10:14:06.921: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:14:16.921: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:14:17.000: INFO: rc: 1
+Feb 12 10:14:17.000: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
+Command stdout:
+
+stderr:
+Error from server (NotFound): pods "ss-2" not found
+
+error:
+exit status 1
+Feb 12 10:14:27.000: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6914 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:14:27.081: INFO: rc: 1
+Feb 12 10:14:27.081: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: 
+Feb 12 10:14:27.081: INFO: Scaling statefulset ss to 0
+STEP: Verifying that stateful set ss was scaled down in reverse order
 [AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
-Aug 30 16:57:46.178: INFO: Deleting all statefulset in ns statefulset-6493
-Aug 30 16:57:46.184: INFO: Scaling statefulset ss to 0
-Aug 30 16:57:56.207: INFO: Waiting for statefulset status.replicas updated to 0
-Aug 30 16:57:56.212: INFO: Deleting statefulset ss
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
+Feb 12 10:14:27.098: INFO: Deleting all statefulset in ns statefulset-6914
+Feb 12 10:14:27.104: INFO: Scaling statefulset ss to 0
+Feb 12 10:14:27.122: INFO: Waiting for statefulset status.replicas updated to 0
+Feb 12 10:14:27.127: INFO: Deleting statefulset ss
 [AfterEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:57:56.231: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-6493" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:14:27.153: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-6914" for this suite.
 
-• [SLOW TEST:20.577 seconds]
+• [SLOW TEST:367.678 seconds]
 [sig-apps] StatefulSet
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
   [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-    Should recreate evicted statefulset [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+    Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance]","total":280,"completed":94,"skipped":1638,"failed":0}
-SSSSSS
+{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]","total":280,"completed":72,"skipped":953,"failed":0}
+SSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Projected configMap 
+  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:57:56.248: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-2805
+Feb 12 10:14:27.170: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6102
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name configmap-test-volume-map-be64058a-324e-4ec6-8ffd-08034b855bd4
+[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name projected-configmap-test-volume-cd5a98bf-00d8-48de-88e2-c295993ec696
 STEP: Creating a pod to test consume configMaps
-Aug 30 16:57:56.429: INFO: Waiting up to 5m0s for pod "pod-configmaps-6ba3572b-870a-4619-82d1-a08a8c8a845e" in namespace "configmap-2805" to be "success or failure"
-Aug 30 16:57:56.434: INFO: Pod "pod-configmaps-6ba3572b-870a-4619-82d1-a08a8c8a845e": Phase="Pending", Reason="", readiness=false. Elapsed: 4.434899ms
-Aug 30 16:57:58.440: INFO: Pod "pod-configmaps-6ba3572b-870a-4619-82d1-a08a8c8a845e": Phase="Running", Reason="", readiness=true. Elapsed: 2.010422495s
-Aug 30 16:58:00.447: INFO: Pod "pod-configmaps-6ba3572b-870a-4619-82d1-a08a8c8a845e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017794371s
+Feb 12 10:14:27.349: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-f9091ed3-3ba7-4706-af19-f8f896236209" in namespace "projected-6102" to be "success or failure"
+Feb 12 10:14:27.354: INFO: Pod "pod-projected-configmaps-f9091ed3-3ba7-4706-af19-f8f896236209": Phase="Pending", Reason="", readiness=false. Elapsed: 5.284272ms
+Feb 12 10:14:29.366: INFO: Pod "pod-projected-configmaps-f9091ed3-3ba7-4706-af19-f8f896236209": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01670394s
+Feb 12 10:14:31.373: INFO: Pod "pod-projected-configmaps-f9091ed3-3ba7-4706-af19-f8f896236209": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023799355s
 STEP: Saw pod success
-Aug 30 16:58:00.447: INFO: Pod "pod-configmaps-6ba3572b-870a-4619-82d1-a08a8c8a845e" satisfied condition "success or failure"
-Aug 30 16:58:00.452: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-configmaps-6ba3572b-870a-4619-82d1-a08a8c8a845e container configmap-volume-test: 
+Feb 12 10:14:31.373: INFO: Pod "pod-projected-configmaps-f9091ed3-3ba7-4706-af19-f8f896236209" satisfied condition "success or failure"
+Feb 12 10:14:31.382: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-projected-configmaps-f9091ed3-3ba7-4706-af19-f8f896236209 container projected-configmap-volume-test: 
 STEP: delete the pod
-Aug 30 16:58:00.521: INFO: Waiting for pod pod-configmaps-6ba3572b-870a-4619-82d1-a08a8c8a845e to disappear
-Aug 30 16:58:00.526: INFO: Pod pod-configmaps-6ba3572b-870a-4619-82d1-a08a8c8a845e no longer exists
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:58:00.526: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-2805" for this suite.
-•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":95,"skipped":1644,"failed":0}
-SSSSSSSSSS
+Feb 12 10:14:31.419: INFO: Waiting for pod pod-projected-configmaps-f9091ed3-3ba7-4706-af19-f8f896236209 to disappear
+Feb 12 10:14:31.424: INFO: Pod pod-projected-configmaps-f9091ed3-3ba7-4706-af19-f8f896236209 no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:14:31.424: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-6102" for this suite.
+•{"msg":"PASSED [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":280,"completed":73,"skipped":957,"failed":0}
+SSS
 ------------------------------
-[sig-cli] Kubectl client Update Demo 
-  should do a rolling update of a replication controller  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Projected configMap 
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:58:00.547: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2382
+Feb 12 10:14:31.441: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-2233
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[BeforeEach] Update Demo
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:324
-[It] should do a rolling update of a replication controller  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the initial replication controller
-Aug 30 16:58:00.709: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-2382'
-Aug 30 16:58:00.963: INFO: stderr: ""
-Aug 30 16:58:00.963: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Aug 30 16:58:00.963: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-2382'
-Aug 30 16:58:01.038: INFO: stderr: ""
-Aug 30 16:58:01.038: INFO: stdout: "update-demo-nautilus-sdxbj update-demo-nautilus-vshvf "
-Aug 30 16:58:01.038: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-sdxbj -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2382'
-Aug 30 16:58:01.099: INFO: stderr: ""
-Aug 30 16:58:01.099: INFO: stdout: ""
-Aug 30 16:58:01.099: INFO: update-demo-nautilus-sdxbj is created but not running
-Aug 30 16:58:06.099: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-2382'
-Aug 30 16:58:06.166: INFO: stderr: ""
-Aug 30 16:58:06.166: INFO: stdout: "update-demo-nautilus-sdxbj update-demo-nautilus-vshvf "
-Aug 30 16:58:06.166: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-sdxbj -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2382'
-Aug 30 16:58:06.231: INFO: stderr: ""
-Aug 30 16:58:06.231: INFO: stdout: "true"
-Aug 30 16:58:06.231: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-sdxbj -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-2382'
-Aug 30 16:58:06.293: INFO: stderr: ""
-Aug 30 16:58:06.293: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Aug 30 16:58:06.293: INFO: validating pod update-demo-nautilus-sdxbj
-Aug 30 16:58:06.429: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Aug 30 16:58:06.429: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Aug 30 16:58:06.429: INFO: update-demo-nautilus-sdxbj is verified up and running
-Aug 30 16:58:06.429: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-vshvf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2382'
-Aug 30 16:58:06.498: INFO: stderr: ""
-Aug 30 16:58:06.498: INFO: stdout: "true"
-Aug 30 16:58:06.498: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-vshvf -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-2382'
-Aug 30 16:58:06.563: INFO: stderr: ""
-Aug 30 16:58:06.563: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Aug 30 16:58:06.563: INFO: validating pod update-demo-nautilus-vshvf
-Aug 30 16:58:06.669: INFO: got data: {
-  "image": "nautilus.jpg"
-}
+[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name cm-test-opt-del-b7304549-b01b-45c3-9e9a-0e588370352e
+STEP: Creating configMap with name cm-test-opt-upd-026246dc-0a22-4620-ae7c-4484e9c9dd35
+STEP: Creating the pod
+STEP: Deleting configmap cm-test-opt-del-b7304549-b01b-45c3-9e9a-0e588370352e
+STEP: Updating configmap cm-test-opt-upd-026246dc-0a22-4620-ae7c-4484e9c9dd35
+STEP: Creating configMap with name cm-test-opt-create-2047c22e-e327-4ab8-af43-69bd4d55213f
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:15:59.287: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-2233" for this suite.
 
-Aug 30 16:58:06.669: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Aug 30 16:58:06.669: INFO: update-demo-nautilus-vshvf is verified up and running
-STEP: rolling-update to new replication controller
-Aug 30 16:58:06.672: INFO: scanned /root for discovery docs: 
-Aug 30 16:58:06.672: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 rolling-update update-demo-nautilus --update-period=1s -f - --namespace=kubectl-2382'
-Aug 30 16:58:29.259: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n"
-Aug 30 16:58:29.259: INFO: stdout: "Created update-demo-kitten\nScaling up update-demo-kitten from 0 to 2, scaling down update-demo-nautilus from 2 to 0 (keep 2 pods available, don't exceed 3 pods)\nScaling update-demo-kitten up to 1\nScaling update-demo-nautilus down to 1\nScaling update-demo-kitten up to 2\nScaling update-demo-nautilus down to 0\nUpdate succeeded. Deleting old controller: update-demo-nautilus\nRenaming update-demo-kitten to update-demo-nautilus\nreplicationcontroller/update-demo-nautilus rolling updated\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Aug 30 16:58:29.259: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-2382'
-Aug 30 16:58:29.334: INFO: stderr: ""
-Aug 30 16:58:29.334: INFO: stdout: "update-demo-kitten-rq8n6 update-demo-kitten-xkh5s "
-Aug 30 16:58:29.334: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-kitten-rq8n6 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2382'
-Aug 30 16:58:29.397: INFO: stderr: ""
-Aug 30 16:58:29.397: INFO: stdout: "true"
-Aug 30 16:58:29.397: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-kitten-rq8n6 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-2382'
-Aug 30 16:58:29.467: INFO: stderr: ""
-Aug 30 16:58:29.467: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
-Aug 30 16:58:29.467: INFO: validating pod update-demo-kitten-rq8n6
-Aug 30 16:58:29.601: INFO: got data: {
-  "image": "kitten.jpg"
-}
-
-Aug 30 16:58:29.601: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
-Aug 30 16:58:29.601: INFO: update-demo-kitten-rq8n6 is verified up and running
-Aug 30 16:58:29.601: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-kitten-xkh5s -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2382'
-Aug 30 16:58:29.671: INFO: stderr: ""
-Aug 30 16:58:29.671: INFO: stdout: "true"
-Aug 30 16:58:29.671: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-kitten-xkh5s -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-2382'
-Aug 30 16:58:29.736: INFO: stderr: ""
-Aug 30 16:58:29.736: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
-Aug 30 16:58:29.736: INFO: validating pod update-demo-kitten-xkh5s
-Aug 30 16:58:29.870: INFO: got data: {
-  "image": "kitten.jpg"
-}
-
-Aug 30 16:58:29.870: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
-Aug 30 16:58:29.870: INFO: update-demo-kitten-xkh5s is verified up and running
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:58:29.870: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-2382" for this suite.
-
-• [SLOW TEST:29.338 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Update Demo
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:322
-    should do a rolling update of a replication controller  [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:87.864 seconds]
+[sig-storage] Projected configMap
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Update Demo should do a rolling update of a replication controller  [Conformance]","total":280,"completed":96,"skipped":1654,"failed":0}
-S
+{"msg":"PASSED [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":74,"skipped":960,"failed":0}
+SSSSSSSSSS
 ------------------------------
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  patching/updating a validating webhook should work [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should be able to deny attaching pod [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:58:29.885: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:15:59.305: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-4481
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-6802
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
 STEP: Setting up server cert
 STEP: Create role binding to let webhook read extension-apiserver-authentication
 STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Aug 30 16:58:30.437: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Aug 30 16:58:32.460: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403510, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403510, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403510, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403510, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:15:59.701: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:16:01.718: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721759, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721759, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721759, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721759, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Aug 30 16:58:35.480: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] patching/updating a validating webhook should work [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a validating webhook configuration
-STEP: Creating a configMap that does not comply to the validation webhook rules
-STEP: Updating a validating webhook configuration's rules to not include the create operation
-STEP: Creating a configMap that does not comply to the validation webhook rules
-STEP: Patching a validating webhook configuration's rules to include the create operation
-STEP: Creating a configMap that does not comply to the validation webhook rules
+Feb 12 10:16:04.741: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should be able to deny attaching pod [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Registering the webhook via the AdmissionRegistration API
+STEP: create a pod
+STEP: 'kubectl attach' the pod, should be denied by the webhook
+Feb 12 10:16:08.921: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 attach --namespace=webhook-6802 to-be-attached-pod -i -c=container1'
+Feb 12 10:16:09.202: INFO: rc: 1
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:58:35.912: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-4481" for this suite.
-STEP: Destroying namespace "webhook-4481-markers" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:16:09.214: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-6802" for this suite.
+STEP: Destroying namespace "webhook-6802-markers" for this suite.
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:6.109 seconds]
+• [SLOW TEST:10.010 seconds]
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  patching/updating a validating webhook should work [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance]","total":280,"completed":97,"skipped":1655,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSS
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should be able to deny attaching pod [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide container's memory request [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:58:35.994: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-4964
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should provide container's memory request [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 16:58:36.163: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7f90bb7d-2770-42cf-8f72-773c196a8552" in namespace "downward-api-4964" to be "success or failure"
-Aug 30 16:58:36.174: INFO: Pod "downwardapi-volume-7f90bb7d-2770-42cf-8f72-773c196a8552": Phase="Pending", Reason="", readiness=false. Elapsed: 10.474979ms
-Aug 30 16:58:38.179: INFO: Pod "downwardapi-volume-7f90bb7d-2770-42cf-8f72-773c196a8552": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01574016s
-Aug 30 16:58:40.185: INFO: Pod "downwardapi-volume-7f90bb7d-2770-42cf-8f72-773c196a8552": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021856357s
-STEP: Saw pod success
-Aug 30 16:58:40.185: INFO: Pod "downwardapi-volume-7f90bb7d-2770-42cf-8f72-773c196a8552" satisfied condition "success or failure"
-Aug 30 16:58:40.190: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod downwardapi-volume-7f90bb7d-2770-42cf-8f72-773c196a8552 container client-container: 
-STEP: delete the pod
-Aug 30 16:58:40.260: INFO: Waiting for pod downwardapi-volume-7f90bb7d-2770-42cf-8f72-773c196a8552 to disappear
-Aug 30 16:58:40.264: INFO: Pod downwardapi-volume-7f90bb7d-2770-42cf-8f72-773c196a8552 no longer exists
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:58:40.264: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-4964" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance]","total":280,"completed":98,"skipped":1682,"failed":0}
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance]","total":280,"completed":75,"skipped":970,"failed":0}
 SSSSSSSS
 ------------------------------
-[sig-api-machinery] ResourceQuota 
-  should create a ResourceQuota and capture the life of a pod. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] Networking Granular Checks: Pods 
+  should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Networking
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:58:40.279: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename resourcequota
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-2994
+Feb 12 10:16:09.316: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename pod-network-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-392
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should create a ResourceQuota and capture the life of a pod. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Counting existing ResourceQuota
-STEP: Creating a ResourceQuota
-STEP: Ensuring resource quota status is calculated
-STEP: Creating a Pod that fits quota
-STEP: Ensuring ResourceQuota status captures the pod usage
-STEP: Not allowing a pod to be created that exceeds remaining quota
-STEP: Not allowing a pod to be created that exceeds remaining quota(validation on extended resources)
-STEP: Ensuring a pod cannot update its resource requirements
-STEP: Ensuring attempts to update pod resource requirements did not change quota usage
-STEP: Deleting the pod
-STEP: Ensuring resource quota status released the pod usage
-[AfterEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:58:53.540: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-2994" for this suite.
+[It] should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Performing setup for networking test in namespace pod-network-test-392
+STEP: creating a selector
+STEP: Creating the service pods in kubernetes
+Feb 12 10:16:09.470: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+STEP: Creating test pods
+Feb 12 10:16:33.604: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.25.1.67:8080/dial?request=hostname&protocol=http&host=172.25.2.30&port=8080&tries=1'] Namespace:pod-network-test-392 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:16:33.604: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:16:34.070: INFO: Waiting for responses: map[]
+Feb 12 10:16:34.076: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.25.1.67:8080/dial?request=hostname&protocol=http&host=172.25.0.23&port=8080&tries=1'] Namespace:pod-network-test-392 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:16:34.076: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:16:34.627: INFO: Waiting for responses: map[]
+Feb 12 10:16:34.634: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.25.1.67:8080/dial?request=hostname&protocol=http&host=172.25.1.66&port=8080&tries=1'] Namespace:pod-network-test-392 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:16:34.634: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:16:35.141: INFO: Waiting for responses: map[]
+[AfterEach] [sig-network] Networking
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:16:35.141: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pod-network-test-392" for this suite.
 
-• [SLOW TEST:13.276 seconds]
-[sig-api-machinery] ResourceQuota
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should create a ResourceQuota and capture the life of a pod. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:25.844 seconds]
+[sig-network] Networking
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
+  Granular Checks: Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
+    should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance]","total":280,"completed":99,"skipped":1690,"failed":0}
-SSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":76,"skipped":978,"failed":0}
+SSSSSS
 ------------------------------
-[sig-storage] Projected configMap 
-  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-node] Downward API 
+  should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-node] Downward API
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:58:53.555: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1140
+Feb 12 10:16:35.160: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-3510
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name projected-configmap-test-volume-map-a4a0c10c-8d35-4a6b-b19e-ccf905550ae2
-STEP: Creating a pod to test consume configMaps
-Aug 30 16:58:53.757: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-18c0f743-8008-46e2-94f8-f65116ba1f47" in namespace "projected-1140" to be "success or failure"
-Aug 30 16:58:53.764: INFO: Pod "pod-projected-configmaps-18c0f743-8008-46e2-94f8-f65116ba1f47": Phase="Pending", Reason="", readiness=false. Elapsed: 7.525193ms
-Aug 30 16:58:55.771: INFO: Pod "pod-projected-configmaps-18c0f743-8008-46e2-94f8-f65116ba1f47": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.013654398s
+[It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward api env vars
+Feb 12 10:16:35.340: INFO: Waiting up to 5m0s for pod "downward-api-15da7c22-d2ec-4890-aa6c-7f32abf90e86" in namespace "downward-api-3510" to be "success or failure"
+Feb 12 10:16:35.344: INFO: Pod "downward-api-15da7c22-d2ec-4890-aa6c-7f32abf90e86": Phase="Pending", Reason="", readiness=false. Elapsed: 3.976823ms
+Feb 12 10:16:37.350: INFO: Pod "downward-api-15da7c22-d2ec-4890-aa6c-7f32abf90e86": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.009676337s
 STEP: Saw pod success
-Aug 30 16:58:55.771: INFO: Pod "pod-projected-configmaps-18c0f743-8008-46e2-94f8-f65116ba1f47" satisfied condition "success or failure"
-Aug 30 16:58:55.780: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-projected-configmaps-18c0f743-8008-46e2-94f8-f65116ba1f47 container projected-configmap-volume-test: 
+Feb 12 10:16:37.350: INFO: Pod "downward-api-15da7c22-d2ec-4890-aa6c-7f32abf90e86" satisfied condition "success or failure"
+Feb 12 10:16:37.357: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-dbhjj pod downward-api-15da7c22-d2ec-4890-aa6c-7f32abf90e86 container dapi-container: 
 STEP: delete the pod
-Aug 30 16:58:55.855: INFO: Waiting for pod pod-projected-configmaps-18c0f743-8008-46e2-94f8-f65116ba1f47 to disappear
-Aug 30 16:58:55.861: INFO: Pod pod-projected-configmaps-18c0f743-8008-46e2-94f8-f65116ba1f47 no longer exists
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:58:55.861: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-1140" for this suite.
-•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":100,"skipped":1706,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-cli] Kubectl client Kubectl run pod 
-  should create a pod from an image when restart is Never  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:58:55.879: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7394
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[BeforeEach] Kubectl run pod
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1754
-[It] should create a pod from an image when restart is Never  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: running the image docker.io/library/httpd:2.4.38-alpine
-Aug 30 16:58:56.042: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 run e2e-test-httpd-pod --restart=Never --generator=run-pod/v1 --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-7394'
-Aug 30 16:58:56.132: INFO: stderr: ""
-Aug 30 16:58:56.132: INFO: stdout: "pod/e2e-test-httpd-pod created\n"
-STEP: verifying the pod e2e-test-httpd-pod was created
-[AfterEach] Kubectl run pod
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1759
-Aug 30 16:58:56.138: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete pods e2e-test-httpd-pod --namespace=kubectl-7394'
-Aug 30 16:59:04.282: INFO: stderr: ""
-Aug 30 16:59:04.282: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:59:04.282: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-7394" for this suite.
-
-• [SLOW TEST:8.417 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Kubectl run pod
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1750
-    should create a pod from an image when restart is Never  [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never  [Conformance]","total":280,"completed":101,"skipped":1730,"failed":0}
+Feb 12 10:16:37.395: INFO: Waiting for pod downward-api-15da7c22-d2ec-4890-aa6c-7f32abf90e86 to disappear
+Feb 12 10:16:37.402: INFO: Pod downward-api-15da7c22-d2ec-4890-aa6c-7f32abf90e86 no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:16:37.402: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-3510" for this suite.
+•{"msg":"PASSED [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]","total":280,"completed":77,"skipped":984,"failed":0}
 SSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected configMap 
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 16:59:04.296: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-4910
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name cm-test-opt-del-6476be7b-c3e8-4204-b0fe-d8148695f67e
-STEP: Creating configMap with name cm-test-opt-upd-57b8491b-8647-4cf8-9712-00cff247cb95
-STEP: Creating the pod
-STEP: Deleting configmap cm-test-opt-del-6476be7b-c3e8-4204-b0fe-d8148695f67e
-STEP: Updating configmap cm-test-opt-upd-57b8491b-8647-4cf8-9712-00cff247cb95
-STEP: Creating configMap with name cm-test-opt-create-67c4d135-d245-40ab-9741-6ba51f9e2989
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:59:08.998: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-4910" for this suite.
-•{"msg":"PASSED [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":102,"skipped":1748,"failed":0}
-SSSSSS
-------------------------------
-[sig-storage] Projected secret 
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] InitContainer [NodeConformance] 
+  should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:59:09.018: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8456
+Feb 12 10:16:37.420: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename init-container
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-4444
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name s-test-opt-del-86c10b45-ec9e-49be-aa9f-e6ab2b4ecad8
-STEP: Creating secret with name s-test-opt-upd-dc86dfc2-25f2-48a1-a1d6-70d14f5a5dec
-STEP: Creating the pod
-STEP: Deleting secret s-test-opt-del-86c10b45-ec9e-49be-aa9f-e6ab2b4ecad8
-STEP: Updating secret s-test-opt-upd-dc86dfc2-25f2-48a1-a1d6-70d14f5a5dec
-STEP: Creating secret with name s-test-opt-create-12fa9a05-9d29-44fb-b216-109fbac5641b
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:59:13.637: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-8456" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":103,"skipped":1754,"failed":0}
-SS
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
+[It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating the pod
+Feb 12 10:16:37.585: INFO: PodSpec: initContainers in spec.initContainers
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:16:40.504: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "init-container-4444" for this suite.
+•{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]","total":280,"completed":78,"skipped":1002,"failed":0}
+SSSSSSSSSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide container's cpu limit [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-node] ConfigMap 
+  should be consumable via the environment [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-node] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:59:13.654: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7210
+Feb 12 10:16:40.528: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-434
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should provide container's cpu limit [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 16:59:13.827: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2302d46c-df17-4be0-8acd-63a386e590a7" in namespace "projected-7210" to be "success or failure"
-Aug 30 16:59:13.834: INFO: Pod "downwardapi-volume-2302d46c-df17-4be0-8acd-63a386e590a7": Phase="Pending", Reason="", readiness=false. Elapsed: 6.862177ms
-Aug 30 16:59:15.839: INFO: Pod "downwardapi-volume-2302d46c-df17-4be0-8acd-63a386e590a7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012580265s
+[It] should be consumable via the environment [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap configmap-434/configmap-test-78a0dff9-1910-458e-b0b6-25b64e058933
+STEP: Creating a pod to test consume configMaps
+Feb 12 10:16:40.699: INFO: Waiting up to 5m0s for pod "pod-configmaps-e3a48399-a6db-40ad-8107-e1f160650c57" in namespace "configmap-434" to be "success or failure"
+Feb 12 10:16:40.707: INFO: Pod "pod-configmaps-e3a48399-a6db-40ad-8107-e1f160650c57": Phase="Pending", Reason="", readiness=false. Elapsed: 7.58767ms
+Feb 12 10:16:42.712: INFO: Pod "pod-configmaps-e3a48399-a6db-40ad-8107-e1f160650c57": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012418421s
+Feb 12 10:16:44.718: INFO: Pod "pod-configmaps-e3a48399-a6db-40ad-8107-e1f160650c57": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018354479s
 STEP: Saw pod success
-Aug 30 16:59:15.839: INFO: Pod "downwardapi-volume-2302d46c-df17-4be0-8acd-63a386e590a7" satisfied condition "success or failure"
-Aug 30 16:59:15.844: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-2302d46c-df17-4be0-8acd-63a386e590a7 container client-container: 
+Feb 12 10:16:44.718: INFO: Pod "pod-configmaps-e3a48399-a6db-40ad-8107-e1f160650c57" satisfied condition "success or failure"
+Feb 12 10:16:44.723: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-configmaps-e3a48399-a6db-40ad-8107-e1f160650c57 container env-test: 
 STEP: delete the pod
-Aug 30 16:59:15.913: INFO: Waiting for pod downwardapi-volume-2302d46c-df17-4be0-8acd-63a386e590a7 to disappear
-Aug 30 16:59:15.918: INFO: Pod downwardapi-volume-2302d46c-df17-4be0-8acd-63a386e590a7 no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 16:59:15.918: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-7210" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance]","total":280,"completed":104,"skipped":1756,"failed":0}
-SSSSSSSSSS
+Feb 12 10:16:44.764: INFO: Waiting for pod pod-configmaps-e3a48399-a6db-40ad-8107-e1f160650c57 to disappear
+Feb 12 10:16:44.768: INFO: Pod pod-configmaps-e3a48399-a6db-40ad-8107-e1f160650c57 no longer exists
+[AfterEach] [sig-node] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:16:44.769: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-434" for this suite.
+•{"msg":"PASSED [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance]","total":280,"completed":79,"skipped":1012,"failed":0}
+SSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] InitContainer [NodeConformance] 
-  should not start app containers if init containers fail on a RestartAlways pod [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  works for CRD preserving unknown fields at the schema root [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 16:59:15.930: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename init-container
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-8547
+Feb 12 10:16:44.785: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename crd-publish-openapi
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-7141
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
-[It] should not start app containers if init containers fail on a RestartAlways pod [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-Aug 30 16:59:16.086: INFO: PodSpec: initContainers in spec.initContainers
-Aug 30 17:00:02.521: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-14f33a98-b13d-4740-9b21-312839e6a27b", GenerateName:"", Namespace:"init-container-8547", SelfLink:"/api/v1/namespaces/init-container-8547/pods/pod-init-14f33a98-b13d-4740-9b21-312839e6a27b", UID:"297636c5-dd7f-4681-8839-3c901e827a97", ResourceVersion:"13160", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63734403556, loc:(*time.Location)(0x7925260)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"86902557"}, Annotations:map[string]string{"cni.projectcalico.org/podIP":"172.25.0.96/32"}, OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-94bws", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc003978b00), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-94bws", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-94bws", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-94bws", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc00412ddf8), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"adoring-wozniak-54dcfd79fc-948mf", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc00326eba0), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc00412de70)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc00412de90)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc00412de98), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc00412de9c), PreemptionPolicy:(*v1.PreemptionPolicy)(nil), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403556, loc:(*time.Location)(0x7925260)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403556, loc:(*time.Location)(0x7925260)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403556, loc:(*time.Location)(0x7925260)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403556, loc:(*time.Location)(0x7925260)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"46.101.135.210", PodIP:"172.25.0.96", PodIPs:[]v1.PodIP{v1.PodIP{IP:"172.25.0.96"}}, StartTime:(*v1.Time)(0xc002f69ec0), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(0xc002f69f00), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc000207730)}, Ready:false, RestartCount:3, Image:"busybox:1.29", ImageID:"docker-pullable://busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"docker://a2988fd2e6a5e8c4c5c111c1b1f6e4ebab8d018d17a8f3c5aa103f97f43c4adc", Started:(*bool)(nil)}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc002f69f40), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:"", Started:(*bool)(nil)}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc002f69ee0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.1", ImageID:"", ContainerID:"", Started:(*bool)(0xc00412df1f)}}, QOSClass:"Burstable", EphemeralContainerStatuses:[]v1.ContainerStatus(nil)}}
-[AfterEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:00:02.522: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "init-container-8547" for this suite.
+[It] works for CRD preserving unknown fields at the schema root [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:16:44.938: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: client-side validation (kubectl create and apply) allows request with any unknown properties
+Feb 12 10:16:47.812: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-7141 create -f -'
+Feb 12 10:16:48.198: INFO: stderr: ""
+Feb 12 10:16:48.198: INFO: stdout: "e2e-test-crd-publish-openapi-8714-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n"
+Feb 12 10:16:48.198: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-7141 delete e2e-test-crd-publish-openapi-8714-crds test-cr'
+Feb 12 10:16:48.293: INFO: stderr: ""
+Feb 12 10:16:48.293: INFO: stdout: "e2e-test-crd-publish-openapi-8714-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n"
+Feb 12 10:16:48.294: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-7141 apply -f -'
+Feb 12 10:16:48.470: INFO: stderr: ""
+Feb 12 10:16:48.470: INFO: stdout: "e2e-test-crd-publish-openapi-8714-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n"
+Feb 12 10:16:48.470: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-7141 delete e2e-test-crd-publish-openapi-8714-crds test-cr'
+Feb 12 10:16:48.561: INFO: stderr: ""
+Feb 12 10:16:48.561: INFO: stdout: "e2e-test-crd-publish-openapi-8714-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n"
+STEP: kubectl explain works to explain CR
+Feb 12 10:16:48.561: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 explain e2e-test-crd-publish-openapi-8714-crds'
+Feb 12 10:16:48.762: INFO: stderr: ""
+Feb 12 10:16:48.762: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-8714-crd\nVERSION:  crd-publish-openapi-test-unknown-at-root.example.com/v1\n\nDESCRIPTION:\n     \n"
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:16:51.605: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-7141" for this suite.
 
-• [SLOW TEST:46.605 seconds]
-[k8s.io] InitContainer [NodeConformance]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should not start app containers if init containers fail on a RestartAlways pod [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:6.839 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  works for CRD preserving unknown fields at the schema root [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance]","total":280,"completed":105,"skipped":1766,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance]","total":280,"completed":80,"skipped":1028,"failed":0}
+SSSSSSSSSSS
 ------------------------------
-[k8s.io] [sig-node] Pods Extended [k8s.io] Delete Grace Period 
-  should be submitted and removed [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] [sig-node] Pods Extended
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-cli] Kubectl client Kubectl logs 
+  should be able to retrieve and filter logs  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:00:02.538: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-5776
+Feb 12 10:16:51.626: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-3894
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Delete Grace Period
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:46
-[It] should be submitted and removed [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-STEP: setting up selector
-STEP: submitting the pod to kubernetes
-STEP: verifying the pod is in kubernetes
-Aug 30 17:00:04.731: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-962893509 proxy -p 0'
-STEP: deleting the pod gracefully
-STEP: verifying the kubelet observed the termination notice
-Aug 30 17:00:09.819: INFO: no pod exists with the name we were looking for, assuming the termination request was observed and completed
-[AfterEach] [k8s.io] [sig-node] Pods Extended
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:00:09.824: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-5776" for this suite.
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[BeforeEach] Kubectl logs
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1358
+STEP: creating an pod
+Feb 12 10:16:51.787: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 run logs-generator --generator=run-pod/v1 --image=gcr.io/kubernetes-e2e-test-images/agnhost:2.8 --namespace=kubectl-3894 -- logs-generator --log-lines-total 100 --run-duration 20s'
+Feb 12 10:16:51.898: INFO: stderr: ""
+Feb 12 10:16:51.898: INFO: stdout: "pod/logs-generator created\n"
+[It] should be able to retrieve and filter logs  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Waiting for log generator to start.
+Feb 12 10:16:51.898: INFO: Waiting up to 5m0s for 1 pods to be running and ready, or succeeded: [logs-generator]
+Feb 12 10:16:51.898: INFO: Waiting up to 5m0s for pod "logs-generator" in namespace "kubectl-3894" to be "running and ready, or succeeded"
+Feb 12 10:16:51.904: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 6.345543ms
+Feb 12 10:16:53.910: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012611716s
+Feb 12 10:16:55.916: INFO: Pod "logs-generator": Phase="Running", Reason="", readiness=true. Elapsed: 4.018628413s
+Feb 12 10:16:55.916: INFO: Pod "logs-generator" satisfied condition "running and ready, or succeeded"
+Feb 12 10:16:55.916: INFO: Wanted all 1 pods to be running and ready, or succeeded. Result: true. Pods: [logs-generator]
+STEP: checking for a matching strings
+Feb 12 10:16:55.916: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 logs logs-generator logs-generator --namespace=kubectl-3894'
+Feb 12 10:16:56.101: INFO: stderr: ""
+Feb 12 10:16:56.101: INFO: stdout: "I0212 10:16:53.157041       1 logs_generator.go:76] 0 PUT /api/v1/namespaces/kube-system/pods/q4xx 576\nI0212 10:16:53.357241       1 logs_generator.go:76] 1 POST /api/v1/namespaces/default/pods/f96 285\nI0212 10:16:53.557319       1 logs_generator.go:76] 2 PUT /api/v1/namespaces/default/pods/jhq 258\nI0212 10:16:53.757181       1 logs_generator.go:76] 3 PUT /api/v1/namespaces/ns/pods/frh 378\nI0212 10:16:53.957185       1 logs_generator.go:76] 4 POST /api/v1/namespaces/ns/pods/k9z 271\nI0212 10:16:54.157225       1 logs_generator.go:76] 5 GET /api/v1/namespaces/default/pods/tww 532\nI0212 10:16:54.357193       1 logs_generator.go:76] 6 PUT /api/v1/namespaces/ns/pods/5nn 558\nI0212 10:16:54.557182       1 logs_generator.go:76] 7 POST /api/v1/namespaces/default/pods/q7r 259\nI0212 10:16:54.757180       1 logs_generator.go:76] 8 GET /api/v1/namespaces/kube-system/pods/bkn7 421\nI0212 10:16:54.957183       1 logs_generator.go:76] 9 POST /api/v1/namespaces/ns/pods/nzm 235\nI0212 10:16:55.157202       1 logs_generator.go:76] 10 GET /api/v1/namespaces/kube-system/pods/tkvv 259\nI0212 10:16:55.357190       1 logs_generator.go:76] 11 GET /api/v1/namespaces/kube-system/pods/l2hd 342\nI0212 10:16:55.557225       1 logs_generator.go:76] 12 GET /api/v1/namespaces/ns/pods/bk4c 221\nI0212 10:16:55.757183       1 logs_generator.go:76] 13 POST /api/v1/namespaces/default/pods/r27w 382\nI0212 10:16:55.957180       1 logs_generator.go:76] 14 POST /api/v1/namespaces/default/pods/jb4h 538\n"
+STEP: limiting log lines
+Feb 12 10:16:56.101: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 logs logs-generator logs-generator --namespace=kubectl-3894 --tail=1'
+Feb 12 10:16:56.189: INFO: stderr: ""
+Feb 12 10:16:56.189: INFO: stdout: "I0212 10:16:56.157178       1 logs_generator.go:76] 15 PUT /api/v1/namespaces/ns/pods/58l2 356\n"
+Feb 12 10:16:56.189: INFO: got output "I0212 10:16:56.157178       1 logs_generator.go:76] 15 PUT /api/v1/namespaces/ns/pods/58l2 356\n"
+STEP: limiting log bytes
+Feb 12 10:16:56.189: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 logs logs-generator logs-generator --namespace=kubectl-3894 --limit-bytes=1'
+Feb 12 10:16:56.301: INFO: stderr: ""
+Feb 12 10:16:56.301: INFO: stdout: "I"
+Feb 12 10:16:56.301: INFO: got output "I"
+STEP: exposing timestamps
+Feb 12 10:16:56.301: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 logs logs-generator logs-generator --namespace=kubectl-3894 --tail=1 --timestamps'
+Feb 12 10:16:56.394: INFO: stderr: ""
+Feb 12 10:16:56.394: INFO: stdout: "2021-02-12T10:16:56.357386271Z I0212 10:16:56.357211       1 logs_generator.go:76] 16 PUT /api/v1/namespaces/default/pods/788 565\n"
+Feb 12 10:16:56.394: INFO: got output "2021-02-12T10:16:56.357386271Z I0212 10:16:56.357211       1 logs_generator.go:76] 16 PUT /api/v1/namespaces/default/pods/788 565\n"
+STEP: restricting to a time range
+Feb 12 10:16:58.895: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 logs logs-generator logs-generator --namespace=kubectl-3894 --since=1s'
+Feb 12 10:16:58.993: INFO: stderr: ""
+Feb 12 10:16:58.993: INFO: stdout: "I0212 10:16:58.157148       1 logs_generator.go:76] 25 POST /api/v1/namespaces/default/pods/rhb 300\nI0212 10:16:58.357184       1 logs_generator.go:76] 26 GET /api/v1/namespaces/default/pods/d2xk 347\nI0212 10:16:58.557179       1 logs_generator.go:76] 27 POST /api/v1/namespaces/default/pods/68jj 361\nI0212 10:16:58.758978       1 logs_generator.go:76] 28 POST /api/v1/namespaces/default/pods/g5f 434\nI0212 10:16:58.957197       1 logs_generator.go:76] 29 PUT /api/v1/namespaces/kube-system/pods/hsjx 326\n"
+Feb 12 10:16:58.993: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 logs logs-generator logs-generator --namespace=kubectl-3894 --since=24h'
+Feb 12 10:16:59.089: INFO: stderr: ""
+Feb 12 10:16:59.090: INFO: stdout: "I0212 10:16:53.157041       1 logs_generator.go:76] 0 PUT /api/v1/namespaces/kube-system/pods/q4xx 576\nI0212 10:16:53.357241       1 logs_generator.go:76] 1 POST /api/v1/namespaces/default/pods/f96 285\nI0212 10:16:53.557319       1 logs_generator.go:76] 2 PUT /api/v1/namespaces/default/pods/jhq 258\nI0212 10:16:53.757181       1 logs_generator.go:76] 3 PUT /api/v1/namespaces/ns/pods/frh 378\nI0212 10:16:53.957185       1 logs_generator.go:76] 4 POST /api/v1/namespaces/ns/pods/k9z 271\nI0212 10:16:54.157225       1 logs_generator.go:76] 5 GET /api/v1/namespaces/default/pods/tww 532\nI0212 10:16:54.357193       1 logs_generator.go:76] 6 PUT /api/v1/namespaces/ns/pods/5nn 558\nI0212 10:16:54.557182       1 logs_generator.go:76] 7 POST /api/v1/namespaces/default/pods/q7r 259\nI0212 10:16:54.757180       1 logs_generator.go:76] 8 GET /api/v1/namespaces/kube-system/pods/bkn7 421\nI0212 10:16:54.957183       1 logs_generator.go:76] 9 POST /api/v1/namespaces/ns/pods/nzm 235\nI0212 10:16:55.157202       1 logs_generator.go:76] 10 GET /api/v1/namespaces/kube-system/pods/tkvv 259\nI0212 10:16:55.357190       1 logs_generator.go:76] 11 GET /api/v1/namespaces/kube-system/pods/l2hd 342\nI0212 10:16:55.557225       1 logs_generator.go:76] 12 GET /api/v1/namespaces/ns/pods/bk4c 221\nI0212 10:16:55.757183       1 logs_generator.go:76] 13 POST /api/v1/namespaces/default/pods/r27w 382\nI0212 10:16:55.957180       1 logs_generator.go:76] 14 POST /api/v1/namespaces/default/pods/jb4h 538\nI0212 10:16:56.157178       1 logs_generator.go:76] 15 PUT /api/v1/namespaces/ns/pods/58l2 356\nI0212 10:16:56.357211       1 logs_generator.go:76] 16 PUT /api/v1/namespaces/default/pods/788 565\nI0212 10:16:56.557224       1 logs_generator.go:76] 17 POST /api/v1/namespaces/kube-system/pods/cm79 274\nI0212 10:16:56.757181       1 logs_generator.go:76] 18 GET /api/v1/namespaces/default/pods/hzv 255\nI0212 10:16:56.957186       1 logs_generator.go:76] 19 PUT /api/v1/namespaces/kube-system/pods/xfx 353\nI0212 10:16:57.157174       1 logs_generator.go:76] 20 POST /api/v1/namespaces/kube-system/pods/4lvv 223\nI0212 10:16:57.357169       1 logs_generator.go:76] 21 PUT /api/v1/namespaces/ns/pods/mrh 279\nI0212 10:16:57.557193       1 logs_generator.go:76] 22 PUT /api/v1/namespaces/ns/pods/9n5w 441\nI0212 10:16:57.757172       1 logs_generator.go:76] 23 GET /api/v1/namespaces/kube-system/pods/ssfd 443\nI0212 10:16:57.957191       1 logs_generator.go:76] 24 PUT /api/v1/namespaces/kube-system/pods/72n6 233\nI0212 10:16:58.157148       1 logs_generator.go:76] 25 POST /api/v1/namespaces/default/pods/rhb 300\nI0212 10:16:58.357184       1 logs_generator.go:76] 26 GET /api/v1/namespaces/default/pods/d2xk 347\nI0212 10:16:58.557179       1 logs_generator.go:76] 27 POST /api/v1/namespaces/default/pods/68jj 361\nI0212 10:16:58.758978       1 logs_generator.go:76] 28 POST /api/v1/namespaces/default/pods/g5f 434\nI0212 10:16:58.957197       1 logs_generator.go:76] 29 PUT /api/v1/namespaces/kube-system/pods/hsjx 326\n"
+[AfterEach] Kubectl logs
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1364
+Feb 12 10:16:59.090: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete pod logs-generator --namespace=kubectl-3894'
+Feb 12 10:17:12.250: INFO: stderr: ""
+Feb 12 10:17:12.250: INFO: stdout: "pod \"logs-generator\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:17:12.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-3894" for this suite.
 
-• [SLOW TEST:7.300 seconds]
-[k8s.io] [sig-node] Pods Extended
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  [k8s.io] Delete Grace Period
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-    should be submitted and removed [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:20.641 seconds]
+[sig-cli] Kubectl client
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Kubectl logs
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1354
+    should be able to retrieve and filter logs  [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] [sig-node] Pods Extended [k8s.io] Delete Grace Period should be submitted and removed [Conformance]","total":280,"completed":106,"skipped":1823,"failed":0}
-SS
+{"msg":"PASSED [sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs  [Conformance]","total":280,"completed":81,"skipped":1039,"failed":0}
+SSSSSSSS
+------------------------------
+[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] 
+  should include custom resource definition resources in discovery documents [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:17:12.267: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename custom-resource-definition
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-6950
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should include custom resource definition resources in discovery documents [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: fetching the /apis discovery document
+STEP: finding the apiextensions.k8s.io API group in the /apis discovery document
+STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis discovery document
+STEP: fetching the /apis/apiextensions.k8s.io discovery document
+STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis/apiextensions.k8s.io discovery document
+STEP: fetching the /apis/apiextensions.k8s.io/v1 discovery document
+STEP: finding customresourcedefinitions resources in the /apis/apiextensions.k8s.io/v1 discovery document
+[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:17:12.438: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "custom-resource-definition-6950" for this suite.
+•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance]","total":280,"completed":82,"skipped":1047,"failed":0}
+SSSSSSSSSSSSS
 ------------------------------
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should honor timeout [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should mutate custom resource with pruning [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:00:09.838: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:17:12.455: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-4513
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-1118
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
 STEP: Setting up server cert
 STEP: Create role binding to let webhook read extension-apiserver-authentication
 STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Aug 30 17:00:10.252: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Aug 30 17:00:12.275: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403610, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403610, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403610, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403610, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:17:12.862: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:17:14.878: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721832, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721832, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721832, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721832, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Aug 30 17:00:15.301: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should honor timeout [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Setting timeout (1s) shorter than webhook latency (5s)
-STEP: Registering slow webhook via the AdmissionRegistration API
-Aug 30 17:00:15.414: INFO: Waiting for webhook configuration to be ready...
-Aug 30 17:00:15.537: INFO: Waiting for webhook configuration to be ready...
-STEP: Request fails when timeout (1s) is shorter than slow webhook latency (5s)
-STEP: Having no error when timeout is shorter than webhook latency and failure policy is ignore
-STEP: Registering slow webhook via the AdmissionRegistration API
-STEP: Having no error when timeout is longer than webhook latency
-STEP: Registering slow webhook via the AdmissionRegistration API
-STEP: Having no error when timeout is empty (defaulted to 10s in v1)
-STEP: Registering slow webhook via the AdmissionRegistration API
+Feb 12 10:17:17.899: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should mutate custom resource with pruning [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:17:17.904: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Registering the mutating webhook for custom resource e2e-test-webhook-8883-crds.webhook.example.com via the AdmissionRegistration API
+STEP: Creating a custom resource that should be mutated by the webhook
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:00:28.342: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-4513" for this suite.
-STEP: Destroying namespace "webhook-4513-markers" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:17:19.502: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-1118" for this suite.
+STEP: Destroying namespace "webhook-1118-markers" for this suite.
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:18.587 seconds]
+• [SLOW TEST:7.140 seconds]
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should honor timeout [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should mutate custom resource with pruning [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance]","total":280,"completed":107,"skipped":1825,"failed":0}
-SS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance]","total":280,"completed":83,"skipped":1060,"failed":0}
+SSSSS
 ------------------------------
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  works for multiple CRDs of same group but different versions [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] Deployment 
+  RollingUpdateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:00:28.426: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename crd-publish-openapi
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-1639
+Feb 12 10:17:19.598: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename deployment
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-7336
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] works for multiple CRDs of same group but different versions [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: CRs in the same group but different versions (one multiversion CRD) show up in OpenAPI documentation
-Aug 30 17:00:28.583: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: CRs in the same group but different versions (two CRDs) show up in OpenAPI documentation
-Aug 30 17:00:39.694: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 17:00:42.545: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:00:54.222: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-1639" for this suite.
+[BeforeEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
+[It] RollingUpdateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:17:19.752: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted)
+Feb 12 10:17:19.766: INFO: Pod name sample-pod: Found 0 pods out of 1
+Feb 12 10:17:24.772: INFO: Pod name sample-pod: Found 1 pods out of 1
+STEP: ensuring each pod is running
+Feb 12 10:17:24.772: INFO: Creating deployment "test-rolling-update-deployment"
+Feb 12 10:17:24.778: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has
+Feb 12 10:17:24.799: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created
+Feb 12 10:17:26.809: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected
+Feb 12 10:17:26.816: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721844, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721844, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721844, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721844, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rolling-update-deployment-67cf4f6444\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:17:28.825: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted)
+[AfterEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
+Feb 12 10:17:28.839: INFO: Deployment "test-rolling-update-deployment":
+&Deployment{ObjectMeta:{test-rolling-update-deployment  deployment-7336 /apis/apps/v1/namespaces/deployment-7336/deployments/test-rolling-update-deployment 6e88aaa2-dba2-439a-ae35-68b810609c48 15589 1 2021-02-12 10:17:24 +0000 UTC   map[name:sample-pod] map[deployment.kubernetes.io/revision:3546343826724305833] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc005a27f08  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2021-02-12 10:17:24 +0000 UTC,LastTransitionTime:2021-02-12 10:17:24 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rolling-update-deployment-67cf4f6444" has successfully progressed.,LastUpdateTime:2021-02-12 10:17:26 +0000 UTC,LastTransitionTime:2021-02-12 10:17:24 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},}
+
+Feb 12 10:17:28.846: INFO: New ReplicaSet "test-rolling-update-deployment-67cf4f6444" of Deployment "test-rolling-update-deployment":
+&ReplicaSet{ObjectMeta:{test-rolling-update-deployment-67cf4f6444  deployment-7336 /apis/apps/v1/namespaces/deployment-7336/replicasets/test-rolling-update-deployment-67cf4f6444 6e2542af-e146-496e-9671-2b65f0d16c71 15578 1 2021-02-12 10:17:24 +0000 UTC   map[name:sample-pod pod-template-hash:67cf4f6444] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305833] [{apps/v1 Deployment test-rolling-update-deployment 6e88aaa2-dba2-439a-ae35-68b810609c48 0xc00595c4d7 0xc00595c4d8}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 67cf4f6444,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod pod-template-hash:67cf4f6444] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00595c558  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},}
+Feb 12 10:17:28.846: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment":
+Feb 12 10:17:28.846: INFO: &ReplicaSet{ObjectMeta:{test-rolling-update-controller  deployment-7336 /apis/apps/v1/namespaces/deployment-7336/replicasets/test-rolling-update-controller 778185fa-01b6-4425-95b8-50a4fe5172d0 15588 2 2021-02-12 10:17:19 +0000 UTC   map[name:sample-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305832] [{apps/v1 Deployment test-rolling-update-deployment 6e88aaa2-dba2-439a-ae35-68b810609c48 0xc00595c3d7 0xc00595c3d8}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod pod:httpd] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc00595c458  ClusterFirst map[]     false false false  PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
+Feb 12 10:17:28.853: INFO: Pod "test-rolling-update-deployment-67cf4f6444-fhqdm" is available:
+&Pod{ObjectMeta:{test-rolling-update-deployment-67cf4f6444-fhqdm test-rolling-update-deployment-67cf4f6444- deployment-7336 /api/v1/namespaces/deployment-7336/pods/test-rolling-update-deployment-67cf4f6444-fhqdm 81c3d697-6695-4812-9cf7-d259aec21b69 15577 0 2021-02-12 10:17:24 +0000 UTC   map[name:sample-pod pod-template-hash:67cf4f6444] map[cni.projectcalico.org/podIP:172.25.1.71/32] [{apps/v1 ReplicaSet test-rolling-update-deployment-67cf4f6444 6e2542af-e146-496e-9671-2b65f0d16c71 0xc0059ace97 0xc0059ace98}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-wldgs,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-wldgs,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-wldgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-xg74t,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:17:24 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:17:26 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:17:26 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:17:24 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:138.68.97.197,PodIP:172.25.1.71,StartTime:2021-02-12 10:17:24 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-02-12 10:17:26 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/agnhost@sha256:daf5332100521b1256d0e3c56d697a238eaec3af48897ed9167cbadd426773b5,ContainerID:docker://f9a7a783fe1dce4fb1eb853c4c49286aad835a8821617314a0461acbb2aa113f,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.71,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+[AfterEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:17:28.854: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-7336" for this suite.
 
-• [SLOW TEST:25.811 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for multiple CRDs of same group but different versions [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:9.275 seconds]
+[sig-apps] Deployment
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  RollingUpdateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance]","total":280,"completed":108,"skipped":1827,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance]","total":280,"completed":84,"skipped":1065,"failed":0}
+SS
 ------------------------------
-[sig-api-machinery] ResourceQuota 
-  should create a ResourceQuota and capture the life of a replication controller. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] Deployment 
+  RecreateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:00:54.239: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename resourcequota
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-3572
+Feb 12 10:17:28.873: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename deployment
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-2680
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should create a ResourceQuota and capture the life of a replication controller. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Counting existing ResourceQuota
-STEP: Creating a ResourceQuota
-STEP: Ensuring resource quota status is calculated
-STEP: Creating a ReplicationController
-STEP: Ensuring resource quota status captures replication controller creation
-STEP: Deleting a ReplicationController
-STEP: Ensuring resource quota status released usage
-[AfterEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:01:05.465: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-3572" for this suite.
-
-• [SLOW TEST:11.241 seconds]
-[sig-api-machinery] ResourceQuota
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should create a ResourceQuota and capture the life of a replication controller. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance]","total":280,"completed":109,"skipped":1882,"failed":0}
-SSS
+[BeforeEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
+[It] RecreateDeployment should delete old pods and create new ones [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:17:29.028: INFO: Creating deployment "test-recreate-deployment"
+Feb 12 10:17:29.036: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1
+Feb 12 10:17:29.051: INFO: deployment "test-recreate-deployment" doesn't have the required revision set
+Feb 12 10:17:31.061: INFO: Waiting deployment "test-recreate-deployment" to complete
+Feb 12 10:17:31.068: INFO: Triggering a new rollout for deployment "test-recreate-deployment"
+Feb 12 10:17:31.083: INFO: Updating deployment test-recreate-deployment
+Feb 12 10:17:31.083: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods
+[AfterEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
+Feb 12 10:17:31.194: INFO: Deployment "test-recreate-deployment":
+&Deployment{ObjectMeta:{test-recreate-deployment  deployment-2680 /apis/apps/v1/namespaces/deployment-2680/deployments/test-recreate-deployment 396caece-7e22-4fc7-97de-5916e2cb5722 15656 2 2021-02-12 10:17:28 +0000 UTC   map[name:sample-pod-3] map[deployment.kubernetes.io/revision:2] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod-3] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00370ccf8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2021-02-12 10:17:31 +0000 UTC,LastTransitionTime:2021-02-12 10:17:31 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "test-recreate-deployment-5f94c574ff" is progressing.,LastUpdateTime:2021-02-12 10:17:31 +0000 UTC,LastTransitionTime:2021-02-12 10:17:29 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},}
+
+Feb 12 10:17:31.204: INFO: New ReplicaSet "test-recreate-deployment-5f94c574ff" of Deployment "test-recreate-deployment":
+&ReplicaSet{ObjectMeta:{test-recreate-deployment-5f94c574ff  deployment-2680 /apis/apps/v1/namespaces/deployment-2680/replicasets/test-recreate-deployment-5f94c574ff 1628c57f-e8c7-4291-a167-4ca628113409 15653 1 2021-02-12 10:17:31 +0000 UTC   map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-recreate-deployment 396caece-7e22-4fc7-97de-5916e2cb5722 0xc00370d2b7 0xc00370d2b8}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 5f94c574ff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00370d328  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
+Feb 12 10:17:31.204: INFO: All old ReplicaSets of Deployment "test-recreate-deployment":
+Feb 12 10:17:31.204: INFO: &ReplicaSet{ObjectMeta:{test-recreate-deployment-799c574856  deployment-2680 /apis/apps/v1/namespaces/deployment-2680/replicasets/test-recreate-deployment-799c574856 e40256c0-dd2d-4455-ac5b-6f4aa6eee12c 15644 2 2021-02-12 10:17:29 +0000 UTC   map[name:sample-pod-3 pod-template-hash:799c574856] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-recreate-deployment 396caece-7e22-4fc7-97de-5916e2cb5722 0xc00370d397 0xc00370d398}] []  []},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 799c574856,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:sample-pod-3 pod-template-hash:799c574856] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00370d438  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
+Feb 12 10:17:31.214: INFO: Pod "test-recreate-deployment-5f94c574ff-cr899" is not available:
+&Pod{ObjectMeta:{test-recreate-deployment-5f94c574ff-cr899 test-recreate-deployment-5f94c574ff- deployment-2680 /api/v1/namespaces/deployment-2680/pods/test-recreate-deployment-5f94c574ff-cr899 a1e7be6b-48d1-4551-97fd-24ab5d5b4e40 15655 0 2021-02-12 10:17:31 +0000 UTC   map[name:sample-pod-3 pod-template-hash:5f94c574ff] map[] [{apps/v1 ReplicaSet test-recreate-deployment-5f94c574ff 1628c57f-e8c7-4291-a167-4ca628113409 0xc0036f05d7 0xc0036f05d8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-tmz4s,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-tmz4s,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-tmz4s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-dbhjj,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:17:31 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:17:31 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:17:31 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:17:31 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:104.248.134.247,PodIP:,StartTime:2021-02-12 10:17:31 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+[AfterEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:17:31.214: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-2680" for this suite.
+•{"msg":"PASSED [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance]","total":280,"completed":85,"skipped":1067,"failed":0}
+SSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] 
-  should be able to convert a non homogeneous list of CRs [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  patching/updating a validating webhook should work [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:01:05.480: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename crd-webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-webhook-1607
+Feb 12 10:17:31.236: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-762
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:125
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
 STEP: Setting up server cert
-STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication
-STEP: Deploying the custom resource conversion webhook pod
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Aug 30 17:01:06.574: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set
-Aug 30 17:01:08.589: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403666, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403666, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403666, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403666, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-78dcf5dd84\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:17:31.867: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:17:33.884: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721851, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721851, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721851, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748721851, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Aug 30 17:01:11.609: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1
-[It] should be able to convert a non homogeneous list of CRs [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:01:11.614: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Creating a v1 custom resource
-STEP: Create a v2 custom resource
-STEP: List CRs in v1
-STEP: List CRs in v2
-[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:01:13.226: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-webhook-1607" for this suite.
-[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:136
+Feb 12 10:17:36.911: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] patching/updating a validating webhook should work [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a validating webhook configuration
+STEP: Creating a configMap that does not comply to the validation webhook rules
+STEP: Updating a validating webhook configuration's rules to not include the create operation
+STEP: Creating a configMap that does not comply to the validation webhook rules
+STEP: Patching a validating webhook configuration's rules to include the create operation
+STEP: Creating a configMap that does not comply to the validation webhook rules
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:17:37.311: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-762" for this suite.
+STEP: Destroying namespace "webhook-762-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:7.822 seconds]
-[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should be able to convert a non homogeneous list of CRs [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:6.219 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  patching/updating a validating webhook should work [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance]","total":280,"completed":110,"skipped":1885,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance]","total":280,"completed":86,"skipped":1083,"failed":0}
+SSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl label 
-  should update the label on a resource  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] [sig-node] Pods Extended [k8s.io] Delete Grace Period 
+  should be submitted and removed [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] [sig-node] Pods Extended
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:01:13.302: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-57
+Feb 12 10:17:37.455: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-2365
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[BeforeEach] Kubectl label
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1275
+[BeforeEach] [k8s.io] Delete Grace Period
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pods.go:53
+[It] should be submitted and removed [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: creating the pod
-Aug 30 17:01:13.470: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-57'
-Aug 30 17:01:13.742: INFO: stderr: ""
-Aug 30 17:01:13.742: INFO: stdout: "pod/pause created\n"
-Aug 30 17:01:13.742: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause]
-Aug 30 17:01:13.742: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-57" to be "running and ready"
-Aug 30 17:01:13.747: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 5.3094ms
-Aug 30 17:01:15.753: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011011236s
-Aug 30 17:01:17.758: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 4.016095287s
-Aug 30 17:01:17.758: INFO: Pod "pause" satisfied condition "running and ready"
-Aug 30 17:01:17.758: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause]
-[It] should update the label on a resource  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: adding the label testing-label with value testing-label-value to a pod
-Aug 30 17:01:17.758: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 label pods pause testing-label=testing-label-value --namespace=kubectl-57'
-Aug 30 17:01:17.832: INFO: stderr: ""
-Aug 30 17:01:17.832: INFO: stdout: "pod/pause labeled\n"
-STEP: verifying the pod has the label testing-label with the value testing-label-value
-Aug 30 17:01:17.832: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pod pause -L testing-label --namespace=kubectl-57'
-Aug 30 17:01:17.915: INFO: stderr: ""
-Aug 30 17:01:17.915: INFO: stdout: "NAME    READY   STATUS    RESTARTS   AGE   TESTING-LABEL\npause   1/1     Running   0          4s    testing-label-value\n"
-STEP: removing the label testing-label of a pod
-Aug 30 17:01:17.915: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 label pods pause testing-label- --namespace=kubectl-57'
-Aug 30 17:01:17.990: INFO: stderr: ""
-Aug 30 17:01:17.990: INFO: stdout: "pod/pause labeled\n"
-STEP: verifying the pod doesn't have the label testing-label
-Aug 30 17:01:17.990: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pod pause -L testing-label --namespace=kubectl-57'
-Aug 30 17:01:18.058: INFO: stderr: ""
-Aug 30 17:01:18.058: INFO: stdout: "NAME    READY   STATUS    RESTARTS   AGE   TESTING-LABEL\npause   1/1     Running   0          5s    \n"
-[AfterEach] Kubectl label
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1282
-STEP: using delete to clean up resources
-Aug 30 17:01:18.058: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete --grace-period=0 --force -f - --namespace=kubectl-57'
-Aug 30 17:01:18.135: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Aug 30 17:01:18.135: INFO: stdout: "pod \"pause\" force deleted\n"
-Aug 30 17:01:18.135: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get rc,svc -l name=pause --no-headers --namespace=kubectl-57'
-Aug 30 17:01:18.205: INFO: stderr: "No resources found in kubectl-57 namespace.\n"
-Aug 30 17:01:18.205: INFO: stdout: ""
-Aug 30 17:01:18.205: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -l name=pause --namespace=kubectl-57 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
-Aug 30 17:01:18.267: INFO: stderr: ""
-Aug 30 17:01:18.267: INFO: stdout: ""
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:01:18.267: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-57" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl label should update the label on a resource  [Conformance]","total":280,"completed":111,"skipped":1919,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+STEP: setting up selector
+STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+Feb 12 10:17:39.643: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-008079477 proxy -p 0'
+STEP: deleting the pod gracefully
+STEP: verifying the kubelet observed the termination notice
+Feb 12 10:17:54.731: INFO: no pod exists with the name we were looking for, assuming the termination request was observed and completed
+[AfterEach] [k8s.io] [sig-node] Pods Extended
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:17:54.736: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-2365" for this suite.
+
+• [SLOW TEST:17.299 seconds]
+[k8s.io] [sig-node] Pods Extended
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  [k8s.io] Delete Grace Period
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+    should be submitted and removed [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-network] Networking Granular Checks: Pods 
-  should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Networking
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [k8s.io] [sig-node] Pods Extended [k8s.io] Delete Grace Period should be submitted and removed [Conformance]","total":280,"completed":87,"skipped":1092,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir wrapper volumes 
+  should not cause race condition when used for configmaps [Serial] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir wrapper volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:01:18.286: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename pod-network-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-4851
+Feb 12 10:17:54.755: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir-wrapper
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-wrapper-8995
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Performing setup for networking test in namespace pod-network-test-4851
-STEP: creating a selector
-STEP: Creating the service pods in kubernetes
-Aug 30 17:01:18.445: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
-STEP: Creating test pods
-Aug 30 17:01:40.571: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.25.1.44:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-4851 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 17:01:40.571: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 17:01:41.062: INFO: Found all expected endpoints: [netserver-0]
-Aug 30 17:01:41.068: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.25.0.101:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-4851 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 17:01:41.068: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 17:01:41.625: INFO: Found all expected endpoints: [netserver-1]
-[AfterEach] [sig-network] Networking
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:01:41.625: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pod-network-test-4851" for this suite.
+[It] should not cause race condition when used for configmaps [Serial] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating 50 configmaps
+STEP: Creating RC which spawns configmap-volume pods
+Feb 12 10:17:55.301: INFO: Pod name wrapped-volume-race-ccd67b99-94f4-4c2d-b2de-4badb0779c15: Found 0 pods out of 5
+Feb 12 10:18:00.310: INFO: Pod name wrapped-volume-race-ccd67b99-94f4-4c2d-b2de-4badb0779c15: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-ccd67b99-94f4-4c2d-b2de-4badb0779c15 in namespace emptydir-wrapper-8995, will wait for the garbage collector to delete the pods
+Feb 12 10:18:12.416: INFO: Deleting ReplicationController wrapped-volume-race-ccd67b99-94f4-4c2d-b2de-4badb0779c15 took: 15.165241ms
+Feb 12 10:18:12.916: INFO: Terminating ReplicationController wrapped-volume-race-ccd67b99-94f4-4c2d-b2de-4badb0779c15 pods took: 500.200301ms
+STEP: Creating RC which spawns configmap-volume pods
+Feb 12 10:18:23.339: INFO: Pod name wrapped-volume-race-e2b323f7-1205-411c-b2c3-b6ed2c1a660f: Found 0 pods out of 5
+Feb 12 10:18:28.351: INFO: Pod name wrapped-volume-race-e2b323f7-1205-411c-b2c3-b6ed2c1a660f: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-e2b323f7-1205-411c-b2c3-b6ed2c1a660f in namespace emptydir-wrapper-8995, will wait for the garbage collector to delete the pods
+Feb 12 10:18:38.471: INFO: Deleting ReplicationController wrapped-volume-race-e2b323f7-1205-411c-b2c3-b6ed2c1a660f took: 13.862105ms
+Feb 12 10:18:38.972: INFO: Terminating ReplicationController wrapped-volume-race-e2b323f7-1205-411c-b2c3-b6ed2c1a660f pods took: 500.230374ms
+STEP: Creating RC which spawns configmap-volume pods
+Feb 12 10:18:52.298: INFO: Pod name wrapped-volume-race-da95039a-fdd2-4c4b-a509-5ab644c60148: Found 0 pods out of 5
+Feb 12 10:18:57.307: INFO: Pod name wrapped-volume-race-da95039a-fdd2-4c4b-a509-5ab644c60148: Found 5 pods out of 5
+STEP: Ensuring each pod is running
+STEP: deleting ReplicationController wrapped-volume-race-da95039a-fdd2-4c4b-a509-5ab644c60148 in namespace emptydir-wrapper-8995, will wait for the garbage collector to delete the pods
+Feb 12 10:19:09.415: INFO: Deleting ReplicationController wrapped-volume-race-da95039a-fdd2-4c4b-a509-5ab644c60148 took: 15.31994ms
+Feb 12 10:19:09.916: INFO: Terminating ReplicationController wrapped-volume-race-da95039a-fdd2-4c4b-a509-5ab644c60148 pods took: 500.443258ms
+STEP: Cleaning up the configMaps
+[AfterEach] [sig-storage] EmptyDir wrapper volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:19:23.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-wrapper-8995" for this suite.
 
-• [SLOW TEST:23.354 seconds]
-[sig-network] Networking
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
-  Granular Checks: Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
-    should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:88.283 seconds]
+[sig-storage] EmptyDir wrapper volumes
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
+  should not cause race condition when used for configmaps [Serial] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":112,"skipped":1965,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance]","total":280,"completed":88,"skipped":1126,"failed":0}
+SSSSSSSSS
 ------------------------------
-[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
-  should have an terminated reason [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] Daemon set [Serial] 
+  should run and stop complex daemon [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:01:41.640: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubelet-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-438
+Feb 12 10:19:23.041: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename daemonsets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-9847
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
-[BeforeEach] when scheduling a busybox command that always fails in a pod
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
-[It] should have an terminated reason [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:01:45.822: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubelet-test-438" for this suite.
-•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance]","total":280,"completed":113,"skipped":1992,"failed":0}
-SS
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
+[It] should run and stop complex daemon [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:19:23.228: INFO: Creating daemon "daemon-set" with a node selector
+STEP: Initially, daemon pods should not be running on any nodes.
+Feb 12 10:19:23.241: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:23.241: INFO: Number of running nodes: 0, number of available pods: 0
+STEP: Change node label to blue, check that daemon pod is launched.
+Feb 12 10:19:23.267: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:23.267: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:19:24.272: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:24.273: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:19:25.273: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:25.273: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:19:26.273: INFO: Number of nodes with available pods: 1
+Feb 12 10:19:26.273: INFO: Number of running nodes: 1, number of available pods: 1
+STEP: Update the node label to green, and wait for daemons to be unscheduled
+Feb 12 10:19:26.304: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:26.307: INFO: Number of running nodes: 0, number of available pods: 0
+STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate
+Feb 12 10:19:26.322: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:26.322: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:19:27.329: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:27.329: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:19:28.328: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:28.328: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:19:29.329: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:29.329: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:19:30.327: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:30.327: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:19:31.328: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:31.329: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:19:32.329: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:32.329: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:19:33.330: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:33.330: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:19:34.328: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:34.328: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:19:35.328: INFO: Number of nodes with available pods: 1
+Feb 12 10:19:35.328: INFO: Number of running nodes: 1, number of available pods: 1
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-9847, will wait for the garbage collector to delete the pods
+Feb 12 10:19:35.407: INFO: Deleting DaemonSet.extensions daemon-set took: 14.297103ms
+Feb 12 10:19:35.908: INFO: Terminating DaemonSet.extensions daemon-set pods took: 500.278642ms
+Feb 12 10:19:38.614: INFO: Number of nodes with available pods: 0
+Feb 12 10:19:38.614: INFO: Number of running nodes: 0, number of available pods: 0
+Feb 12 10:19:38.619: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-9847/daemonsets","resourceVersion":"17328"},"items":null}
+
+Feb 12 10:19:38.625: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-9847/pods","resourceVersion":"17328"},"items":null}
+
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:19:38.661: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "daemonsets-9847" for this suite.
+
+• [SLOW TEST:15.642 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should run and stop complex daemon [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Projected configMap 
-  should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance]","total":280,"completed":89,"skipped":1135,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:01:45.837: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3218
+Feb 12 10:19:38.687: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-8480
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name projected-configmap-test-volume-map-952c75f3-d0da-43e9-954e-44216fdaa816
-STEP: Creating a pod to test consume configMaps
-Aug 30 17:01:46.006: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-866744b3-90b6-4cdb-b8d9-46af8d4ec473" in namespace "projected-3218" to be "success or failure"
-Aug 30 17:01:46.011: INFO: Pod "pod-projected-configmaps-866744b3-90b6-4cdb-b8d9-46af8d4ec473": Phase="Pending", Reason="", readiness=false. Elapsed: 4.851974ms
-Aug 30 17:01:48.017: INFO: Pod "pod-projected-configmaps-866744b3-90b6-4cdb-b8d9-46af8d4ec473": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010980775s
+[It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir 0777 on tmpfs
+Feb 12 10:19:38.854: INFO: Waiting up to 5m0s for pod "pod-48cc16da-b865-46d7-bd0e-cae3d53b199c" in namespace "emptydir-8480" to be "success or failure"
+Feb 12 10:19:38.860: INFO: Pod "pod-48cc16da-b865-46d7-bd0e-cae3d53b199c": Phase="Pending", Reason="", readiness=false. Elapsed: 5.089863ms
+Feb 12 10:19:40.864: INFO: Pod "pod-48cc16da-b865-46d7-bd0e-cae3d53b199c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.009930366s
 STEP: Saw pod success
-Aug 30 17:01:48.018: INFO: Pod "pod-projected-configmaps-866744b3-90b6-4cdb-b8d9-46af8d4ec473" satisfied condition "success or failure"
-Aug 30 17:01:48.023: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod pod-projected-configmaps-866744b3-90b6-4cdb-b8d9-46af8d4ec473 container projected-configmap-volume-test: 
+Feb 12 10:19:40.865: INFO: Pod "pod-48cc16da-b865-46d7-bd0e-cae3d53b199c" satisfied condition "success or failure"
+Feb 12 10:19:40.869: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-48cc16da-b865-46d7-bd0e-cae3d53b199c container test-container: 
 STEP: delete the pod
-Aug 30 17:01:48.089: INFO: Waiting for pod pod-projected-configmaps-866744b3-90b6-4cdb-b8d9-46af8d4ec473 to disappear
-Aug 30 17:01:48.094: INFO: Pod pod-projected-configmaps-866744b3-90b6-4cdb-b8d9-46af8d4ec473 no longer exists
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:01:48.094: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-3218" for this suite.
-•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","total":280,"completed":114,"skipped":1994,"failed":0}
-
+Feb 12 10:19:40.939: INFO: Waiting for pod pod-48cc16da-b865-46d7-bd0e-cae3d53b199c to disappear
+Feb 12 10:19:40.943: INFO: Pod pod-48cc16da-b865-46d7-bd0e-cae3d53b199c no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:19:40.943: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-8480" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":90,"skipped":1163,"failed":0}
+SSSSSSS
 ------------------------------
-[k8s.io] Container Runtime blackbox test on terminated container 
-  should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  works for CRD preserving unknown fields in an embedded object [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:01:48.111: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-runtime
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-6797
+Feb 12 10:19:40.961: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename crd-publish-openapi
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-9995
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the container
-STEP: wait for the container to reach Succeeded
-STEP: get the container status
-STEP: the container should be terminated
-STEP: the termination message should be set
-Aug 30 17:01:50.313: INFO: Expected: &{OK} to match Container's Termination Message: OK --
-STEP: delete the container
-[AfterEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:01:50.332: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-runtime-6797" for this suite.
-•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":280,"completed":115,"skipped":1994,"failed":0}
-SSSSSSSS
+[It] works for CRD preserving unknown fields in an embedded object [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:19:41.115: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: client-side validation (kubectl create and apply) allows request with any unknown properties
+Feb 12 10:19:43.972: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-9995 create -f -'
+Feb 12 10:19:44.447: INFO: stderr: ""
+Feb 12 10:19:44.447: INFO: stdout: "e2e-test-crd-publish-openapi-814-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n"
+Feb 12 10:19:44.447: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-9995 delete e2e-test-crd-publish-openapi-814-crds test-cr'
+Feb 12 10:19:44.532: INFO: stderr: ""
+Feb 12 10:19:44.532: INFO: stdout: "e2e-test-crd-publish-openapi-814-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n"
+Feb 12 10:19:44.532: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-9995 apply -f -'
+Feb 12 10:19:44.691: INFO: stderr: ""
+Feb 12 10:19:44.691: INFO: stdout: "e2e-test-crd-publish-openapi-814-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n"
+Feb 12 10:19:44.691: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-9995 delete e2e-test-crd-publish-openapi-814-crds test-cr'
+Feb 12 10:19:44.775: INFO: stderr: ""
+Feb 12 10:19:44.775: INFO: stdout: "e2e-test-crd-publish-openapi-814-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n"
+STEP: kubectl explain works to explain CR
+Feb 12 10:19:44.775: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 explain e2e-test-crd-publish-openapi-814-crds'
+Feb 12 10:19:44.953: INFO: stderr: ""
+Feb 12 10:19:44.953: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-814-crd\nVERSION:  crd-publish-openapi-test-unknown-in-nested.example.com/v1\n\nDESCRIPTION:\n     preserve-unknown-properties in nested field for Testing\n\nFIELDS:\n   apiVersion\t\n     APIVersion defines the versioned schema of this representation of an\n     object. Servers should convert recognized schemas to the latest internal\n     value, and may reject unrecognized values. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n   kind\t\n     Kind is a string value representing the REST resource this object\n     represents. Servers may infer this from the endpoint the client submits\n     requests to. Cannot be updated. In CamelCase. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n   metadata\t\n     Standard object's metadata. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   spec\t\n     Specification of Waldo\n\n   status\t\n     Status of Waldo\n\n"
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:19:47.782: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-9995" for this suite.
+
+• [SLOW TEST:6.841 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  works for CRD preserving unknown fields in an embedded object [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance]","total":280,"completed":91,"skipped":1170,"failed":0}
+SSSSSSS
 ------------------------------
 [sig-network] Services 
-  should be able to create a functioning NodePort service [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should be able to change the type from ClusterIP to ExternalName [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:01:50.344: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:19:47.804: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename services
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-3363
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-4291
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should be able to create a functioning NodePort service [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating service nodeport-test with type=NodePort in namespace services-3363
-STEP: creating replication controller nodeport-test in namespace services-3363
-I0830 17:01:50.529824      23 runners.go:189] Created replication controller with name: nodeport-test, namespace: services-3363, replica count: 2
-I0830 17:01:53.580185      23 runners.go:189] nodeport-test Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-Aug 30 17:01:53.580: INFO: Creating new exec pod
-Aug 30 17:01:58.603: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=services-3363 execpod6b9zn -- /bin/sh -x -c nc -zv -t -w 2 nodeport-test 80'
-Aug 30 17:01:59.165: INFO: stderr: "+ nc -zv -t -w 2 nodeport-test 80\nConnection to nodeport-test 80 port [tcp/http] succeeded!\n"
-Aug 30 17:01:59.165: INFO: stdout: ""
-Aug 30 17:01:59.165: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=services-3363 execpod6b9zn -- /bin/sh -x -c nc -zv -t -w 2 10.240.23.81 80'
-Aug 30 17:01:59.774: INFO: stderr: "+ nc -zv -t -w 2 10.240.23.81 80\nConnection to 10.240.23.81 80 port [tcp/http] succeeded!\n"
-Aug 30 17:01:59.774: INFO: stdout: ""
-Aug 30 17:01:59.774: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=services-3363 execpod6b9zn -- /bin/sh -x -c nc -zv -t -w 2 46.101.153.64 31140'
-Aug 30 17:02:00.410: INFO: stderr: "+ nc -zv -t -w 2 46.101.153.64 31140\nConnection to 46.101.153.64 31140 port [tcp/31140] succeeded!\n"
-Aug 30 17:02:00.410: INFO: stdout: ""
-Aug 30 17:02:00.410: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=services-3363 execpod6b9zn -- /bin/sh -x -c nc -zv -t -w 2 46.101.135.210 31140'
-Aug 30 17:02:00.936: INFO: stderr: "+ nc -zv -t -w 2 46.101.135.210 31140\nConnection to 46.101.135.210 31140 port [tcp/31140] succeeded!\n"
-Aug 30 17:02:00.937: INFO: stdout: ""
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
+[It] should be able to change the type from ClusterIP to ExternalName [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating a service clusterip-service with the type=ClusterIP in namespace services-4291
+STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service
+STEP: creating service externalsvc in namespace services-4291
+STEP: creating replication controller externalsvc in namespace services-4291
+I0212 10:19:48.009713      21 runners.go:189] Created replication controller with name: externalsvc, namespace: services-4291, replica count: 2
+I0212 10:19:51.060315      21 runners.go:189] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+STEP: changing the ClusterIP service to type=ExternalName
+Feb 12 10:19:51.088: INFO: Creating new exec pod
+Feb 12 10:19:53.113: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=services-4291 execpodsb7f6 -- /bin/sh -x -c nslookup clusterip-service'
+Feb 12 10:19:53.706: INFO: stderr: "+ nslookup clusterip-service\n"
+Feb 12 10:19:53.706: INFO: stdout: "Server:\t\t10.240.16.10\nAddress:\t10.240.16.10#53\n\nclusterip-service.services-4291.svc.cluster.local\tcanonical name = externalsvc.services-4291.svc.cluster.local.\nName:\texternalsvc.services-4291.svc.cluster.local\nAddress: 10.240.18.165\n\n"
+STEP: deleting ReplicationController externalsvc in namespace services-4291, will wait for the garbage collector to delete the pods
+Feb 12 10:19:53.774: INFO: Deleting ReplicationController externalsvc took: 12.790633ms
+Feb 12 10:19:54.274: INFO: Terminating ReplicationController externalsvc pods took: 500.155755ms
+Feb 12 10:20:02.406: INFO: Cleaning up the ClusterIP to ExternalName test service
 [AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:00.937: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-3363" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:20:02.434: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-4291" for this suite.
 [AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
 
-• [SLOW TEST:10.607 seconds]
+• [SLOW TEST:14.658 seconds]
 [sig-network] Services
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should be able to create a functioning NodePort service [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should be able to change the type from ClusterIP to ExternalName [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Services should be able to create a functioning NodePort service [Conformance]","total":280,"completed":116,"skipped":2002,"failed":0}
-SSSSSSSSSSS
+{"msg":"PASSED [sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance]","total":280,"completed":92,"skipped":1177,"failed":0}
+SSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] Proxy version v1 
+  should proxy through a service and a pod  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] version v1
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:02:00.951: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-3687
+Feb 12 10:20:02.464: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename proxy
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-1718
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0644 on node default medium
-Aug 30 17:02:01.123: INFO: Waiting up to 5m0s for pod "pod-11d5a9ec-5053-4477-9667-fc1f8b575e00" in namespace "emptydir-3687" to be "success or failure"
-Aug 30 17:02:01.129: INFO: Pod "pod-11d5a9ec-5053-4477-9667-fc1f8b575e00": Phase="Pending", Reason="", readiness=false. Elapsed: 5.556649ms
-Aug 30 17:02:03.134: INFO: Pod "pod-11d5a9ec-5053-4477-9667-fc1f8b575e00": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010437215s
-STEP: Saw pod success
-Aug 30 17:02:03.134: INFO: Pod "pod-11d5a9ec-5053-4477-9667-fc1f8b575e00" satisfied condition "success or failure"
-Aug 30 17:02:03.138: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-11d5a9ec-5053-4477-9667-fc1f8b575e00 container test-container: 
-STEP: delete the pod
-Aug 30 17:02:03.206: INFO: Waiting for pod pod-11d5a9ec-5053-4477-9667-fc1f8b575e00 to disappear
-Aug 30 17:02:03.209: INFO: Pod pod-11d5a9ec-5053-4477-9667-fc1f8b575e00 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:03.209: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-3687" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":117,"skipped":2013,"failed":0}
-S
+[It] should proxy through a service and a pod  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: starting an echo server on multiple ports
+STEP: creating replication controller proxy-service-dcbp5 in namespace proxy-1718
+I0212 10:20:02.682969      21 runners.go:189] Created replication controller with name: proxy-service-dcbp5, namespace: proxy-1718, replica count: 1
+I0212 10:20:03.737193      21 runners.go:189] proxy-service-dcbp5 Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+I0212 10:20:04.737479      21 runners.go:189] proxy-service-dcbp5 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0212 10:20:05.737875      21 runners.go:189] proxy-service-dcbp5 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0212 10:20:06.738290      21 runners.go:189] proxy-service-dcbp5 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0212 10:20:07.738646      21 runners.go:189] proxy-service-dcbp5 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0212 10:20:08.738892      21 runners.go:189] proxy-service-dcbp5 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0212 10:20:09.739097      21 runners.go:189] proxy-service-dcbp5 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0212 10:20:10.739317      21 runners.go:189] proxy-service-dcbp5 Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
+I0212 10:20:11.739525      21 runners.go:189] proxy-service-dcbp5 Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+Feb 12 10:20:11.746: INFO: setup took 9.112280578s, starting test cases
+STEP: running 16 cases, 20 attempts per case, 320 total attempts
+Feb 12 10:20:11.808: INFO: (0) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 61.317572ms)
+Feb 12 10:20:11.808: INFO: (0) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 61.65124ms)
+Feb 12 10:20:11.808: INFO: (0) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 61.296912ms)
+Feb 12 10:20:11.808: INFO: (0) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 61.636951ms)
+Feb 12 10:20:11.808: INFO: (0) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:1080/proxy/: test<... (200; 61.627086ms)
+Feb 12 10:20:11.808: INFO: (0) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 61.400464ms)
+Feb 12 10:20:11.816: INFO: (0) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 69.0764ms)
+Feb 12 10:20:11.816: INFO: (0) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 69.096406ms)
+Feb 12 10:20:11.816: INFO: (0) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 69.039116ms)
+Feb 12 10:20:11.816: INFO: (0) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 69.259909ms)
+Feb 12 10:20:11.818: INFO: (0) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname2/proxy/: bar (200; 71.476569ms)
+Feb 12 10:20:11.818: INFO: (0) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test (200; 10.09006ms)
+Feb 12 10:20:11.833: INFO: (1) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 11.71397ms)
+Feb 12 10:20:11.833: INFO: (1) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test<... (200; 12.61526ms)
+Feb 12 10:20:11.834: INFO: (1) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname2/proxy/: tls qux (200; 12.650929ms)
+Feb 12 10:20:11.834: INFO: (1) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 12.3139ms)
+Feb 12 10:20:11.837: INFO: (1) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname1/proxy/: tls baz (200; 16.418408ms)
+Feb 12 10:20:11.837: INFO: (1) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 16.190274ms)
+Feb 12 10:20:11.839: INFO: (1) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 17.810926ms)
+Feb 12 10:20:11.839: INFO: (1) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname2/proxy/: bar (200; 18.047203ms)
+Feb 12 10:20:11.839: INFO: (1) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 18.359994ms)
+Feb 12 10:20:11.885: INFO: (2) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 45.897268ms)
+Feb 12 10:20:11.886: INFO: (2) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:1080/proxy/: test<... (200; 46.463239ms)
+Feb 12 10:20:11.886: INFO: (2) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 45.754322ms)
+Feb 12 10:20:11.886: INFO: (2) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 46.162334ms)
+Feb 12 10:20:11.885: INFO: (2) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 45.658173ms)
+Feb 12 10:20:11.888: INFO: (2) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test (200; 48.17919ms)
+Feb 12 10:20:11.890: INFO: (2) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname1/proxy/: tls baz (200; 50.592512ms)
+Feb 12 10:20:11.893: INFO: (2) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 53.363237ms)
+Feb 12 10:20:11.927: INFO: (2) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 87.341323ms)
+Feb 12 10:20:11.927: INFO: (2) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 87.173233ms)
+Feb 12 10:20:11.927: INFO: (2) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 87.315604ms)
+Feb 12 10:20:11.940: INFO: (3) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 12.708293ms)
+Feb 12 10:20:11.941: INFO: (3) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 13.26233ms)
+Feb 12 10:20:11.941: INFO: (3) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 13.169264ms)
+Feb 12 10:20:11.941: INFO: (3) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname2/proxy/: bar (200; 13.216493ms)
+Feb 12 10:20:11.941: INFO: (3) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 13.472967ms)
+Feb 12 10:20:11.941: INFO: (3) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: ... (200; 13.775674ms)
+Feb 12 10:20:11.941: INFO: (3) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:1080/proxy/: test<... (200; 14.099881ms)
+Feb 12 10:20:11.942: INFO: (3) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 13.991996ms)
+Feb 12 10:20:11.945: INFO: (3) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname2/proxy/: tls qux (200; 18.132767ms)
+Feb 12 10:20:11.946: INFO: (3) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 18.235589ms)
+Feb 12 10:20:11.946: INFO: (3) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 18.44662ms)
+Feb 12 10:20:11.946: INFO: (3) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 18.330948ms)
+Feb 12 10:20:11.946: INFO: (3) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 18.428326ms)
+Feb 12 10:20:11.959: INFO: (4) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 12.863407ms)
+Feb 12 10:20:11.959: INFO: (4) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 13.021668ms)
+Feb 12 10:20:11.959: INFO: (4) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:1080/proxy/: test<... (200; 13.250534ms)
+Feb 12 10:20:11.960: INFO: (4) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 13.063372ms)
+Feb 12 10:20:11.960: INFO: (4) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:460/proxy/: tls baz (200; 13.422298ms)
+Feb 12 10:20:11.960: INFO: (4) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 13.596231ms)
+Feb 12 10:20:11.962: INFO: (4) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test (200; 17.965478ms)
+Feb 12 10:20:11.965: INFO: (4) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname2/proxy/: bar (200; 18.338517ms)
+Feb 12 10:20:11.965: INFO: (4) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname2/proxy/: tls qux (200; 18.121194ms)
+Feb 12 10:20:11.965: INFO: (4) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 18.23645ms)
+Feb 12 10:20:11.965: INFO: (4) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 18.383259ms)
+Feb 12 10:20:11.965: INFO: (4) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 18.359496ms)
+Feb 12 10:20:11.967: INFO: (4) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 19.958169ms)
+Feb 12 10:20:11.977: INFO: (5) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 10.265065ms)
+Feb 12 10:20:11.979: INFO: (5) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 12.464904ms)
+Feb 12 10:20:11.980: INFO: (5) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 12.890158ms)
+Feb 12 10:20:11.980: INFO: (5) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:460/proxy/: tls baz (200; 13.082835ms)
+Feb 12 10:20:11.980: INFO: (5) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 13.209894ms)
+Feb 12 10:20:11.980: INFO: (5) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test<... (200; 18.831494ms)
+Feb 12 10:20:11.986: INFO: (5) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 19.090043ms)
+Feb 12 10:20:11.988: INFO: (5) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname2/proxy/: bar (200; 21.330865ms)
+Feb 12 10:20:11.988: INFO: (5) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 21.632318ms)
+Feb 12 10:20:11.993: INFO: (5) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 25.937573ms)
+Feb 12 10:20:12.000: INFO: (5) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname2/proxy/: tls qux (200; 33.533425ms)
+Feb 12 10:20:12.001: INFO: (5) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 34.566107ms)
+Feb 12 10:20:12.014: INFO: (6) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 11.878693ms)
+Feb 12 10:20:12.016: INFO: (6) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 13.882129ms)
+Feb 12 10:20:12.016: INFO: (6) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 14.157085ms)
+Feb 12 10:20:12.016: INFO: (6) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 13.895701ms)
+Feb 12 10:20:12.016: INFO: (6) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test<... (200; 14.627086ms)
+Feb 12 10:20:12.017: INFO: (6) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 14.959252ms)
+Feb 12 10:20:12.017: INFO: (6) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 14.661434ms)
+Feb 12 10:20:12.017: INFO: (6) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 15.160412ms)
+Feb 12 10:20:12.021: INFO: (6) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 19.087913ms)
+Feb 12 10:20:12.022: INFO: (6) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname2/proxy/: tls qux (200; 19.871585ms)
+Feb 12 10:20:12.022: INFO: (6) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 20.173118ms)
+Feb 12 10:20:12.022: INFO: (6) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 20.04828ms)
+Feb 12 10:20:12.022: INFO: (6) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname2/proxy/: bar (200; 20.258707ms)
+Feb 12 10:20:12.034: INFO: (7) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 11.432932ms)
+Feb 12 10:20:12.034: INFO: (7) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 11.807926ms)
+Feb 12 10:20:12.034: INFO: (7) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 11.710565ms)
+Feb 12 10:20:12.035: INFO: (7) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:1080/proxy/: test<... (200; 12.083922ms)
+Feb 12 10:20:12.035: INFO: (7) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 12.181676ms)
+Feb 12 10:20:12.035: INFO: (7) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname2/proxy/: tls qux (200; 12.617366ms)
+Feb 12 10:20:12.035: INFO: (7) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 12.464319ms)
+Feb 12 10:20:12.035: INFO: (7) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:460/proxy/: tls baz (200; 12.719033ms)
+Feb 12 10:20:12.036: INFO: (7) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 12.930708ms)
+Feb 12 10:20:12.036: INFO: (7) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 13.033376ms)
+Feb 12 10:20:12.036: INFO: (7) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test<... (200; 15.767224ms)
+Feb 12 10:20:12.095: INFO: (8) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 15.60155ms)
+Feb 12 10:20:12.095: INFO: (8) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 15.883878ms)
+Feb 12 10:20:12.095: INFO: (8) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test (200; 16.548348ms)
+Feb 12 10:20:12.096: INFO: (8) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 16.218163ms)
+Feb 12 10:20:12.096: INFO: (8) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname1/proxy/: tls baz (200; 16.500971ms)
+Feb 12 10:20:12.096: INFO: (8) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 17.219804ms)
+Feb 12 10:20:12.099: INFO: (8) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 19.607891ms)
+Feb 12 10:20:12.101: INFO: (8) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname2/proxy/: bar (200; 22.309675ms)
+Feb 12 10:20:12.103: INFO: (8) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 24.47402ms)
+Feb 12 10:20:12.103: INFO: (8) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 24.245054ms)
+Feb 12 10:20:12.119: INFO: (9) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 14.893232ms)
+Feb 12 10:20:12.119: INFO: (9) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 15.51462ms)
+Feb 12 10:20:12.119: INFO: (9) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test<... (200; 15.623453ms)
+Feb 12 10:20:12.120: INFO: (9) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 16.376922ms)
+Feb 12 10:20:12.120: INFO: (9) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 16.22948ms)
+Feb 12 10:20:12.162: INFO: (9) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname1/proxy/: tls baz (200; 58.433474ms)
+Feb 12 10:20:12.162: INFO: (9) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname2/proxy/: tls qux (200; 57.953874ms)
+Feb 12 10:20:12.164: INFO: (9) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 60.129881ms)
+Feb 12 10:20:12.164: INFO: (9) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname2/proxy/: bar (200; 60.492482ms)
+Feb 12 10:20:12.165: INFO: (9) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 60.617899ms)
+Feb 12 10:20:12.165: INFO: (9) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 60.291107ms)
+Feb 12 10:20:12.179: INFO: (10) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 13.983286ms)
+Feb 12 10:20:12.179: INFO: (10) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 14.041276ms)
+Feb 12 10:20:12.179: INFO: (10) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:460/proxy/: tls baz (200; 14.129984ms)
+Feb 12 10:20:12.179: INFO: (10) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 14.114753ms)
+Feb 12 10:20:12.179: INFO: (10) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:1080/proxy/: test<... (200; 13.977723ms)
+Feb 12 10:20:12.179: INFO: (10) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 14.360818ms)
+Feb 12 10:20:12.180: INFO: (10) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 14.395846ms)
+Feb 12 10:20:12.180: INFO: (10) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 14.505056ms)
+Feb 12 10:20:12.180: INFO: (10) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 14.674592ms)
+Feb 12 10:20:12.180: INFO: (10) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test (200; 17.334648ms)
+Feb 12 10:20:12.210: INFO: (11) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname2/proxy/: tls qux (200; 17.448324ms)
+Feb 12 10:20:12.210: INFO: (11) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:1080/proxy/: test<... (200; 17.495916ms)
+Feb 12 10:20:12.210: INFO: (11) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 17.735437ms)
+Feb 12 10:20:12.210: INFO: (11) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 17.619922ms)
+Feb 12 10:20:12.210: INFO: (11) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test<... (200; 15.993585ms)
+Feb 12 10:20:12.232: INFO: (12) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 16.26205ms)
+Feb 12 10:20:12.232: INFO: (12) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 16.779253ms)
+Feb 12 10:20:12.232: INFO: (12) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 16.090522ms)
+Feb 12 10:20:12.233: INFO: (12) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: ... (200; 59.874345ms)
+Feb 12 10:20:12.300: INFO: (13) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 60.01489ms)
+Feb 12 10:20:12.300: INFO: (13) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test (200; 59.779268ms)
+Feb 12 10:20:12.300: INFO: (13) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname1/proxy/: tls baz (200; 59.917323ms)
+Feb 12 10:20:12.300: INFO: (13) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:1080/proxy/: test<... (200; 59.847811ms)
+Feb 12 10:20:12.300: INFO: (13) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 59.998564ms)
+Feb 12 10:20:12.300: INFO: (13) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 60.214731ms)
+Feb 12 10:20:12.300: INFO: (13) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:460/proxy/: tls baz (200; 60.150953ms)
+Feb 12 10:20:12.300: INFO: (13) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 59.935204ms)
+Feb 12 10:20:12.341: INFO: (13) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname2/proxy/: bar (200; 101.43769ms)
+Feb 12 10:20:12.341: INFO: (13) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 101.384569ms)
+Feb 12 10:20:12.341: INFO: (13) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 101.534516ms)
+Feb 12 10:20:12.341: INFO: (13) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 101.572045ms)
+Feb 12 10:20:12.353: INFO: (14) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 11.819036ms)
+Feb 12 10:20:12.354: INFO: (14) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 11.815932ms)
+Feb 12 10:20:12.354: INFO: (14) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 11.613498ms)
+Feb 12 10:20:12.354: INFO: (14) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 11.806952ms)
+Feb 12 10:20:12.354: INFO: (14) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 12.095003ms)
+Feb 12 10:20:12.354: INFO: (14) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 11.658244ms)
+Feb 12 10:20:12.355: INFO: (14) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 13.271267ms)
+Feb 12 10:20:12.357: INFO: (14) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:1080/proxy/: test<... (200; 15.17639ms)
+Feb 12 10:20:12.358: INFO: (14) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 15.690788ms)
+Feb 12 10:20:12.358: INFO: (14) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:460/proxy/: tls baz (200; 15.779844ms)
+Feb 12 10:20:12.358: INFO: (14) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test<... (200; 13.012735ms)
+Feb 12 10:20:12.373: INFO: (15) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 12.869649ms)
+Feb 12 10:20:12.373: INFO: (15) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 12.954545ms)
+Feb 12 10:20:12.373: INFO: (15) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 12.929659ms)
+Feb 12 10:20:12.374: INFO: (15) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 13.551342ms)
+Feb 12 10:20:12.375: INFO: (15) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname2/proxy/: tls qux (200; 15.018652ms)
+Feb 12 10:20:12.375: INFO: (15) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:460/proxy/: tls baz (200; 15.285398ms)
+Feb 12 10:20:12.375: INFO: (15) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 14.999985ms)
+Feb 12 10:20:12.377: INFO: (15) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname1/proxy/: tls baz (200; 16.654253ms)
+Feb 12 10:20:12.377: INFO: (15) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 16.9991ms)
+Feb 12 10:20:12.378: INFO: (15) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 18.281803ms)
+Feb 12 10:20:12.380: INFO: (15) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 19.758724ms)
+Feb 12 10:20:12.381: INFO: (15) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname2/proxy/: bar (200; 21.17823ms)
+Feb 12 10:20:12.395: INFO: (16) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 13.253007ms)
+Feb 12 10:20:12.396: INFO: (16) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test (200; 13.866758ms)
+Feb 12 10:20:12.396: INFO: (16) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 14.456737ms)
+Feb 12 10:20:12.396: INFO: (16) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 14.164175ms)
+Feb 12 10:20:12.396: INFO: (16) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:460/proxy/: tls baz (200; 14.55286ms)
+Feb 12 10:20:12.397: INFO: (16) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:1080/proxy/: test<... (200; 15.184598ms)
+Feb 12 10:20:12.397: INFO: (16) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname2/proxy/: tls qux (200; 14.368309ms)
+Feb 12 10:20:12.397: INFO: (16) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname1/proxy/: tls baz (200; 15.050388ms)
+Feb 12 10:20:12.399: INFO: (16) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 16.984288ms)
+Feb 12 10:20:12.401: INFO: (16) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 19.134275ms)
+Feb 12 10:20:12.401: INFO: (16) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 19.390398ms)
+Feb 12 10:20:12.401: INFO: (16) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 19.836516ms)
+Feb 12 10:20:12.416: INFO: (17) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 13.823992ms)
+Feb 12 10:20:12.416: INFO: (17) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 14.305604ms)
+Feb 12 10:20:12.416: INFO: (17) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 14.265806ms)
+Feb 12 10:20:12.416: INFO: (17) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:1080/proxy/: test<... (200; 14.037734ms)
+Feb 12 10:20:12.416: INFO: (17) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 14.271889ms)
+Feb 12 10:20:12.416: INFO: (17) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname2/proxy/: bar (200; 13.923765ms)
+Feb 12 10:20:12.416: INFO: (17) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 14.117238ms)
+Feb 12 10:20:12.416: INFO: (17) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname2/proxy/: tls qux (200; 14.109258ms)
+Feb 12 10:20:12.416: INFO: (17) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 14.335346ms)
+Feb 12 10:20:12.419: INFO: (17) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test<... (200; 13.339055ms)
+Feb 12 10:20:12.434: INFO: (18) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 13.496037ms)
+Feb 12 10:20:12.434: INFO: (18) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 13.899642ms)
+Feb 12 10:20:12.435: INFO: (18) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:1080/proxy/: ... (200; 13.953297ms)
+Feb 12 10:20:12.435: INFO: (18) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname1/proxy/: tls baz (200; 13.961239ms)
+Feb 12 10:20:12.435: INFO: (18) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:460/proxy/: tls baz (200; 14.041632ms)
+Feb 12 10:20:12.435: INFO: (18) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 13.882829ms)
+Feb 12 10:20:12.435: INFO: (18) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 13.925965ms)
+Feb 12 10:20:12.435: INFO: (18) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: ... (200; 11.702375ms)
+Feb 12 10:20:12.454: INFO: (19) /api/v1/namespaces/proxy-1718/pods/http:proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 11.763898ms)
+Feb 12 10:20:12.454: INFO: (19) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq/proxy/: test (200; 11.586672ms)
+Feb 12 10:20:12.455: INFO: (19) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:460/proxy/: tls baz (200; 11.946387ms)
+Feb 12 10:20:12.455: INFO: (19) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname1/proxy/: foo (200; 11.801204ms)
+Feb 12 10:20:12.456: INFO: (19) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:462/proxy/: tls qux (200; 13.622602ms)
+Feb 12 10:20:12.496: INFO: (19) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname2/proxy/: tls qux (200; 53.05271ms)
+Feb 12 10:20:12.496: INFO: (19) /api/v1/namespaces/proxy-1718/services/https:proxy-service-dcbp5:tlsportname1/proxy/: tls baz (200; 52.921163ms)
+Feb 12 10:20:12.496: INFO: (19) /api/v1/namespaces/proxy-1718/pods/https:proxy-service-dcbp5-c26kq:443/proxy/: test<... (200; 53.059953ms)
+Feb 12 10:20:12.496: INFO: (19) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:162/proxy/: bar (200; 53.320995ms)
+Feb 12 10:20:12.498: INFO: (19) /api/v1/namespaces/proxy-1718/pods/proxy-service-dcbp5-c26kq:160/proxy/: foo (200; 55.392549ms)
+Feb 12 10:20:12.498: INFO: (19) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname1/proxy/: foo (200; 55.470692ms)
+Feb 12 10:20:12.498: INFO: (19) /api/v1/namespaces/proxy-1718/services/http:proxy-service-dcbp5:portname2/proxy/: bar (200; 55.015643ms)
+Feb 12 10:20:12.503: INFO: (19) /api/v1/namespaces/proxy-1718/services/proxy-service-dcbp5:portname2/proxy/: bar (200; 60.803708ms)
+STEP: deleting ReplicationController proxy-service-dcbp5 in namespace proxy-1718, will wait for the garbage collector to delete the pods
+Feb 12 10:20:12.574: INFO: Deleting ReplicationController proxy-service-dcbp5 took: 14.387447ms
+Feb 12 10:20:12.674: INFO: Terminating ReplicationController proxy-service-dcbp5 pods took: 100.310339ms
+[AfterEach] version v1
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:20:22.275: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "proxy-1718" for this suite.
+
+• [SLOW TEST:19.830 seconds]
+[sig-network] Proxy
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  version v1
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:57
+    should proxy through a service and a pod  [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-cli] Kubectl client Kubectl expose 
-  should create services for rc  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-network] Proxy version v1 should proxy through a service and a pod  [Conformance]","total":280,"completed":93,"skipped":1189,"failed":0}
+SSS
+------------------------------
+[k8s.io] Container Runtime blackbox test when starting a container that exits 
+  should run with the expected status [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Container Runtime
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:02:03.223: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-6864
+Feb 12 10:20:22.295: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-runtime
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-462
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[It] should create services for rc  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating Agnhost RC
-Aug 30 17:02:03.384: INFO: namespace kubectl-6864
-Aug 30 17:02:03.384: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-6864'
-Aug 30 17:02:03.643: INFO: stderr: ""
-Aug 30 17:02:03.643: INFO: stdout: "replicationcontroller/agnhost-master created\n"
-STEP: Waiting for Agnhost master to start.
-Aug 30 17:02:04.649: INFO: Selector matched 1 pods for map[app:agnhost]
-Aug 30 17:02:04.649: INFO: Found 0 / 1
-Aug 30 17:02:05.649: INFO: Selector matched 1 pods for map[app:agnhost]
-Aug 30 17:02:05.649: INFO: Found 1 / 1
-Aug 30 17:02:05.649: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
-Aug 30 17:02:05.653: INFO: Selector matched 1 pods for map[app:agnhost]
-Aug 30 17:02:05.653: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
-Aug 30 17:02:05.653: INFO: wait on agnhost-master startup in kubectl-6864 
-Aug 30 17:02:05.654: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 logs agnhost-master-pfd9s agnhost-master --namespace=kubectl-6864'
-Aug 30 17:02:05.877: INFO: stderr: ""
-Aug 30 17:02:05.877: INFO: stdout: "Paused\n"
-STEP: exposing RC
-Aug 30 17:02:05.877: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 expose rc agnhost-master --name=rm2 --port=1234 --target-port=6379 --namespace=kubectl-6864'
-Aug 30 17:02:05.978: INFO: stderr: ""
-Aug 30 17:02:05.978: INFO: stdout: "service/rm2 exposed\n"
-Aug 30 17:02:05.985: INFO: Service rm2 in namespace kubectl-6864 found.
-STEP: exposing service
-Aug 30 17:02:07.994: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=kubectl-6864'
-Aug 30 17:02:08.078: INFO: stderr: ""
-Aug 30 17:02:08.078: INFO: stdout: "service/rm3 exposed\n"
-Aug 30 17:02:08.082: INFO: Service rm3 in namespace kubectl-6864 found.
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:10.092: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-6864" for this suite.
+[It] should run with the expected status [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount'
+STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase'
+STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition
+STEP: Container 'terminate-cmd-rpa': should get the expected 'State'
+STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance]
+STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount'
+STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase'
+STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition
+STEP: Container 'terminate-cmd-rpof': should get the expected 'State'
+STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance]
+STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount'
+STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase'
+STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition
+STEP: Container 'terminate-cmd-rpn': should get the expected 'State'
+STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance]
+[AfterEach] [k8s.io] Container Runtime
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:20:45.797: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-runtime-462" for this suite.
 
-• [SLOW TEST:6.881 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Kubectl expose
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1188
-    should create services for rc  [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:23.519 seconds]
+[k8s.io] Container Runtime
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  blackbox test
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38
+    when starting a container that exits
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:39
+      should run with the expected status [NodeConformance] [Conformance]
+      /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Kubectl expose should create services for rc  [Conformance]","total":280,"completed":118,"skipped":2014,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [k8s.io] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance]","total":280,"completed":94,"skipped":1192,"failed":0}
+SSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Watchers 
-  should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] Daemon set [Serial] 
+  should retry creating failed daemon pods [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:02:10.105: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename watch
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-3684
+Feb 12 10:20:45.818: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename daemonsets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-5510
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a watch on configmaps with a certain label
-STEP: creating a new configmap
-STEP: modifying the configmap once
-STEP: changing the label value of the configmap
-STEP: Expecting to observe a delete notification for the watched object
-Aug 30 17:02:10.288: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-3684 /api/v1/namespaces/watch-3684/configmaps/e2e-watch-test-label-changed 458597cf-d454-454b-98ec-41e7a65afcb6 14268 0 2020-08-30 17:02:10 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Aug 30 17:02:10.289: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-3684 /api/v1/namespaces/watch-3684/configmaps/e2e-watch-test-label-changed 458597cf-d454-454b-98ec-41e7a65afcb6 14269 0 2020-08-30 17:02:10 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-Aug 30 17:02:10.290: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-3684 /api/v1/namespaces/watch-3684/configmaps/e2e-watch-test-label-changed 458597cf-d454-454b-98ec-41e7a65afcb6 14270 0 2020-08-30 17:02:10 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-STEP: modifying the configmap a second time
-STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements
-STEP: changing the label value of the configmap back
-STEP: modifying the configmap a third time
-STEP: deleting the configmap
-STEP: Expecting to observe an add notification for the watched object when the label value was restored
-Aug 30 17:02:20.336: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-3684 /api/v1/namespaces/watch-3684/configmaps/e2e-watch-test-label-changed 458597cf-d454-454b-98ec-41e7a65afcb6 14330 0 2020-08-30 17:02:10 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Aug 30 17:02:20.337: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-3684 /api/v1/namespaces/watch-3684/configmaps/e2e-watch-test-label-changed 458597cf-d454-454b-98ec-41e7a65afcb6 14331 0 2020-08-30 17:02:10 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
-Aug 30 17:02:20.337: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed  watch-3684 /api/v1/namespaces/watch-3684/configmaps/e2e-watch-test-label-changed 458597cf-d454-454b-98ec-41e7a65afcb6 14332 0 2020-08-30 17:02:10 +0000 UTC   map[watch-this-configmap:label-changed-and-restored] map[] [] []  []},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},}
-[AfterEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:20.337: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "watch-3684" for this suite.
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
+[It] should retry creating failed daemon pods [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a simple DaemonSet "daemon-set"
+STEP: Check that daemon pods launch on every node of the cluster.
+Feb 12 10:20:46.047: INFO: Number of nodes with available pods: 0
+Feb 12 10:20:46.048: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod
+Feb 12 10:20:47.065: INFO: Number of nodes with available pods: 0
+Feb 12 10:20:47.065: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod
+Feb 12 10:20:48.063: INFO: Number of nodes with available pods: 1
+Feb 12 10:20:48.063: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod
+Feb 12 10:20:49.061: INFO: Number of nodes with available pods: 3
+Feb 12 10:20:49.061: INFO: Number of running nodes: 3, number of available pods: 3
+STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived.
+Feb 12 10:20:49.099: INFO: Number of nodes with available pods: 3
+Feb 12 10:20:49.099: INFO: Number of running nodes: 3, number of available pods: 3
+STEP: Wait for the failed daemon pod to be completely deleted.
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-5510, will wait for the garbage collector to delete the pods
+Feb 12 10:20:49.181: INFO: Deleting DaemonSet.extensions daemon-set took: 12.911289ms
+Feb 12 10:20:49.681: INFO: Terminating DaemonSet.extensions daemon-set pods took: 500.283243ms
+Feb 12 10:21:02.387: INFO: Number of nodes with available pods: 0
+Feb 12 10:21:02.387: INFO: Number of running nodes: 0, number of available pods: 0
+Feb 12 10:21:02.393: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-5510/daemonsets","resourceVersion":"18104"},"items":null}
 
-• [SLOW TEST:10.248 seconds]
-[sig-api-machinery] Watchers
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should observe an object deletion if it stops meeting the requirements of the selector [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance]","total":280,"completed":119,"skipped":2054,"failed":0}
-SSSSSSS
+Feb 12 10:21:02.398: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-5510/pods","resourceVersion":"18104"},"items":null}
+
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:21:02.434: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "daemonsets-5510" for this suite.
+
+• [SLOW TEST:16.633 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should retry creating failed daemon pods [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] Lease 
-  lease API should be available [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Lease
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:02:20.353: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename lease-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in lease-test-1261
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] lease API should be available [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [k8s.io] Lease
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:20.593: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "lease-test-1261" for this suite.
-•{"msg":"PASSED [k8s.io] Lease lease API should be available [Conformance]","total":280,"completed":120,"skipped":2061,"failed":0}
-SSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance]","total":280,"completed":95,"skipped":1205,"failed":0}
+SSSSSSSSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Downward API volume 
+  should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:02:20.607: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-8704
+Feb 12 10:21:02.451: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-8532
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name configmap-test-volume-10024e4b-0673-46d4-a59c-168f9c3eb2bb
-STEP: Creating a pod to test consume configMaps
-Aug 30 17:02:20.820: INFO: Waiting up to 5m0s for pod "pod-configmaps-1f921bde-4a8d-436c-bc19-e69be507ca35" in namespace "configmap-8704" to be "success or failure"
-Aug 30 17:02:20.828: INFO: Pod "pod-configmaps-1f921bde-4a8d-436c-bc19-e69be507ca35": Phase="Pending", Reason="", readiness=false. Elapsed: 8.262462ms
-Aug 30 17:02:22.834: INFO: Pod "pod-configmaps-1f921bde-4a8d-436c-bc19-e69be507ca35": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014015603s
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should provide container's cpu limit [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Feb 12 10:21:02.636: INFO: Waiting up to 5m0s for pod "downwardapi-volume-73efe267-f9db-46b3-9453-70fa2f7b0eb1" in namespace "downward-api-8532" to be "success or failure"
+Feb 12 10:21:02.642: INFO: Pod "downwardapi-volume-73efe267-f9db-46b3-9453-70fa2f7b0eb1": Phase="Pending", Reason="", readiness=false. Elapsed: 5.038898ms
+Feb 12 10:21:04.679: INFO: Pod "downwardapi-volume-73efe267-f9db-46b3-9453-70fa2f7b0eb1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.042368651s
 STEP: Saw pod success
-Aug 30 17:02:22.834: INFO: Pod "pod-configmaps-1f921bde-4a8d-436c-bc19-e69be507ca35" satisfied condition "success or failure"
-Aug 30 17:02:22.840: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-configmaps-1f921bde-4a8d-436c-bc19-e69be507ca35 container configmap-volume-test: 
+Feb 12 10:21:04.679: INFO: Pod "downwardapi-volume-73efe267-f9db-46b3-9453-70fa2f7b0eb1" satisfied condition "success or failure"
+Feb 12 10:21:04.686: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-73efe267-f9db-46b3-9453-70fa2f7b0eb1 container client-container: 
 STEP: delete the pod
-Aug 30 17:02:22.916: INFO: Waiting for pod pod-configmaps-1f921bde-4a8d-436c-bc19-e69be507ca35 to disappear
-Aug 30 17:02:22.920: INFO: Pod pod-configmaps-1f921bde-4a8d-436c-bc19-e69be507ca35 no longer exists
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:22.920: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-8704" for this suite.
-•{"msg":"PASSED [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":280,"completed":121,"skipped":2078,"failed":0}
-SSSSSSSSSSSSSSSSSS
+Feb 12 10:21:04.779: INFO: Waiting for pod downwardapi-volume-73efe267-f9db-46b3-9453-70fa2f7b0eb1 to disappear
+Feb 12 10:21:04.787: INFO: Pod downwardapi-volume-73efe267-f9db-46b3-9453-70fa2f7b0eb1 no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:21:04.788: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-8532" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance]","total":280,"completed":96,"skipped":1216,"failed":0}
+SSS
 ------------------------------
 [k8s.io] Container Runtime blackbox test on terminated container 
-  should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:02:22.936: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:21:04.856: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename container-runtime
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-1440
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-7582
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[It] should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: create the container
-STEP: wait for the container to reach Succeeded
+STEP: wait for the container to reach Failed
 STEP: get the container status
 STEP: the container should be terminated
 STEP: the termination message should be set
-Aug 30 17:02:25.127: INFO: Expected: &{} to match Container's Termination Message:  --
+Feb 12 10:21:07.142: INFO: Expected: &{DONE} to match Container's Termination Message: DONE --
 STEP: delete the container
 [AfterEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:25.147: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-runtime-1440" for this suite.
-•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":280,"completed":122,"skipped":2096,"failed":0}
-SSSSSSSSSSSS
-------------------------------
-[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition 
-  listing custom resource definition objects works  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:02:25.159: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename custom-resource-definition
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-2270
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] listing custom resource definition objects works  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:02:25.313: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:31.775: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "custom-resource-definition-2270" for this suite.
-
-• [SLOW TEST:6.640 seconds]
-[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  Simple CustomResourceDefinition
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/custom_resource_definition.go:47
-    listing custom resource definition objects works  [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works  [Conformance]","total":280,"completed":123,"skipped":2108,"failed":0}
-SSSSSSS
-------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide podname only [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:02:31.799: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6793
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should provide podname only [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 17:02:32.019: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8e12a98a-2731-4540-8ae6-a95603d3080d" in namespace "projected-6793" to be "success or failure"
-Aug 30 17:02:32.039: INFO: Pod "downwardapi-volume-8e12a98a-2731-4540-8ae6-a95603d3080d": Phase="Pending", Reason="", readiness=false. Elapsed: 20.358295ms
-Aug 30 17:02:34.044: INFO: Pod "downwardapi-volume-8e12a98a-2731-4540-8ae6-a95603d3080d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025392523s
-Aug 30 17:02:36.052: INFO: Pod "downwardapi-volume-8e12a98a-2731-4540-8ae6-a95603d3080d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.033588598s
-STEP: Saw pod success
-Aug 30 17:02:36.053: INFO: Pod "downwardapi-volume-8e12a98a-2731-4540-8ae6-a95603d3080d" satisfied condition "success or failure"
-Aug 30 17:02:36.057: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-8e12a98a-2731-4540-8ae6-a95603d3080d container client-container: 
-STEP: delete the pod
-Aug 30 17:02:36.131: INFO: Waiting for pod downwardapi-volume-8e12a98a-2731-4540-8ae6-a95603d3080d to disappear
-Aug 30 17:02:36.135: INFO: Pod downwardapi-volume-8e12a98a-2731-4540-8ae6-a95603d3080d no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:36.135: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-6793" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance]","total":280,"completed":124,"skipped":2115,"failed":0}
-SSSSSSSSSSSSSSS
-------------------------------
-[sig-node] ConfigMap 
-  should be consumable via the environment [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:02:36.152: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-7069
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable via the environment [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap configmap-7069/configmap-test-84365d81-32fb-4dfd-beb6-b4901b98ef50
-STEP: Creating a pod to test consume configMaps
-Aug 30 17:02:36.326: INFO: Waiting up to 5m0s for pod "pod-configmaps-2850384b-6feb-4d77-b16e-8d9349084fcb" in namespace "configmap-7069" to be "success or failure"
-Aug 30 17:02:36.333: INFO: Pod "pod-configmaps-2850384b-6feb-4d77-b16e-8d9349084fcb": Phase="Pending", Reason="", readiness=false. Elapsed: 7.26094ms
-Aug 30 17:02:38.339: INFO: Pod "pod-configmaps-2850384b-6feb-4d77-b16e-8d9349084fcb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012751914s
-STEP: Saw pod success
-Aug 30 17:02:38.339: INFO: Pod "pod-configmaps-2850384b-6feb-4d77-b16e-8d9349084fcb" satisfied condition "success or failure"
-Aug 30 17:02:38.343: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-configmaps-2850384b-6feb-4d77-b16e-8d9349084fcb container env-test: 
-STEP: delete the pod
-Aug 30 17:02:38.372: INFO: Waiting for pod pod-configmaps-2850384b-6feb-4d77-b16e-8d9349084fcb to disappear
-Aug 30 17:02:38.376: INFO: Pod pod-configmaps-2850384b-6feb-4d77-b16e-8d9349084fcb no longer exists
-[AfterEach] [sig-node] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:38.376: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-7069" for this suite.
-•{"msg":"PASSED [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance]","total":280,"completed":125,"skipped":2130,"failed":0}
-SSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-storage] Projected configMap 
-  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:02:38.389: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-2403
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name projected-configmap-test-volume-0ac8bb3a-09b8-43a4-8e13-18b20290c084
-STEP: Creating a pod to test consume configMaps
-Aug 30 17:02:38.574: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-e34b1de0-5978-4ed2-8961-37cccfe4ec83" in namespace "projected-2403" to be "success or failure"
-Aug 30 17:02:38.582: INFO: Pod "pod-projected-configmaps-e34b1de0-5978-4ed2-8961-37cccfe4ec83": Phase="Pending", Reason="", readiness=false. Elapsed: 8.457295ms
-Aug 30 17:02:40.588: INFO: Pod "pod-projected-configmaps-e34b1de0-5978-4ed2-8961-37cccfe4ec83": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014286519s
-Aug 30 17:02:42.593: INFO: Pod "pod-projected-configmaps-e34b1de0-5978-4ed2-8961-37cccfe4ec83": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.019620186s
-STEP: Saw pod success
-Aug 30 17:02:42.593: INFO: Pod "pod-projected-configmaps-e34b1de0-5978-4ed2-8961-37cccfe4ec83" satisfied condition "success or failure"
-Aug 30 17:02:42.598: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-projected-configmaps-e34b1de0-5978-4ed2-8961-37cccfe4ec83 container projected-configmap-volume-test: 
-STEP: delete the pod
-Aug 30 17:02:42.676: INFO: Waiting for pod pod-projected-configmaps-e34b1de0-5978-4ed2-8961-37cccfe4ec83 to disappear
-Aug 30 17:02:42.681: INFO: Pod pod-projected-configmaps-e34b1de0-5978-4ed2-8961-37cccfe4ec83 no longer exists
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:42.681: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-2403" for this suite.
-•{"msg":"PASSED [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":280,"completed":126,"skipped":2151,"failed":0}
-SSSSSS
-------------------------------
-[k8s.io] Docker Containers 
-  should use the image defaults if command and args are blank [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Docker Containers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:02:42.695: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename containers
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-2522
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should use the image defaults if command and args are blank [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [k8s.io] Docker Containers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:46.896: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "containers-2522" for this suite.
-•{"msg":"PASSED [k8s.io] Docker Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance]","total":280,"completed":127,"skipped":2157,"failed":0}
-SSSSSSS
-------------------------------
-[k8s.io] Pods 
-  should be submitted and removed [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:02:46.919: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-354
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
-[It] should be submitted and removed [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-STEP: setting up watch
-STEP: submitting the pod to kubernetes
-Aug 30 17:02:47.089: INFO: observed the pod list
-STEP: verifying the pod is in kubernetes
-STEP: verifying pod creation was observed
-STEP: deleting the pod gracefully
-STEP: verifying the kubelet observed the termination notice
-Aug 30 17:02:54.152: INFO: no pod exists with the name we were looking for, assuming the termination request was observed and completed
-STEP: verifying pod deletion was observed
-[AfterEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:54.157: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-354" for this suite.
-
-• [SLOW TEST:7.257 seconds]
-[k8s.io] Pods
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should be submitted and removed [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [k8s.io] Pods should be submitted and removed [NodeConformance] [Conformance]","total":280,"completed":128,"skipped":2164,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-api-machinery] Watchers 
-  should be able to start watching from a specific resource version [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:02:54.179: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename watch
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-7321
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to start watching from a specific resource version [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a new configmap
-STEP: modifying the configmap once
-STEP: modifying the configmap a second time
-STEP: deleting the configmap
-STEP: creating a watch on configmaps from the resource version returned by the first update
-STEP: Expecting to observe notifications for all changes to the configmap after the first update
-Aug 30 17:02:54.396: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version  watch-7321 /api/v1/namespaces/watch-7321/configmaps/e2e-watch-test-resource-version c3a994f0-d5a7-40e7-b9ea-4b145e214b1b 14791 0 2020-08-30 17:02:54 +0000 UTC   map[watch-this-configmap:from-resource-version] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Aug 30 17:02:54.397: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version  watch-7321 /api/v1/namespaces/watch-7321/configmaps/e2e-watch-test-resource-version c3a994f0-d5a7-40e7-b9ea-4b145e214b1b 14792 0 2020-08-30 17:02:54 +0000 UTC   map[watch-this-configmap:from-resource-version] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-[AfterEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:02:54.397: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "watch-7321" for this suite.
-•{"msg":"PASSED [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance]","total":280,"completed":129,"skipped":2198,"failed":0}
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:21:07.164: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-runtime-7582" for this suite.
+•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":280,"completed":97,"skipped":1219,"failed":0}
 SSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should mutate configmap [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should mutate custom resource with different stored version [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:02:54.411: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:21:07.186: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-4521
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-195
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
 STEP: Setting up server cert
 STEP: Create role binding to let webhook read extension-apiserver-authentication
 STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Aug 30 17:02:55.081: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Aug 30 17:02:57.097: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403775, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403775, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403775, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403775, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:21:07.597: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:21:09.613: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722067, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722067, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722067, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722067, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Aug 30 17:03:00.117: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should mutate configmap [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering the mutating configmap webhook via the AdmissionRegistration API
-STEP: create a configmap that should be updated by the webhook
+Feb 12 10:21:12.633: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should mutate custom resource with different stored version [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:21:12.639: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Registering the mutating webhook for custom resource e2e-test-webhook-7542-crds.webhook.example.com via the AdmissionRegistration API
+STEP: Creating a custom resource while v1 is storage version
+STEP: Patching Custom Resource Definition to set v2 as storage
+STEP: Patching the custom resource while v2 is storage version
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:03:00.374: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-4521" for this suite.
-STEP: Destroying namespace "webhook-4521-markers" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:21:13.908: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-195" for this suite.
+STEP: Destroying namespace "webhook-195-markers" for this suite.
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:6.051 seconds]
+• [SLOW TEST:6.810 seconds]
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should mutate configmap [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance]","total":280,"completed":130,"skipped":2218,"failed":0}
-SSSSSS
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should mutate custom resource with different stored version [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:03:00.463: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-8947
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0777 on tmpfs
-Aug 30 17:03:00.638: INFO: Waiting up to 5m0s for pod "pod-89aae775-d4b0-4c4a-8457-b2048dcefb07" in namespace "emptydir-8947" to be "success or failure"
-Aug 30 17:03:00.643: INFO: Pod "pod-89aae775-d4b0-4c4a-8457-b2048dcefb07": Phase="Pending", Reason="", readiness=false. Elapsed: 4.998337ms
-Aug 30 17:03:02.649: INFO: Pod "pod-89aae775-d4b0-4c4a-8457-b2048dcefb07": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.01126009s
-STEP: Saw pod success
-Aug 30 17:03:02.649: INFO: Pod "pod-89aae775-d4b0-4c4a-8457-b2048dcefb07" satisfied condition "success or failure"
-Aug 30 17:03:02.654: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-89aae775-d4b0-4c4a-8457-b2048dcefb07 container test-container: 
-STEP: delete the pod
-Aug 30 17:03:02.720: INFO: Waiting for pod pod-89aae775-d4b0-4c4a-8457-b2048dcefb07 to disappear
-Aug 30 17:03:02.724: INFO: Pod pod-89aae775-d4b0-4c4a-8457-b2048dcefb07 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:03:02.724: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-8947" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":131,"skipped":2224,"failed":0}
-SSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance]","total":280,"completed":98,"skipped":1239,"failed":0}
+SSSSS
 ------------------------------
-[sig-api-machinery] Namespaces [Serial] 
-  should ensure that all pods are removed when a namespace is deleted [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Namespaces [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] ConfigMap 
+  binary data should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:03:02.747: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename namespaces
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in namespaces-4938
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should ensure that all pods are removed when a namespace is deleted [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a test namespace
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-9326
+Feb 12 10:21:13.998: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-3998
 STEP: Waiting for a default service account to be provisioned in namespace
-STEP: Creating a pod in the namespace
-STEP: Waiting for the pod to have running status
-STEP: Deleting the namespace
-STEP: Waiting for the namespace to be removed.
-STEP: Recreating the namespace
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-2227
-STEP: Verifying there are no pods in the namespace
-[AfterEach] [sig-api-machinery] Namespaces [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:03:16.291: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "namespaces-4938" for this suite.
-STEP: Destroying namespace "nsdeletetest-9326" for this suite.
-Aug 30 17:03:16.307: INFO: Namespace nsdeletetest-9326 was already deleted
-STEP: Destroying namespace "nsdeletetest-2227" for this suite.
-
-• [SLOW TEST:13.570 seconds]
-[sig-api-machinery] Namespaces [Serial]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should ensure that all pods are removed when a namespace is deleted [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance]","total":280,"completed":132,"skipped":2234,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[It] binary data should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name configmap-test-upd-a01cda3c-8062-4289-ae1e-77ff7c45ea99
+STEP: Creating the pod
+STEP: Waiting for pod with text data
+STEP: Waiting for pod with binary data
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:21:16.289: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-3998" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":99,"skipped":1244,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  works for multiple CRDs of different groups [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  works for CRD without validation schema [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:03:16.318: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:21:16.308: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename crd-publish-openapi
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-7038
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-4197
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] works for multiple CRDs of different groups [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: CRs in different groups (two CRDs) show up in OpenAPI documentation
-Aug 30 17:03:16.473: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 17:03:19.326: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+[It] works for CRD without validation schema [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:21:16.468: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: client-side validation (kubectl create and apply) allows request with any unknown properties
+Feb 12 10:21:18.820: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-4197 create -f -'
+Feb 12 10:21:19.202: INFO: stderr: ""
+Feb 12 10:21:19.202: INFO: stdout: "e2e-test-crd-publish-openapi-9336-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n"
+Feb 12 10:21:19.202: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-4197 delete e2e-test-crd-publish-openapi-9336-crds test-cr'
+Feb 12 10:21:19.346: INFO: stderr: ""
+Feb 12 10:21:19.346: INFO: stdout: "e2e-test-crd-publish-openapi-9336-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n"
+Feb 12 10:21:19.346: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-4197 apply -f -'
+Feb 12 10:21:19.578: INFO: stderr: ""
+Feb 12 10:21:19.578: INFO: stdout: "e2e-test-crd-publish-openapi-9336-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n"
+Feb 12 10:21:19.578: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-4197 delete e2e-test-crd-publish-openapi-9336-crds test-cr'
+Feb 12 10:21:19.673: INFO: stderr: ""
+Feb 12 10:21:19.673: INFO: stdout: "e2e-test-crd-publish-openapi-9336-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n"
+STEP: kubectl explain works to explain CR without validation schema
+Feb 12 10:21:19.673: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 explain e2e-test-crd-publish-openapi-9336-crds'
+Feb 12 10:21:19.820: INFO: stderr: ""
+Feb 12 10:21:19.820: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-9336-crd\nVERSION:  crd-publish-openapi-test-empty.example.com/v1\n\nDESCRIPTION:\n     \n"
 [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:03:31.090: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-7038" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:21:22.735: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-4197" for this suite.
 
-• [SLOW TEST:14.787 seconds]
+• [SLOW TEST:6.444 seconds]
 [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for multiple CRDs of different groups [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance]","total":280,"completed":133,"skipped":2283,"failed":0}
-SSSSSSS
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  works for CRD without validation schema [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-api-machinery] Secrets 
-  should be consumable via the environment [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:03:31.105: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-5267
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable via the environment [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating secret secrets-5267/secret-test-9e1c267f-fb7b-45ca-bebe-6b0a560e851f
-STEP: Creating a pod to test consume secrets
-Aug 30 17:03:31.288: INFO: Waiting up to 5m0s for pod "pod-configmaps-a8b34ba1-4e78-40e0-9a49-61595e258d4b" in namespace "secrets-5267" to be "success or failure"
-Aug 30 17:03:31.295: INFO: Pod "pod-configmaps-a8b34ba1-4e78-40e0-9a49-61595e258d4b": Phase="Pending", Reason="", readiness=false. Elapsed: 7.259044ms
-Aug 30 17:03:33.300: INFO: Pod "pod-configmaps-a8b34ba1-4e78-40e0-9a49-61595e258d4b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012362146s
-Aug 30 17:03:35.307: INFO: Pod "pod-configmaps-a8b34ba1-4e78-40e0-9a49-61595e258d4b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018984346s
-STEP: Saw pod success
-Aug 30 17:03:35.307: INFO: Pod "pod-configmaps-a8b34ba1-4e78-40e0-9a49-61595e258d4b" satisfied condition "success or failure"
-Aug 30 17:03:35.311: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-configmaps-a8b34ba1-4e78-40e0-9a49-61595e258d4b container env-test: 
-STEP: delete the pod
-Aug 30 17:03:35.390: INFO: Waiting for pod pod-configmaps-a8b34ba1-4e78-40e0-9a49-61595e258d4b to disappear
-Aug 30 17:03:35.393: INFO: Pod pod-configmaps-a8b34ba1-4e78-40e0-9a49-61595e258d4b no longer exists
-[AfterEach] [sig-api-machinery] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:03:35.393: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-5267" for this suite.
-•{"msg":"PASSED [sig-api-machinery] Secrets should be consumable via the environment [NodeConformance] [Conformance]","total":280,"completed":134,"skipped":2290,"failed":0}
-SSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance]","total":280,"completed":100,"skipped":1281,"failed":0}
+SSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Aggregator 
-  Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Aggregator
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] ResourceQuota 
+  should verify ResourceQuota with terminating scopes. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:03:35.408: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename aggregator
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in aggregator-4474
+Feb 12 10:21:22.753: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename resourcequota
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-6967
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] Aggregator
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:76
-Aug 30 17:03:35.557: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-[It] Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering the sample API server.
-Aug 30 17:03:36.120: INFO: deployment "sample-apiserver-deployment" doesn't have the required revision set
-Aug 30 17:03:38.220: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Aug 30 17:03:40.226: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Aug 30 17:03:42.227: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Aug 30 17:03:44.226: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734403816, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-867766ffc6\" is progressing."}}, CollisionCount:(*int32)(nil)}
-Aug 30 17:03:47.348: INFO: Waited 1.108543883s for the sample-apiserver to be ready to handle requests.
-[AfterEach] [sig-api-machinery] Aggregator
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/aggregator.go:67
-[AfterEach] [sig-api-machinery] Aggregator
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:03:48.324: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "aggregator-4474" for this suite.
+[It] should verify ResourceQuota with terminating scopes. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a ResourceQuota with terminating scope
+STEP: Ensuring ResourceQuota status is calculated
+STEP: Creating a ResourceQuota with not terminating scope
+STEP: Ensuring ResourceQuota status is calculated
+STEP: Creating a long running pod
+STEP: Ensuring resource quota with not terminating scope captures the pod usage
+STEP: Ensuring resource quota with terminating scope ignored the pod usage
+STEP: Deleting the pod
+STEP: Ensuring resource quota status released the pod usage
+STEP: Creating a terminating pod
+STEP: Ensuring resource quota with terminating scope captures the pod usage
+STEP: Ensuring resource quota with not terminating scope ignored the pod usage
+STEP: Deleting the pod
+STEP: Ensuring resource quota status released the pod usage
+[AfterEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:21:39.100: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-6967" for this suite.
 
-• [SLOW TEST:12.938 seconds]
-[sig-api-machinery] Aggregator
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:16.371 seconds]
+[sig-api-machinery] ResourceQuota
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should verify ResourceQuota with terminating scopes. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Aggregator Should be able to support the 1.10 Sample API Server using the current Aggregator [Conformance]","total":280,"completed":135,"skipped":2301,"failed":0}
-SSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance]","total":280,"completed":101,"skipped":1299,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Garbage collector 
-  should orphan pods created by rc if delete options say so [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] [sig-node] Events 
+  should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] [sig-node] Events
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:03:48.349: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-9227
+Feb 12 10:21:39.125: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename events
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in events-7601
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should orphan pods created by rc if delete options say so [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the rc
-STEP: delete the rc
-STEP: wait for the rc to be deleted
-STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods
-STEP: Gathering metrics
-W0830 17:04:28.567640      23 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-Aug 30 17:04:28.567: INFO: For apiserver_request_total:
-For apiserver_request_latency_seconds:
-For apiserver_init_events_total:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
+[It] should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+STEP: retrieving the pod
+Feb 12 10:21:43.328: INFO: &Pod{ObjectMeta:{send-events-95705413-cb8d-4ed1-ae65-c0ed9b345cb9  events-7601 /api/v1/namespaces/events-7601/pods/send-events-95705413-cb8d-4ed1-ae65-c0ed9b345cb9 5287c3f3-6d8d-4a32-8383-eed8431a5d93 18607 0 2021-02-12 10:21:39 +0000 UTC   map[name:foo time:291869683] map[cni.projectcalico.org/podIP:172.25.1.98/32] [] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-2rf2z,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-2rf2z,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:p,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[serve-hostname],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:,HostPort:0,ContainerPort:80,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-2rf2z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-xg74t,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:21:39 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:21:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:21:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:21:39 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:138.68.97.197,PodIP:172.25.1.98,StartTime:2021-02-12 10:21:39 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:p,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-02-12 10:21:40 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/agnhost@sha256:daf5332100521b1256d0e3c56d697a238eaec3af48897ed9167cbadd426773b5,ContainerID:docker://9158b6951a3ec00fb8ed71b7af6e7ab177a83a1061909d0c986a48f7f7ad18f2,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.98,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
 
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:04:28.567: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-9227" for this suite.
+STEP: checking for scheduler event about the pod
+Feb 12 10:21:45.334: INFO: Saw scheduler event for our pod.
+STEP: checking for kubelet event about the pod
+Feb 12 10:21:47.341: INFO: Saw kubelet event for our pod.
+STEP: deleting the pod
+[AfterEach] [k8s.io] [sig-node] Events
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:21:47.354: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "events-7601" for this suite.
 
-• [SLOW TEST:40.231 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should orphan pods created by rc if delete options say so [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:8.244 seconds]
+[k8s.io] [sig-node] Events
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance]","total":280,"completed":136,"skipped":2317,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [k8s.io] [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]","total":280,"completed":102,"skipped":1375,"failed":0}
+SSSS
 ------------------------------
-[sig-network] Networking Granular Checks: Pods 
-  should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Networking
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] Job 
+  should delete a job [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] Job
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:04:28.582: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename pod-network-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-7268
+Feb 12 10:21:47.371: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename job
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in job-8073
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Performing setup for networking test in namespace pod-network-test-7268
-STEP: creating a selector
-STEP: Creating the service pods in kubernetes
-Aug 30 17:04:28.751: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
-STEP: Creating test pods
-Aug 30 17:04:48.856: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.25.1.53 8081 | grep -v '^\s*$'] Namespace:pod-network-test-7268 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 17:04:48.856: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 17:04:50.335: INFO: Found all expected endpoints: [netserver-0]
-Aug 30 17:04:50.340: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.25.0.125 8081 | grep -v '^\s*$'] Namespace:pod-network-test-7268 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 17:04:50.340: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 17:04:51.757: INFO: Found all expected endpoints: [netserver-1]
-[AfterEach] [sig-network] Networking
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:04:51.757: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pod-network-test-7268" for this suite.
+[It] should delete a job [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a job
+STEP: Ensuring active pods == parallelism
+STEP: delete a job
+STEP: deleting Job.batch foo in namespace job-8073, will wait for the garbage collector to delete the pods
+Feb 12 10:21:51.627: INFO: Deleting Job.batch foo took: 25.659119ms
+Feb 12 10:21:52.127: INFO: Terminating Job.batch foo pods took: 500.324678ms
+STEP: Ensuring job was deleted
+[AfterEach] [sig-apps] Job
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:22:32.445: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "job-8073" for this suite.
 
-• [SLOW TEST:23.188 seconds]
-[sig-network] Networking
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
-  Granular Checks: Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
-    should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:45.091 seconds]
+[sig-apps] Job
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should delete a job [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":137,"skipped":2354,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-apps] Job should delete a job [Conformance]","total":280,"completed":103,"skipped":1379,"failed":0}
+SSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] Job 
+  should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] Job
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:04:51.771: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9563
+Feb 12 10:22:32.463: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename job
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in job-9476
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating projection with secret that has name projected-secret-test-39cd66b7-9392-4a92-9f49-b427de43bcba
-STEP: Creating a pod to test consume secrets
-Aug 30 17:04:51.956: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-99cd0842-b006-4743-84fa-c9284dc44f32" in namespace "projected-9563" to be "success or failure"
-Aug 30 17:04:51.964: INFO: Pod "pod-projected-secrets-99cd0842-b006-4743-84fa-c9284dc44f32": Phase="Pending", Reason="", readiness=false. Elapsed: 7.900014ms
-Aug 30 17:04:53.970: INFO: Pod "pod-projected-secrets-99cd0842-b006-4743-84fa-c9284dc44f32": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.013665981s
-STEP: Saw pod success
-Aug 30 17:04:53.970: INFO: Pod "pod-projected-secrets-99cd0842-b006-4743-84fa-c9284dc44f32" satisfied condition "success or failure"
-Aug 30 17:04:53.975: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod pod-projected-secrets-99cd0842-b006-4743-84fa-c9284dc44f32 container projected-secret-volume-test: 
-STEP: delete the pod
-Aug 30 17:04:54.042: INFO: Waiting for pod pod-projected-secrets-99cd0842-b006-4743-84fa-c9284dc44f32 to disappear
-Aug 30 17:04:54.046: INFO: Pod pod-projected-secrets-99cd0842-b006-4743-84fa-c9284dc44f32 no longer exists
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:04:54.046: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-9563" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":138,"skipped":2388,"failed":0}
-SSSSSSSSS
+[It] should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a job
+STEP: Ensuring job reaches completions
+[AfterEach] [sig-apps] Job
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:22:40.635: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "job-9476" for this suite.
+
+• [SLOW TEST:8.193 seconds]
+[sig-apps] Job
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]","total":280,"completed":104,"skipped":1399,"failed":0}
+SSSSS
+------------------------------
+[sig-network] DNS 
+  should provide DNS for pods for Subdomain [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:04:54.066: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename statefulset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-8383
+Feb 12 10:22:40.658: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename dns
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-1017
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
-STEP: Creating service test in namespace statefulset-8383
-[It] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating stateful set ss in namespace statefulset-8383
-STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-8383
-Aug 30 17:04:54.244: INFO: Found 0 stateful pods, waiting for 1
-Aug 30 17:05:04.254: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod
-Aug 30 17:05:05.057: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Aug 30 17:05:05.751: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Aug 30 17:05:05.751: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Aug 30 17:05:05.751: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Aug 30 17:05:05.756: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
-Aug 30 17:05:15.763: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
-Aug 30 17:05:15.763: INFO: Waiting for statefulset status.replicas updated to 0
-Aug 30 17:05:15.788: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
-Aug 30 17:05:15.788: INFO: ss-0  adoring-wozniak-54dcfd79fc-6rshr  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:04:54 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:06 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:06 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:04:54 +0000 UTC  }]
-Aug 30 17:05:15.788: INFO: 
-Aug 30 17:05:15.788: INFO: StatefulSet ss has not reached scale 3, at 1
-Aug 30 17:05:16.795: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.994154815s
-Aug 30 17:05:17.801: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.986909983s
-Aug 30 17:05:18.807: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.981038002s
-Aug 30 17:05:19.813: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.974754748s
-Aug 30 17:05:20.820: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.96874811s
-Aug 30 17:05:21.826: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.962073015s
-Aug 30 17:05:22.833: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.956400676s
-Aug 30 17:05:23.839: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.949317203s
-Aug 30 17:05:24.845: INFO: Verifying statefulset ss doesn't scale past 3 for another 943.481667ms
-STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-8383
-Aug 30 17:05:25.852: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:05:26.404: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Aug 30 17:05:26.405: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Aug 30 17:05:26.405: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Aug 30 17:05:26.405: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:05:26.979: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n"
-Aug 30 17:05:26.979: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Aug 30 17:05:26.979: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Aug 30 17:05:26.979: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:05:27.553: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n"
-Aug 30 17:05:27.553: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Aug 30 17:05:27.553: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Aug 30 17:05:27.560: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-Aug 30 17:05:27.560: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
-Aug 30 17:05:27.560: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Scale down will not halt with unhealthy stateful pod
-Aug 30 17:05:27.566: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Aug 30 17:05:28.098: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Aug 30 17:05:28.098: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Aug 30 17:05:28.098: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Aug 30 17:05:28.098: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Aug 30 17:05:28.682: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Aug 30 17:05:28.682: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Aug 30 17:05:28.682: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Aug 30 17:05:28.682: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Aug 30 17:05:29.238: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Aug 30 17:05:29.238: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Aug 30 17:05:29.238: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Aug 30 17:05:29.238: INFO: Waiting for statefulset status.replicas updated to 0
-Aug 30 17:05:29.243: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2
-Aug 30 17:05:39.254: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
-Aug 30 17:05:39.254: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
-Aug 30 17:05:39.254: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
-Aug 30 17:05:39.271: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
-Aug 30 17:05:39.271: INFO: ss-0  adoring-wozniak-54dcfd79fc-6rshr  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:04:54 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:28 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:28 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:04:54 +0000 UTC  }]
-Aug 30 17:05:39.271: INFO: ss-1  adoring-wozniak-54dcfd79fc-948mf  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:39.271: INFO: ss-2  adoring-wozniak-54dcfd79fc-948mf  Running         [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:39.271: INFO: 
-Aug 30 17:05:39.271: INFO: StatefulSet ss has not reached scale 0, at 3
-Aug 30 17:05:40.278: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
-Aug 30 17:05:40.278: INFO: ss-0  adoring-wozniak-54dcfd79fc-6rshr  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:04:54 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:28 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:28 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:04:54 +0000 UTC  }]
-Aug 30 17:05:40.278: INFO: ss-1  adoring-wozniak-54dcfd79fc-948mf  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:40.279: INFO: ss-2  adoring-wozniak-54dcfd79fc-948mf  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:40.279: INFO: 
-Aug 30 17:05:40.279: INFO: StatefulSet ss has not reached scale 0, at 3
-Aug 30 17:05:41.284: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
-Aug 30 17:05:41.284: INFO: ss-1  adoring-wozniak-54dcfd79fc-948mf  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:41.284: INFO: ss-2  adoring-wozniak-54dcfd79fc-948mf  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:41.287: INFO: 
-Aug 30 17:05:41.287: INFO: StatefulSet ss has not reached scale 0, at 2
-Aug 30 17:05:42.294: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
-Aug 30 17:05:42.294: INFO: ss-1  adoring-wozniak-54dcfd79fc-948mf  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:42.294: INFO: ss-2  adoring-wozniak-54dcfd79fc-948mf  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:42.294: INFO: 
-Aug 30 17:05:42.294: INFO: StatefulSet ss has not reached scale 0, at 2
-Aug 30 17:05:43.300: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
-Aug 30 17:05:43.300: INFO: ss-1  adoring-wozniak-54dcfd79fc-948mf  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:43.300: INFO: ss-2  adoring-wozniak-54dcfd79fc-948mf  Running  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:43.300: INFO: 
-Aug 30 17:05:43.300: INFO: StatefulSet ss has not reached scale 0, at 2
-Aug 30 17:05:44.305: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
-Aug 30 17:05:44.305: INFO: ss-1  adoring-wozniak-54dcfd79fc-948mf  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:44.305: INFO: 
-Aug 30 17:05:44.305: INFO: StatefulSet ss has not reached scale 0, at 1
-Aug 30 17:05:45.312: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
-Aug 30 17:05:45.312: INFO: ss-1  adoring-wozniak-54dcfd79fc-948mf  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:45.312: INFO: 
-Aug 30 17:05:45.312: INFO: StatefulSet ss has not reached scale 0, at 1
-Aug 30 17:05:46.318: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
-Aug 30 17:05:46.318: INFO: ss-1  adoring-wozniak-54dcfd79fc-948mf  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:46.318: INFO: 
-Aug 30 17:05:46.318: INFO: StatefulSet ss has not reached scale 0, at 1
-Aug 30 17:05:47.325: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
-Aug 30 17:05:47.325: INFO: ss-1  adoring-wozniak-54dcfd79fc-948mf  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:47.325: INFO: 
-Aug 30 17:05:47.325: INFO: StatefulSet ss has not reached scale 0, at 1
-Aug 30 17:05:48.332: INFO: POD   NODE                              PHASE    GRACE  CONDITIONS
-Aug 30 17:05:48.332: INFO: ss-1  adoring-wozniak-54dcfd79fc-948mf  Pending  30s    [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  } {Ready False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:29 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2020-08-30 17:05:15 +0000 UTC  }]
-Aug 30 17:05:48.332: INFO: 
-Aug 30 17:05:48.332: INFO: StatefulSet ss has not reached scale 0, at 1
-STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-8383
-Aug 30 17:05:49.337: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:05:49.563: INFO: rc: 1
-Aug 30 17:05:49.563: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-error: unable to upgrade connection: container not found ("webserver")
-
-error:
-exit status 1
-Aug 30 17:05:59.563: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:05:59.631: INFO: rc: 1
-Aug 30 17:05:59.631: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:06:09.632: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:06:09.698: INFO: rc: 1
-Aug 30 17:06:09.698: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:06:19.699: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:06:19.766: INFO: rc: 1
-Aug 30 17:06:19.766: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:06:29.767: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:06:29.833: INFO: rc: 1
-Aug 30 17:06:29.833: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:06:39.834: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:06:39.912: INFO: rc: 1
-Aug 30 17:06:39.912: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:06:49.912: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:06:49.990: INFO: rc: 1
-Aug 30 17:06:49.990: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:06:59.991: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:07:00.068: INFO: rc: 1
-Aug 30 17:07:00.068: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:07:10.068: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:07:10.141: INFO: rc: 1
-Aug 30 17:07:10.141: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:07:20.141: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:07:20.209: INFO: rc: 1
-Aug 30 17:07:20.209: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:07:30.209: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:07:30.275: INFO: rc: 1
-Aug 30 17:07:30.275: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:07:40.275: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:07:40.341: INFO: rc: 1
-Aug 30 17:07:40.341: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:07:50.341: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:07:50.411: INFO: rc: 1
-Aug 30 17:07:50.411: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:08:00.412: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:08:00.495: INFO: rc: 1
-Aug 30 17:08:00.495: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:08:10.496: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:08:10.563: INFO: rc: 1
-Aug 30 17:08:10.563: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:08:20.563: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:08:20.634: INFO: rc: 1
-Aug 30 17:08:20.634: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:08:30.634: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:08:30.701: INFO: rc: 1
-Aug 30 17:08:30.701: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:08:40.701: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:08:40.768: INFO: rc: 1
-Aug 30 17:08:40.768: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:08:50.769: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:08:50.836: INFO: rc: 1
-Aug 30 17:08:50.836: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:09:00.836: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:09:00.903: INFO: rc: 1
-Aug 30 17:09:00.903: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:09:10.903: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:09:10.983: INFO: rc: 1
-Aug 30 17:09:10.983: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:09:20.983: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:09:21.058: INFO: rc: 1
-Aug 30 17:09:21.058: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:09:31.058: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:09:31.130: INFO: rc: 1
-Aug 30 17:09:31.130: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
-
-stderr:
-Error from server (NotFound): pods "ss-1" not found
-
-error:
-exit status 1
-Aug 30 17:09:41.131: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:09:41.199: INFO: rc: 1
-Aug 30 17:09:41.199: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
+[It] should provide DNS for pods for Subdomain [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a test headless service
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-1017.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service-2.dns-1017.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-1017.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service-2.dns-1017.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1017.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
 
-stderr:
-Error from server (NotFound): pods "ss-1" not found
+STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-1017.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service-2.dns-1017.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-1017.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service-2.dns-1017.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-1017.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
 
-error:
-exit status 1
-Aug 30 17:09:51.199: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:09:51.266: INFO: rc: 1
-Aug 30 17:09:51.266: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
+STEP: creating a pod to probe DNS
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Feb 12 10:22:42.943: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local from pod dns-1017/dns-test-e68180d9-f749-484a-b012-09748215a9b5: the server could not find the requested resource (get pods dns-test-e68180d9-f749-484a-b012-09748215a9b5)
+Feb 12 10:22:42.987: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local from pod dns-1017/dns-test-e68180d9-f749-484a-b012-09748215a9b5: the server could not find the requested resource (get pods dns-test-e68180d9-f749-484a-b012-09748215a9b5)
+Feb 12 10:22:42.996: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-1017.svc.cluster.local from pod dns-1017/dns-test-e68180d9-f749-484a-b012-09748215a9b5: the server could not find the requested resource (get pods dns-test-e68180d9-f749-484a-b012-09748215a9b5)
+Feb 12 10:22:43.006: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-1017.svc.cluster.local from pod dns-1017/dns-test-e68180d9-f749-484a-b012-09748215a9b5: the server could not find the requested resource (get pods dns-test-e68180d9-f749-484a-b012-09748215a9b5)
+Feb 12 10:22:43.195: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local from pod dns-1017/dns-test-e68180d9-f749-484a-b012-09748215a9b5: the server could not find the requested resource (get pods dns-test-e68180d9-f749-484a-b012-09748215a9b5)
+Feb 12 10:22:43.203: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local from pod dns-1017/dns-test-e68180d9-f749-484a-b012-09748215a9b5: the server could not find the requested resource (get pods dns-test-e68180d9-f749-484a-b012-09748215a9b5)
+Feb 12 10:22:43.214: INFO: Unable to read jessie_udp@dns-test-service-2.dns-1017.svc.cluster.local from pod dns-1017/dns-test-e68180d9-f749-484a-b012-09748215a9b5: the server could not find the requested resource (get pods dns-test-e68180d9-f749-484a-b012-09748215a9b5)
+Feb 12 10:22:43.226: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-1017.svc.cluster.local from pod dns-1017/dns-test-e68180d9-f749-484a-b012-09748215a9b5: the server could not find the requested resource (get pods dns-test-e68180d9-f749-484a-b012-09748215a9b5)
+Feb 12 10:22:43.353: INFO: Lookups using dns-1017/dns-test-e68180d9-f749-484a-b012-09748215a9b5 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local wheezy_udp@dns-test-service-2.dns-1017.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-1017.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-1017.svc.cluster.local jessie_udp@dns-test-service-2.dns-1017.svc.cluster.local jessie_tcp@dns-test-service-2.dns-1017.svc.cluster.local]
+
+Feb 12 10:22:49.315: INFO: DNS probes using dns-1017/dns-test-e68180d9-f749-484a-b012-09748215a9b5 succeeded
 
-stderr:
-Error from server (NotFound): pods "ss-1" not found
+STEP: deleting the pod
+STEP: deleting the test headless service
+[AfterEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:22:49.360: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-1017" for this suite.
 
-error:
-exit status 1
-Aug 30 17:10:01.266: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:10:01.344: INFO: rc: 1
-Aug 30 17:10:01.344: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
+• [SLOW TEST:8.722 seconds]
+[sig-network] DNS
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should provide DNS for pods for Subdomain [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Subdomain [Conformance]","total":280,"completed":105,"skipped":1404,"failed":0}
+SSS
+------------------------------
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should mutate pod and apply defaults after mutation [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:22:49.382: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-7077
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Feb 12 10:22:50.088: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:22:52.107: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722169, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722169, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722169, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722169, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Feb 12 10:22:55.136: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should mutate pod and apply defaults after mutation [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Registering the mutating pod webhook via the AdmissionRegistration API
+STEP: create a pod that should be updated by the webhook
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:22:55.483: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-7077" for this suite.
+STEP: Destroying namespace "webhook-7077-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-stderr:
-Error from server (NotFound): pods "ss-1" not found
+• [SLOW TEST:6.215 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should mutate pod and apply defaults after mutation [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance]","total":280,"completed":106,"skipped":1407,"failed":0}
+SSSSS
+------------------------------
+[sig-cli] Kubectl client Kubectl run pod 
+  should create a pod from an image when restart is Never  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:22:55.598: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-4896
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[BeforeEach] Kubectl run pod
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1760
+[It] should create a pod from an image when restart is Never  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: running the image docker.io/library/httpd:2.4.38-alpine
+Feb 12 10:22:55.778: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 run e2e-test-httpd-pod --restart=Never --generator=run-pod/v1 --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-4896'
+Feb 12 10:22:55.914: INFO: stderr: ""
+Feb 12 10:22:55.914: INFO: stdout: "pod/e2e-test-httpd-pod created\n"
+STEP: verifying the pod e2e-test-httpd-pod was created
+[AfterEach] Kubectl run pod
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1765
+Feb 12 10:22:55.921: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete pods e2e-test-httpd-pod --namespace=kubectl-4896'
+Feb 12 10:23:02.235: INFO: stderr: ""
+Feb 12 10:23:02.235: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:23:02.235: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-4896" for this suite.
 
-error:
-exit status 1
-Aug 30 17:10:11.345: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:10:11.420: INFO: rc: 1
-Aug 30 17:10:11.420: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
+• [SLOW TEST:6.656 seconds]
+[sig-cli] Kubectl client
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Kubectl run pod
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1756
+    should create a pod from an image when restart is Never  [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never  [Conformance]","total":280,"completed":107,"skipped":1412,"failed":0}
+SSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-network] Services 
+  should serve a basic endpoint from pods  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:23:02.254: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename services
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-5841
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
+[It] should serve a basic endpoint from pods  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating service endpoint-test2 in namespace services-5841
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-5841 to expose endpoints map[]
+Feb 12 10:23:02.466: INFO: Get endpoints failed (7.156522ms elapsed, ignoring for 5s): endpoints "endpoint-test2" not found
+Feb 12 10:23:03.473: INFO: successfully validated that service endpoint-test2 in namespace services-5841 exposes endpoints map[] (1.014205464s elapsed)
+STEP: Creating pod pod1 in namespace services-5841
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-5841 to expose endpoints map[pod1:[80]]
+Feb 12 10:23:05.517: INFO: successfully validated that service endpoint-test2 in namespace services-5841 exposes endpoints map[pod1:[80]] (2.031658103s elapsed)
+STEP: Creating pod pod2 in namespace services-5841
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-5841 to expose endpoints map[pod1:[80] pod2:[80]]
+Feb 12 10:23:08.591: INFO: successfully validated that service endpoint-test2 in namespace services-5841 exposes endpoints map[pod1:[80] pod2:[80]] (3.062290219s elapsed)
+STEP: Deleting pod pod1 in namespace services-5841
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-5841 to expose endpoints map[pod2:[80]]
+Feb 12 10:23:08.627: INFO: successfully validated that service endpoint-test2 in namespace services-5841 exposes endpoints map[pod2:[80]] (22.817932ms elapsed)
+STEP: Deleting pod pod2 in namespace services-5841
+STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-5841 to expose endpoints map[]
+Feb 12 10:23:08.643: INFO: successfully validated that service endpoint-test2 in namespace services-5841 exposes endpoints map[] (5.394993ms elapsed)
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:23:08.668: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-5841" for this suite.
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
 
-stderr:
-Error from server (NotFound): pods "ss-1" not found
+• [SLOW TEST:6.430 seconds]
+[sig-network] Services
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should serve a basic endpoint from pods  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-network] Services should serve a basic endpoint from pods  [Conformance]","total":280,"completed":108,"skipped":1433,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected downwardAPI 
+  should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:23:08.686: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-924
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating the pod
+Feb 12 10:23:11.452: INFO: Successfully updated pod "annotationupdateae4e8c8c-c414-4e94-aba2-d56ac3a7d1fe"
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:23:13.485: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-924" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance]","total":280,"completed":109,"skipped":1464,"failed":0}
+SSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client Kubectl expose 
+  should create services for rc  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:23:13.502: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-4326
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[It] should create services for rc  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating Agnhost RC
+Feb 12 10:23:13.670: INFO: namespace kubectl-4326
+Feb 12 10:23:13.670: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-4326'
+Feb 12 10:23:13.962: INFO: stderr: ""
+Feb 12 10:23:13.963: INFO: stdout: "replicationcontroller/agnhost-master created\n"
+STEP: Waiting for Agnhost master to start.
+Feb 12 10:23:14.970: INFO: Selector matched 1 pods for map[app:agnhost]
+Feb 12 10:23:14.970: INFO: Found 0 / 1
+Feb 12 10:23:15.969: INFO: Selector matched 1 pods for map[app:agnhost]
+Feb 12 10:23:15.969: INFO: Found 1 / 1
+Feb 12 10:23:15.969: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+Feb 12 10:23:15.975: INFO: Selector matched 1 pods for map[app:agnhost]
+Feb 12 10:23:15.975: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Feb 12 10:23:15.975: INFO: wait on agnhost-master startup in kubectl-4326 
+Feb 12 10:23:15.975: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 logs agnhost-master-c99wn agnhost-master --namespace=kubectl-4326'
+Feb 12 10:23:16.222: INFO: stderr: ""
+Feb 12 10:23:16.222: INFO: stdout: "Paused\n"
+STEP: exposing RC
+Feb 12 10:23:16.222: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 expose rc agnhost-master --name=rm2 --port=1234 --target-port=6379 --namespace=kubectl-4326'
+Feb 12 10:23:16.357: INFO: stderr: ""
+Feb 12 10:23:16.357: INFO: stdout: "service/rm2 exposed\n"
+Feb 12 10:23:16.363: INFO: Service rm2 in namespace kubectl-4326 found.
+STEP: exposing service
+Feb 12 10:23:18.373: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 expose service rm2 --name=rm3 --port=2345 --target-port=6379 --namespace=kubectl-4326'
+Feb 12 10:23:18.468: INFO: stderr: ""
+Feb 12 10:23:18.468: INFO: stdout: "service/rm3 exposed\n"
+Feb 12 10:23:18.474: INFO: Service rm3 in namespace kubectl-4326 found.
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:23:20.486: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-4326" for this suite.
 
-error:
-exit status 1
-Aug 30 17:10:21.420: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:10:21.494: INFO: rc: 1
-Aug 30 17:10:21.494: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
+• [SLOW TEST:7.003 seconds]
+[sig-cli] Kubectl client
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Kubectl expose
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1189
+    should create services for rc  [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-cli] Kubectl client Kubectl expose should create services for rc  [Conformance]","total":280,"completed":110,"skipped":1474,"failed":0}
+SSSSSS
+------------------------------
+[sig-apps] Daemon set [Serial] 
+  should run and stop simple daemon [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:23:20.505: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename daemonsets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-4566
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
+[It] should run and stop simple daemon [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating simple DaemonSet "daemon-set"
+STEP: Check that daemon pods launch on every node of the cluster.
+Feb 12 10:23:20.736: INFO: Number of nodes with available pods: 0
+Feb 12 10:23:20.736: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod
+Feb 12 10:23:21.750: INFO: Number of nodes with available pods: 0
+Feb 12 10:23:21.750: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod
+Feb 12 10:23:22.766: INFO: Number of nodes with available pods: 1
+Feb 12 10:23:22.766: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod
+Feb 12 10:23:23.754: INFO: Number of nodes with available pods: 3
+Feb 12 10:23:23.754: INFO: Number of running nodes: 3, number of available pods: 3
+STEP: Stop a daemon pod, check that the daemon pod is revived.
+Feb 12 10:23:23.788: INFO: Number of nodes with available pods: 2
+Feb 12 10:23:23.788: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:23:24.801: INFO: Number of nodes with available pods: 2
+Feb 12 10:23:24.801: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:23:25.812: INFO: Number of nodes with available pods: 2
+Feb 12 10:23:25.812: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:23:26.802: INFO: Number of nodes with available pods: 2
+Feb 12 10:23:26.802: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:23:27.802: INFO: Number of nodes with available pods: 2
+Feb 12 10:23:27.802: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:23:28.801: INFO: Number of nodes with available pods: 2
+Feb 12 10:23:28.801: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:23:29.802: INFO: Number of nodes with available pods: 2
+Feb 12 10:23:29.802: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:23:30.802: INFO: Number of nodes with available pods: 2
+Feb 12 10:23:30.802: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:23:31.801: INFO: Number of nodes with available pods: 2
+Feb 12 10:23:31.801: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:23:32.802: INFO: Number of nodes with available pods: 2
+Feb 12 10:23:32.802: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:23:33.801: INFO: Number of nodes with available pods: 2
+Feb 12 10:23:33.801: INFO: Node nostalgic-fermat-cdc5d8777-r6snm is running more than one daemon pod
+Feb 12 10:23:34.803: INFO: Number of nodes with available pods: 3
+Feb 12 10:23:34.803: INFO: Number of running nodes: 3, number of available pods: 3
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-4566, will wait for the garbage collector to delete the pods
+Feb 12 10:23:34.876: INFO: Deleting DaemonSet.extensions daemon-set took: 11.889502ms
+Feb 12 10:23:34.977: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.332675ms
+Feb 12 10:23:42.483: INFO: Number of nodes with available pods: 0
+Feb 12 10:23:42.483: INFO: Number of running nodes: 0, number of available pods: 0
+Feb 12 10:23:42.488: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-4566/daemonsets","resourceVersion":"19784"},"items":null}
 
-stderr:
-Error from server (NotFound): pods "ss-1" not found
+Feb 12 10:23:42.493: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-4566/pods","resourceVersion":"19784"},"items":null}
 
-error:
-exit status 1
-Aug 30 17:10:31.495: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:10:31.564: INFO: rc: 1
-Aug 30 17:10:31.564: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:23:42.521: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "daemonsets-4566" for this suite.
 
-stderr:
-Error from server (NotFound): pods "ss-1" not found
+• [SLOW TEST:22.037 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should run and stop simple daemon [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance]","total":280,"completed":111,"skipped":1480,"failed":0}
+SSSSS
+------------------------------
+[sig-apps] ReplicationController 
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] ReplicationController
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:23:42.543: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename replication-controller
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-8922
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating replication controller my-hostname-basic-7b66f192-5398-4e6a-8f08-942a21f5f4a6
+Feb 12 10:23:42.713: INFO: Pod name my-hostname-basic-7b66f192-5398-4e6a-8f08-942a21f5f4a6: Found 0 pods out of 1
+Feb 12 10:23:47.721: INFO: Pod name my-hostname-basic-7b66f192-5398-4e6a-8f08-942a21f5f4a6: Found 1 pods out of 1
+Feb 12 10:23:47.721: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-7b66f192-5398-4e6a-8f08-942a21f5f4a6" are running
+Feb 12 10:23:47.727: INFO: Pod "my-hostname-basic-7b66f192-5398-4e6a-8f08-942a21f5f4a6-nf8cx" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-02-12 10:23:42 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-02-12 10:23:44 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-02-12 10:23:44 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-02-12 10:23:42 +0000 UTC Reason: Message:}])
+Feb 12 10:23:47.727: INFO: Trying to dial the pod
+Feb 12 10:23:52.839: INFO: Controller my-hostname-basic-7b66f192-5398-4e6a-8f08-942a21f5f4a6: Got expected result from replica 1 [my-hostname-basic-7b66f192-5398-4e6a-8f08-942a21f5f4a6-nf8cx]: "my-hostname-basic-7b66f192-5398-4e6a-8f08-942a21f5f4a6-nf8cx", 1 of 1 required successes so far
+[AfterEach] [sig-apps] ReplicationController
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:23:52.839: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replication-controller-8922" for this suite.
 
-error:
-exit status 1
-Aug 30 17:10:41.564: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:10:41.633: INFO: rc: 1
-Aug 30 17:10:41.633: INFO: Waiting 10s to retry failed RunHostCmd: error running /usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true:
-Command stdout:
+• [SLOW TEST:10.316 seconds]
+[sig-apps] ReplicationController
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-apps] ReplicationController should serve a basic image on each replica with a public image  [Conformance]","total":280,"completed":112,"skipped":1485,"failed":0}
+SSS
+------------------------------
+[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] 
+  should be able to convert from CR v1 to CR v2 [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:23:52.859: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename crd-webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-webhook-3308
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:125
+STEP: Setting up server cert
+STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication
+STEP: Deploying the custom resource conversion webhook pod
+STEP: Wait for the deployment to be ready
+Feb 12 10:23:53.717: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set
+Feb 12 10:23:55.742: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722233, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722233, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722233, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722233, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-78dcf5dd84\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Feb 12 10:23:58.766: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1
+[It] should be able to convert from CR v1 to CR v2 [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:23:58.772: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Creating a v1 custom resource
+STEP: v2 custom resource should be converted
+[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:24:00.191: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-webhook-3308" for this suite.
+[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:136
 
-stderr:
-Error from server (NotFound): pods "ss-1" not found
+• [SLOW TEST:7.440 seconds]
+[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should be able to convert from CR v1 to CR v2 [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance]","total":280,"completed":113,"skipped":1488,"failed":0}
+SSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:24:00.304: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-688
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir 0644 on node default medium
+Feb 12 10:24:00.552: INFO: Waiting up to 5m0s for pod "pod-f04ebffa-952d-4e61-a24b-3a0c56f83e88" in namespace "emptydir-688" to be "success or failure"
+Feb 12 10:24:00.560: INFO: Pod "pod-f04ebffa-952d-4e61-a24b-3a0c56f83e88": Phase="Pending", Reason="", readiness=false. Elapsed: 7.891939ms
+Feb 12 10:24:02.568: INFO: Pod "pod-f04ebffa-952d-4e61-a24b-3a0c56f83e88": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015473226s
+Feb 12 10:24:04.573: INFO: Pod "pod-f04ebffa-952d-4e61-a24b-3a0c56f83e88": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021157962s
+STEP: Saw pod success
+Feb 12 10:24:04.573: INFO: Pod "pod-f04ebffa-952d-4e61-a24b-3a0c56f83e88" satisfied condition "success or failure"
+Feb 12 10:24:04.578: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-f04ebffa-952d-4e61-a24b-3a0c56f83e88 container test-container: 
+STEP: delete the pod
+Feb 12 10:24:04.614: INFO: Waiting for pod pod-f04ebffa-952d-4e61-a24b-3a0c56f83e88 to disappear
+Feb 12 10:24:04.619: INFO: Pod pod-f04ebffa-952d-4e61-a24b-3a0c56f83e88 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:24:04.619: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-688" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":114,"skipped":1508,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] ResourceQuota 
+  should create a ResourceQuota and capture the life of a service. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:24:04.641: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename resourcequota
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-1377
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should create a ResourceQuota and capture the life of a service. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Counting existing ResourceQuota
+STEP: Creating a ResourceQuota
+STEP: Ensuring resource quota status is calculated
+STEP: Creating a Service
+STEP: Ensuring resource quota status captures service creation
+STEP: Deleting a Service
+STEP: Ensuring resource quota status released usage
+[AfterEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:24:15.899: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-1377" for this suite.
 
-error:
-exit status 1
-Aug 30 17:10:51.633: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8383 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:10:51.705: INFO: rc: 1
-Aug 30 17:10:51.705: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: 
-Aug 30 17:10:51.705: INFO: Scaling statefulset ss to 0
-Aug 30 17:10:51.723: INFO: Waiting for statefulset status.replicas updated to 0
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
-Aug 30 17:10:51.728: INFO: Deleting all statefulset in ns statefulset-8383
-Aug 30 17:10:51.733: INFO: Scaling statefulset ss to 0
-Aug 30 17:10:51.749: INFO: Waiting for statefulset status.replicas updated to 0
-Aug 30 17:10:51.753: INFO: Deleting statefulset ss
-[AfterEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:10:51.774: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-8383" for this suite.
+• [SLOW TEST:11.275 seconds]
+[sig-api-machinery] ResourceQuota
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should create a ResourceQuota and capture the life of a service. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance]","total":280,"completed":115,"skipped":1595,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  works for multiple CRDs of different groups [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:24:15.918: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename crd-publish-openapi
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-6396
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] works for multiple CRDs of different groups [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: CRs in different groups (two CRDs) show up in OpenAPI documentation
+Feb 12 10:24:16.078: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:24:18.985: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:24:30.313: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-6396" for this suite.
 
-• [SLOW TEST:357.721 seconds]
-[sig-apps] StatefulSet
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-    Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:14.412 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  works for multiple CRDs of different groups [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]","total":280,"completed":139,"skipped":2397,"failed":0}
-[sig-storage] Projected downwardAPI 
-  should update annotations on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance]","total":280,"completed":116,"skipped":1620,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:10:51.787: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:24:30.334: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1949
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6820
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should update annotations on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name projected-configmap-test-volume-map-c5d4395b-3bc1-4d2a-bcc2-6e888483085e
+STEP: Creating a pod to test consume configMaps
+Feb 12 10:24:30.532: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-42f300be-031c-4c8f-aebe-1531cfca339e" in namespace "projected-6820" to be "success or failure"
+Feb 12 10:24:30.538: INFO: Pod "pod-projected-configmaps-42f300be-031c-4c8f-aebe-1531cfca339e": Phase="Pending", Reason="", readiness=false. Elapsed: 5.687329ms
+Feb 12 10:24:32.550: INFO: Pod "pod-projected-configmaps-42f300be-031c-4c8f-aebe-1531cfca339e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.017578718s
+STEP: Saw pod success
+Feb 12 10:24:32.550: INFO: Pod "pod-projected-configmaps-42f300be-031c-4c8f-aebe-1531cfca339e" satisfied condition "success or failure"
+Feb 12 10:24:32.559: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-projected-configmaps-42f300be-031c-4c8f-aebe-1531cfca339e container projected-configmap-volume-test: 
+STEP: delete the pod
+Feb 12 10:24:32.612: INFO: Waiting for pod pod-projected-configmaps-42f300be-031c-4c8f-aebe-1531cfca339e to disappear
+Feb 12 10:24:32.617: INFO: Pod pod-projected-configmaps-42f300be-031c-4c8f-aebe-1531cfca339e no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:24:32.617: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-6820" for this suite.
+•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":117,"skipped":1650,"failed":0}
+SSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:24:32.634: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-2595
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir 0666 on node default medium
+Feb 12 10:24:32.812: INFO: Waiting up to 5m0s for pod "pod-65be91f3-d568-43d7-9597-782479a0bf66" in namespace "emptydir-2595" to be "success or failure"
+Feb 12 10:24:32.817: INFO: Pod "pod-65be91f3-d568-43d7-9597-782479a0bf66": Phase="Pending", Reason="", readiness=false. Elapsed: 5.081566ms
+Feb 12 10:24:34.823: INFO: Pod "pod-65be91f3-d568-43d7-9597-782479a0bf66": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011295949s
+STEP: Saw pod success
+Feb 12 10:24:34.823: INFO: Pod "pod-65be91f3-d568-43d7-9597-782479a0bf66" satisfied condition "success or failure"
+Feb 12 10:24:34.829: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-65be91f3-d568-43d7-9597-782479a0bf66 container test-container: 
+STEP: delete the pod
+Feb 12 10:24:34.864: INFO: Waiting for pod pod-65be91f3-d568-43d7-9597-782479a0bf66 to disappear
+Feb 12 10:24:34.868: INFO: Pod pod-65be91f3-d568-43d7-9597-782479a0bf66 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:24:34.868: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-2595" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":118,"skipped":1671,"failed":0}
+SSSSSSSSSSSSS
+------------------------------
+[k8s.io] Lease 
+  lease API should be available [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Lease
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:24:34.888: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename lease-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in lease-test-3780
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] lease API should be available [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[AfterEach] [k8s.io] Lease
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:24:35.157: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "lease-test-3780" for this suite.
+•{"msg":"PASSED [k8s.io] Lease lease API should be available [Conformance]","total":280,"completed":119,"skipped":1684,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Container Runtime blackbox test on terminated container 
+  should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Container Runtime
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:24:35.183: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-runtime
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-3679
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: create the container
+STEP: wait for the container to reach Succeeded
+STEP: get the container status
+STEP: the container should be terminated
+STEP: the termination message should be set
+Feb 12 10:24:37.400: INFO: Expected: &{DONE} to match Container's Termination Message: DONE --
+STEP: delete the container
+[AfterEach] [k8s.io] Container Runtime
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:24:37.420: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-runtime-3679" for this suite.
+•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]","total":280,"completed":120,"skipped":1712,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected secret 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:24:37.444: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8714
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating projection with secret that has name projected-secret-test-map-8ab0a55d-5cf7-4392-8d7f-9dabf0e95d17
+STEP: Creating a pod to test consume secrets
+Feb 12 10:24:37.618: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-031b2128-3a28-49c1-9789-fde5357589dc" in namespace "projected-8714" to be "success or failure"
+Feb 12 10:24:37.624: INFO: Pod "pod-projected-secrets-031b2128-3a28-49c1-9789-fde5357589dc": Phase="Pending", Reason="", readiness=false. Elapsed: 6.657764ms
+Feb 12 10:24:39.630: INFO: Pod "pod-projected-secrets-031b2128-3a28-49c1-9789-fde5357589dc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012314395s
+Feb 12 10:24:41.637: INFO: Pod "pod-projected-secrets-031b2128-3a28-49c1-9789-fde5357589dc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018746169s
+STEP: Saw pod success
+Feb 12 10:24:41.637: INFO: Pod "pod-projected-secrets-031b2128-3a28-49c1-9789-fde5357589dc" satisfied condition "success or failure"
+Feb 12 10:24:41.652: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-projected-secrets-031b2128-3a28-49c1-9789-fde5357589dc container projected-secret-volume-test: 
+STEP: delete the pod
+Feb 12 10:24:41.722: INFO: Waiting for pod pod-projected-secrets-031b2128-3a28-49c1-9789-fde5357589dc to disappear
+Feb 12 10:24:41.727: INFO: Pod pod-projected-secrets-031b2128-3a28-49c1-9789-fde5357589dc no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:24:41.728: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-8714" for this suite.
+•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":121,"skipped":1754,"failed":0}
+SSSSSSS
+------------------------------
+[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition 
+  getting/updating/patching custom resource definition status sub-resource works  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:24:41.755: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename custom-resource-definition
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-6190
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] getting/updating/patching custom resource definition status sub-resource works  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:24:41.935: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:24:42.010: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "custom-resource-definition-6190" for this suite.
+•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works  [Conformance]","total":280,"completed":122,"skipped":1761,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected configMap 
+  updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:24:42.031: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-5330
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating projection with configMap that has name projected-configmap-test-upd-4ff9f615-014e-47dc-bf9e-14d56f37a933
 STEP: Creating the pod
-Aug 30 17:10:54.523: INFO: Successfully updated pod "annotationupdate2de23eec-5a56-45aa-800b-b70f41425341"
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:10:58.570: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-1949" for this suite.
+STEP: Updating configmap projected-configmap-test-upd-4ff9f615-014e-47dc-bf9e-14d56f37a933
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:24:46.409: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-5330" for this suite.
+•{"msg":"PASSED [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":123,"skipped":1791,"failed":0}
+SSSSSSS
+------------------------------
+[sig-apps] ReplicaSet 
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] ReplicaSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:24:46.427: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename replicaset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replicaset-3808
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:24:46.586: INFO: Creating ReplicaSet my-hostname-basic-0405a9ab-989d-4f78-b830-c41214bde5bd
+Feb 12 10:24:46.599: INFO: Pod name my-hostname-basic-0405a9ab-989d-4f78-b830-c41214bde5bd: Found 0 pods out of 1
+Feb 12 10:24:51.605: INFO: Pod name my-hostname-basic-0405a9ab-989d-4f78-b830-c41214bde5bd: Found 1 pods out of 1
+Feb 12 10:24:51.605: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-0405a9ab-989d-4f78-b830-c41214bde5bd" is running
+Feb 12 10:24:51.612: INFO: Pod "my-hostname-basic-0405a9ab-989d-4f78-b830-c41214bde5bd-q6572" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-02-12 10:24:46 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-02-12 10:24:48 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-02-12 10:24:48 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2021-02-12 10:24:46 +0000 UTC Reason: Message:}])
+Feb 12 10:24:51.612: INFO: Trying to dial the pod
+Feb 12 10:24:56.723: INFO: Controller my-hostname-basic-0405a9ab-989d-4f78-b830-c41214bde5bd: Got expected result from replica 1 [my-hostname-basic-0405a9ab-989d-4f78-b830-c41214bde5bd-q6572]: "my-hostname-basic-0405a9ab-989d-4f78-b830-c41214bde5bd-q6572", 1 of 1 required successes so far
+[AfterEach] [sig-apps] ReplicaSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:24:56.723: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replicaset-3808" for this suite.
 
-• [SLOW TEST:6.797 seconds]
-[sig-storage] Projected downwardAPI
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:34
-  should update annotations on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:10.314 seconds]
+[sig-apps] ReplicaSet
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should serve a basic image on each replica with a public image  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance]","total":280,"completed":140,"skipped":2397,"failed":0}
-SSSSSSSSSS
+{"msg":"PASSED [sig-apps] ReplicaSet should serve a basic image on each replica with a public image  [Conformance]","total":280,"completed":124,"skipped":1798,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should include webhook resources in discovery documents [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should be able to deny custom resource creation, update and deletion [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:10:58.585: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:24:56.742: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-9188
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-7409
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
 STEP: Setting up server cert
 STEP: Create role binding to let webhook read extension-apiserver-authentication
 STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Aug 30 17:10:59.469: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created
-Aug 30 17:11:01.488: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404259, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404259, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404259, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404259, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:24:57.354: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:24:59.376: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722297, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722297, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722297, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722297, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Aug 30 17:11:04.508: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should include webhook resources in discovery documents [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: fetching the /apis discovery document
-STEP: finding the admissionregistration.k8s.io API group in the /apis discovery document
-STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis discovery document
-STEP: fetching the /apis/admissionregistration.k8s.io discovery document
-STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis/admissionregistration.k8s.io discovery document
-STEP: fetching the /apis/admissionregistration.k8s.io/v1 discovery document
-STEP: finding mutatingwebhookconfigurations and validatingwebhookconfigurations resources in the /apis/admissionregistration.k8s.io/v1 discovery document
+Feb 12 10:25:02.399: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should be able to deny custom resource creation, update and deletion [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:25:02.405: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Registering the custom resource webhook via the AdmissionRegistration API
+STEP: Creating a custom resource that should be denied by the webhook
+STEP: Creating a custom resource whose deletion would be denied by the webhook
+STEP: Updating the custom resource with disallowed data should be denied
+STEP: Deleting the custom resource should be denied
+STEP: Remove the offending key and value from the custom resource data
+STEP: Deleting the updated custom resource should be successful
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:11:04.520: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-9188" for this suite.
-STEP: Destroying namespace "webhook-9188-markers" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:25:04.379: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-7409" for this suite.
+STEP: Destroying namespace "webhook-7409-markers" for this suite.
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+
+• [SLOW TEST:7.752 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should be able to deny custom resource creation, update and deletion [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance]","total":280,"completed":125,"skipped":1828,"failed":0}
+SSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Secrets 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:25:04.495: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-8455
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating secret with name secret-test-map-2df3b59f-c916-47e6-8d1b-386f869f2b29
+STEP: Creating a pod to test consume secrets
+Feb 12 10:25:04.701: INFO: Waiting up to 5m0s for pod "pod-secrets-a7ec9314-1ab2-4024-9614-6e5dae6a220c" in namespace "secrets-8455" to be "success or failure"
+Feb 12 10:25:04.706: INFO: Pod "pod-secrets-a7ec9314-1ab2-4024-9614-6e5dae6a220c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.183208ms
+Feb 12 10:25:06.712: INFO: Pod "pod-secrets-a7ec9314-1ab2-4024-9614-6e5dae6a220c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010644648s
+STEP: Saw pod success
+Feb 12 10:25:06.712: INFO: Pod "pod-secrets-a7ec9314-1ab2-4024-9614-6e5dae6a220c" satisfied condition "success or failure"
+Feb 12 10:25:06.717: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-secrets-a7ec9314-1ab2-4024-9614-6e5dae6a220c container secret-volume-test: 
+STEP: delete the pod
+Feb 12 10:25:06.803: INFO: Waiting for pod pod-secrets-a7ec9314-1ab2-4024-9614-6e5dae6a220c to disappear
+Feb 12 10:25:06.808: INFO: Pod pod-secrets-a7ec9314-1ab2-4024-9614-6e5dae6a220c no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:25:06.808: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-8455" for this suite.
+•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":126,"skipped":1847,"failed":0}
+SSSSSSS
+------------------------------
+[sig-api-machinery] ResourceQuota 
+  should create a ResourceQuota and capture the life of a configMap. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:25:06.828: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename resourcequota
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-6519
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should create a ResourceQuota and capture the life of a configMap. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Counting existing ResourceQuota
+STEP: Creating a ResourceQuota
+STEP: Ensuring resource quota status is calculated
+STEP: Creating a ConfigMap
+STEP: Ensuring resource quota status captures configMap creation
+STEP: Deleting a ConfigMap
+STEP: Ensuring resource quota status released usage
+[AfterEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:25:23.100: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-6519" for this suite.
 
-• [SLOW TEST:6.023 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should include webhook resources in discovery documents [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:16.291 seconds]
+[sig-api-machinery] ResourceQuota
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should create a ResourceQuota and capture the life of a configMap. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance]","total":280,"completed":141,"skipped":2407,"failed":0}
-SSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance]","total":280,"completed":127,"skipped":1854,"failed":0}
+SSSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide container's memory limit [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Secrets 
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:11:04.609: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-6143
+Feb 12 10:25:23.119: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-8773
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should provide container's memory limit [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 17:11:04.782: INFO: Waiting up to 5m0s for pod "downwardapi-volume-a60e8e4a-34e1-47b8-8ed7-051d5b5d7447" in namespace "downward-api-6143" to be "success or failure"
-Aug 30 17:11:04.788: INFO: Pod "downwardapi-volume-a60e8e4a-34e1-47b8-8ed7-051d5b5d7447": Phase="Pending", Reason="", readiness=false. Elapsed: 6.035248ms
-Aug 30 17:11:06.794: INFO: Pod "downwardapi-volume-a60e8e4a-34e1-47b8-8ed7-051d5b5d7447": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012018932s
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating secret with name secret-test-7cfce5dd-58b4-46d0-afbf-5c07301252c2
+STEP: Creating a pod to test consume secrets
+Feb 12 10:25:23.321: INFO: Waiting up to 5m0s for pod "pod-secrets-95d48916-c101-4dc5-800a-da9e0717bd3e" in namespace "secrets-8773" to be "success or failure"
+Feb 12 10:25:23.330: INFO: Pod "pod-secrets-95d48916-c101-4dc5-800a-da9e0717bd3e": Phase="Pending", Reason="", readiness=false. Elapsed: 8.83052ms
+Feb 12 10:25:25.337: INFO: Pod "pod-secrets-95d48916-c101-4dc5-800a-da9e0717bd3e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015459342s
 STEP: Saw pod success
-Aug 30 17:11:06.794: INFO: Pod "downwardapi-volume-a60e8e4a-34e1-47b8-8ed7-051d5b5d7447" satisfied condition "success or failure"
-Aug 30 17:11:06.799: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod downwardapi-volume-a60e8e4a-34e1-47b8-8ed7-051d5b5d7447 container client-container: 
+Feb 12 10:25:25.337: INFO: Pod "pod-secrets-95d48916-c101-4dc5-800a-da9e0717bd3e" satisfied condition "success or failure"
+Feb 12 10:25:25.342: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-secrets-95d48916-c101-4dc5-800a-da9e0717bd3e container secret-volume-test: 
 STEP: delete the pod
-Aug 30 17:11:06.869: INFO: Waiting for pod downwardapi-volume-a60e8e4a-34e1-47b8-8ed7-051d5b5d7447 to disappear
-Aug 30 17:11:06.873: INFO: Pod downwardapi-volume-a60e8e4a-34e1-47b8-8ed7-051d5b5d7447 no longer exists
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:11:06.873: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-6143" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance]","total":280,"completed":142,"skipped":2416,"failed":0}
-SS
-------------------------------
-[sig-api-machinery] Watchers 
-  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:11:06.886: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename watch
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-1253
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to restart watching from the last resource version observed by the previous watch [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a watch on configmaps
-STEP: creating a new configmap
-STEP: modifying the configmap once
-STEP: closing the watch once it receives two notifications
-Aug 30 17:11:07.072: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed  watch-1253 /api/v1/namespaces/watch-1253/configmaps/e2e-watch-test-watch-closed 3302fdde-e7aa-4496-9b46-e8565092ab38 17435 0 2020-08-30 17:11:07 +0000 UTC   map[watch-this-configmap:watch-closed-and-restarted] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Aug 30 17:11:07.073: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed  watch-1253 /api/v1/namespaces/watch-1253/configmaps/e2e-watch-test-watch-closed 3302fdde-e7aa-4496-9b46-e8565092ab38 17436 0 2020-08-30 17:11:07 +0000 UTC   map[watch-this-configmap:watch-closed-and-restarted] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-STEP: modifying the configmap a second time, while the watch is closed
-STEP: creating a new watch on configmaps from the last resource version observed by the first watch
-STEP: deleting the configmap
-STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed
-Aug 30 17:11:07.105: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed  watch-1253 /api/v1/namespaces/watch-1253/configmaps/e2e-watch-test-watch-closed 3302fdde-e7aa-4496-9b46-e8565092ab38 17437 0 2020-08-30 17:11:07 +0000 UTC   map[watch-this-configmap:watch-closed-and-restarted] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Aug 30 17:11:07.105: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed  watch-1253 /api/v1/namespaces/watch-1253/configmaps/e2e-watch-test-watch-closed 3302fdde-e7aa-4496-9b46-e8565092ab38 17438 0 2020-08-30 17:11:07 +0000 UTC   map[watch-this-configmap:watch-closed-and-restarted] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-[AfterEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:11:07.105: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "watch-1253" for this suite.
-•{"msg":"PASSED [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance]","total":280,"completed":143,"skipped":2418,"failed":0}
-SSS
+Feb 12 10:25:25.382: INFO: Waiting for pod pod-secrets-95d48916-c101-4dc5-800a-da9e0717bd3e to disappear
+Feb 12 10:25:25.393: INFO: Pod pod-secrets-95d48916-c101-4dc5-800a-da9e0717bd3e no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:25:25.393: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-8773" for this suite.
+•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":128,"skipped":1859,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-network] DNS 
-  should provide DNS for pods for Hostname [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] ResourceQuota 
+  should create a ResourceQuota and capture the life of a replica set. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:11:07.122: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename dns
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-3450
+Feb 12 10:25:25.418: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename resourcequota
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-2177
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide DNS for pods for Hostname [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a test headless service
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-3450.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-2.dns-test-service-2.dns-3450.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/wheezy_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3450.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
-
-STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-3450.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-2.dns-test-service-2.dns-3450.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/jessie_hosts@dns-querier-2;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3450.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
-
-STEP: creating a pod to probe DNS
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Aug 30 17:11:12.056: INFO: DNS probes using dns-3450/dns-test-b62cf087-422c-4373-9bf3-5dcaa266f63a succeeded
+[It] should create a ResourceQuota and capture the life of a replica set. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Counting existing ResourceQuota
+STEP: Creating a ResourceQuota
+STEP: Ensuring resource quota status is calculated
+STEP: Creating a ReplicaSet
+STEP: Ensuring resource quota status captures replicaset creation
+STEP: Deleting a ReplicaSet
+STEP: Ensuring resource quota status released usage
+[AfterEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:25:36.667: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-2177" for this suite.
 
-STEP: deleting the pod
-STEP: deleting the test headless service
-[AfterEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:11:12.104: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-3450" for this suite.
-•{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Hostname [LinuxOnly] [Conformance]","total":280,"completed":144,"skipped":2421,"failed":0}
-SSSS
+• [SLOW TEST:11.271 seconds]
+[sig-api-machinery] ResourceQuota
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should create a ResourceQuota and capture the life of a replica set. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] 
-  should include custom resource definition resources in discovery documents [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance]","total":280,"completed":129,"skipped":1896,"failed":0}
+[sig-api-machinery] Secrets 
+  should be consumable from pods in env vars [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:11:12.120: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename custom-resource-definition
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-196
+Feb 12 10:25:36.689: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-3545
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should include custom resource definition resources in discovery documents [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: fetching the /apis discovery document
-STEP: finding the apiextensions.k8s.io API group in the /apis discovery document
-STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis discovery document
-STEP: fetching the /apis/apiextensions.k8s.io discovery document
-STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis/apiextensions.k8s.io discovery document
-STEP: fetching the /apis/apiextensions.k8s.io/v1 discovery document
-STEP: finding customresourcedefinitions resources in the /apis/apiextensions.k8s.io/v1 discovery document
-[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:11:12.295: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "custom-resource-definition-196" for this suite.
-•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance]","total":280,"completed":145,"skipped":2425,"failed":0}
-
+[It] should be consumable from pods in env vars [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating secret with name secret-test-3a1842a9-f675-44c8-9524-de1611fdeba3
+STEP: Creating a pod to test consume secrets
+Feb 12 10:25:36.871: INFO: Waiting up to 5m0s for pod "pod-secrets-d783a363-e51f-418b-bb4f-cdc039b52c3f" in namespace "secrets-3545" to be "success or failure"
+Feb 12 10:25:36.879: INFO: Pod "pod-secrets-d783a363-e51f-418b-bb4f-cdc039b52c3f": Phase="Pending", Reason="", readiness=false. Elapsed: 8.011266ms
+Feb 12 10:25:38.887: INFO: Pod "pod-secrets-d783a363-e51f-418b-bb4f-cdc039b52c3f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016362705s
+STEP: Saw pod success
+Feb 12 10:25:38.888: INFO: Pod "pod-secrets-d783a363-e51f-418b-bb4f-cdc039b52c3f" satisfied condition "success or failure"
+Feb 12 10:25:38.894: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-secrets-d783a363-e51f-418b-bb4f-cdc039b52c3f container secret-env-test: 
+STEP: delete the pod
+Feb 12 10:25:38.985: INFO: Waiting for pod pod-secrets-d783a363-e51f-418b-bb4f-cdc039b52c3f to disappear
+Feb 12 10:25:38.989: INFO: Pod pod-secrets-d783a363-e51f-418b-bb4f-cdc039b52c3f no longer exists
+[AfterEach] [sig-api-machinery] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:25:38.989: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-3545" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance]","total":280,"completed":130,"skipped":1896,"failed":0}
+S
 ------------------------------
-[sig-network] Networking Granular Checks: Pods 
-  should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Networking
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Probing container 
+  should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:11:12.309: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename pod-network-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-9339
+Feb 12 10:25:39.013: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-623
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Performing setup for networking test in namespace pod-network-test-9339
-STEP: creating a selector
-STEP: Creating the service pods in kubernetes
-Aug 30 17:11:12.467: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
-STEP: Creating test pods
-Aug 30 17:11:32.566: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.25.0.132:8080/dial?request=hostname&protocol=http&host=172.25.1.58&port=8080&tries=1'] Namespace:pod-network-test-9339 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 17:11:32.566: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 17:11:33.059: INFO: Waiting for responses: map[]
-Aug 30 17:11:33.064: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://172.25.0.132:8080/dial?request=hostname&protocol=http&host=172.25.0.131&port=8080&tries=1'] Namespace:pod-network-test-9339 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
-Aug 30 17:11:33.064: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-Aug 30 17:11:33.606: INFO: Waiting for responses: map[]
-[AfterEach] [sig-network] Networking
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:11:33.606: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pod-network-test-9339" for this suite.
+[BeforeEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating pod test-webserver-0a44351d-9038-4d97-a273-437384cf5c03 in namespace container-probe-623
+Feb 12 10:25:41.200: INFO: Started pod test-webserver-0a44351d-9038-4d97-a273-437384cf5c03 in namespace container-probe-623
+STEP: checking the pod's current state and verifying that restartCount is present
+Feb 12 10:25:41.206: INFO: Initial restart count of pod test-webserver-0a44351d-9038-4d97-a273-437384cf5c03 is 0
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:29:42.198: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-623" for this suite.
 
-• [SLOW TEST:21.313 seconds]
-[sig-network] Networking
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
-  Granular Checks: Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
-    should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:243.211 seconds]
+[k8s.io] Probing container
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":146,"skipped":2425,"failed":0}
-SS
+{"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":280,"completed":131,"skipped":1897,"failed":0}
+SSSSSSSSSSSS
 ------------------------------
-[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
-  should execute prestop http hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Probing container 
+  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:11:33.622: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-lifecycle-hook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-3270
+Feb 12 10:29:42.225: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-6343
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] when create a pod with lifecycle hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
-STEP: create the container to handle the HTTPGet hook request.
-[It] should execute prestop http hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the pod with lifecycle hook
-STEP: delete the pod with lifecycle hook
-Aug 30 17:11:37.862: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Aug 30 17:11:37.867: INFO: Pod pod-with-prestop-http-hook still exists
-Aug 30 17:11:39.868: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Aug 30 17:11:39.874: INFO: Pod pod-with-prestop-http-hook still exists
-Aug 30 17:11:41.868: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Aug 30 17:11:41.873: INFO: Pod pod-with-prestop-http-hook still exists
-Aug 30 17:11:43.868: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Aug 30 17:11:43.873: INFO: Pod pod-with-prestop-http-hook still exists
-Aug 30 17:11:45.867: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Aug 30 17:11:45.875: INFO: Pod pod-with-prestop-http-hook still exists
-Aug 30 17:11:47.868: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Aug 30 17:11:47.874: INFO: Pod pod-with-prestop-http-hook still exists
-Aug 30 17:11:49.867: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Aug 30 17:11:49.873: INFO: Pod pod-with-prestop-http-hook still exists
-Aug 30 17:11:51.867: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Aug 30 17:11:51.874: INFO: Pod pod-with-prestop-http-hook still exists
-Aug 30 17:11:53.868: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Aug 30 17:11:53.874: INFO: Pod pod-with-prestop-http-hook still exists
-Aug 30 17:11:55.868: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
-Aug 30 17:11:55.874: INFO: Pod pod-with-prestop-http-hook no longer exists
-STEP: check prestop hook
-[AfterEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:11:55.928: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-lifecycle-hook-3270" for this suite.
+[BeforeEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:30:08.420: INFO: Container started at 2021-02-12 10:29:43 +0000 UTC, pod became ready at 2021-02-12 10:30:07 +0000 UTC
+[AfterEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:08.420: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-6343" for this suite.
 
-• [SLOW TEST:22.326 seconds]
-[k8s.io] Container Lifecycle Hook
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  when create a pod with lifecycle hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
-    should execute prestop http hook properly [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance]","total":280,"completed":147,"skipped":2427,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-storage] Projected configMap 
-  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:11:55.949: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7121
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name projected-configmap-test-volume-25b95375-48b8-47e8-ab05-05e6da7b655c
-STEP: Creating a pod to test consume configMaps
-Aug 30 17:11:56.135: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-ba6c501f-446a-4d18-a11b-7caa195a3427" in namespace "projected-7121" to be "success or failure"
-Aug 30 17:11:56.140: INFO: Pod "pod-projected-configmaps-ba6c501f-446a-4d18-a11b-7caa195a3427": Phase="Pending", Reason="", readiness=false. Elapsed: 5.427442ms
-Aug 30 17:11:58.147: INFO: Pod "pod-projected-configmaps-ba6c501f-446a-4d18-a11b-7caa195a3427": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011633213s
-STEP: Saw pod success
-Aug 30 17:11:58.147: INFO: Pod "pod-projected-configmaps-ba6c501f-446a-4d18-a11b-7caa195a3427" satisfied condition "success or failure"
-Aug 30 17:11:58.151: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-projected-configmaps-ba6c501f-446a-4d18-a11b-7caa195a3427 container projected-configmap-volume-test: 
-STEP: delete the pod
-Aug 30 17:11:58.221: INFO: Waiting for pod pod-projected-configmaps-ba6c501f-446a-4d18-a11b-7caa195a3427 to disappear
-Aug 30 17:11:58.226: INFO: Pod pod-projected-configmaps-ba6c501f-446a-4d18-a11b-7caa195a3427 no longer exists
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:11:58.226: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-7121" for this suite.
-•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":148,"skipped":2451,"failed":0}
-SSSSSSSSSSSSSSSSS
+• [SLOW TEST:26.213 seconds]
+[k8s.io] Probing container
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] InitContainer [NodeConformance] 
-  should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:11:58.243: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename init-container
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-2355
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
-[It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-Aug 30 17:11:58.410: INFO: PodSpec: initContainers in spec.initContainers
-[AfterEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:12:01.349: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "init-container-2355" for this suite.
-•{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]","total":280,"completed":149,"skipped":2468,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [k8s.io] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]","total":280,"completed":132,"skipped":1909,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] Daemon set [Serial] 
-  should rollback without unnecessary restarts [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-cli] Kubectl client Guestbook application 
+  should create and stop a working application  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:12:01.365: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename daemonsets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-257
+Feb 12 10:30:08.443: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2761
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
-[It] should rollback without unnecessary restarts [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:12:01.558: INFO: Create a RollingUpdate DaemonSet
-Aug 30 17:12:01.567: INFO: Check that daemon pods launch on every node of the cluster
-Aug 30 17:12:01.578: INFO: Number of nodes with available pods: 0
-Aug 30 17:12:01.578: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 17:12:02.589: INFO: Number of nodes with available pods: 0
-Aug 30 17:12:02.589: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 17:12:03.588: INFO: Number of nodes with available pods: 2
-Aug 30 17:12:03.588: INFO: Number of running nodes: 2, number of available pods: 2
-Aug 30 17:12:03.588: INFO: Update the DaemonSet to trigger a rollout
-Aug 30 17:12:03.602: INFO: Updating DaemonSet daemon-set
-Aug 30 17:12:14.626: INFO: Roll back the DaemonSet before rollout is complete
-Aug 30 17:12:14.641: INFO: Updating DaemonSet daemon-set
-Aug 30 17:12:14.641: INFO: Make sure DaemonSet rollback is complete
-Aug 30 17:12:14.647: INFO: Wrong image for pod: daemon-set-nkmgl. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent.
-Aug 30 17:12:14.647: INFO: Pod daemon-set-nkmgl is not available
-Aug 30 17:12:15.658: INFO: Wrong image for pod: daemon-set-nkmgl. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent.
-Aug 30 17:12:15.658: INFO: Pod daemon-set-nkmgl is not available
-Aug 30 17:12:17.663: INFO: Pod daemon-set-jk2v5 is not available
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-257, will wait for the garbage collector to delete the pods
-Aug 30 17:12:17.751: INFO: Deleting DaemonSet.extensions daemon-set took: 13.73808ms
-Aug 30 17:12:18.251: INFO: Terminating DaemonSet.extensions daemon-set pods took: 500.17222ms
-Aug 30 17:13:34.357: INFO: Number of nodes with available pods: 0
-Aug 30 17:13:34.357: INFO: Number of running nodes: 0, number of available pods: 0
-Aug 30 17:13:34.364: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-257/daemonsets","resourceVersion":"18363"},"items":null}
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[It] should create and stop a working application  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating all guestbook components
+Feb 12 10:30:08.602: INFO: apiVersion: v1
+kind: Service
+metadata:
+  name: agnhost-slave
+  labels:
+    app: agnhost
+    role: slave
+    tier: backend
+spec:
+  ports:
+  - port: 6379
+  selector:
+    app: agnhost
+    role: slave
+    tier: backend
+
+Feb 12 10:30:08.602: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-2761'
+Feb 12 10:30:08.860: INFO: stderr: ""
+Feb 12 10:30:08.860: INFO: stdout: "service/agnhost-slave created\n"
+Feb 12 10:30:08.860: INFO: apiVersion: v1
+kind: Service
+metadata:
+  name: agnhost-master
+  labels:
+    app: agnhost
+    role: master
+    tier: backend
+spec:
+  ports:
+  - port: 6379
+    targetPort: 6379
+  selector:
+    app: agnhost
+    role: master
+    tier: backend
+
+Feb 12 10:30:08.860: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-2761'
+Feb 12 10:30:09.080: INFO: stderr: ""
+Feb 12 10:30:09.080: INFO: stdout: "service/agnhost-master created\n"
+Feb 12 10:30:09.080: INFO: apiVersion: v1
+kind: Service
+metadata:
+  name: frontend
+  labels:
+    app: guestbook
+    tier: frontend
+spec:
+  # if your cluster supports it, uncomment the following to automatically create
+  # an external load-balanced IP for the frontend service.
+  # type: LoadBalancer
+  ports:
+  - port: 80
+  selector:
+    app: guestbook
+    tier: frontend
+
+Feb 12 10:30:09.080: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-2761'
+Feb 12 10:30:09.351: INFO: stderr: ""
+Feb 12 10:30:09.351: INFO: stdout: "service/frontend created\n"
+Feb 12 10:30:09.351: INFO: apiVersion: apps/v1
+kind: Deployment
+metadata:
+  name: frontend
+spec:
+  replicas: 3
+  selector:
+    matchLabels:
+      app: guestbook
+      tier: frontend
+  template:
+    metadata:
+      labels:
+        app: guestbook
+        tier: frontend
+    spec:
+      containers:
+      - name: guestbook-frontend
+        image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8
+        args: [ "guestbook", "--backend-port", "6379" ]
+        resources:
+          requests:
+            cpu: 100m
+            memory: 100Mi
+        ports:
+        - containerPort: 80
 
-Aug 30 17:13:34.370: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-257/pods","resourceVersion":"18363"},"items":null}
+Feb 12 10:30:09.351: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-2761'
+Feb 12 10:30:09.511: INFO: stderr: ""
+Feb 12 10:30:09.511: INFO: stdout: "deployment.apps/frontend created\n"
+Feb 12 10:30:09.511: INFO: apiVersion: apps/v1
+kind: Deployment
+metadata:
+  name: agnhost-master
+spec:
+  replicas: 1
+  selector:
+    matchLabels:
+      app: agnhost
+      role: master
+      tier: backend
+  template:
+    metadata:
+      labels:
+        app: agnhost
+        role: master
+        tier: backend
+    spec:
+      containers:
+      - name: master
+        image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8
+        args: [ "guestbook", "--http-port", "6379" ]
+        resources:
+          requests:
+            cpu: 100m
+            memory: 100Mi
+        ports:
+        - containerPort: 6379
 
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:13:34.388: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "daemonsets-257" for this suite.
+Feb 12 10:30:09.511: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-2761'
+Feb 12 10:30:09.672: INFO: stderr: ""
+Feb 12 10:30:09.672: INFO: stdout: "deployment.apps/agnhost-master created\n"
+Feb 12 10:30:09.672: INFO: apiVersion: apps/v1
+kind: Deployment
+metadata:
+  name: agnhost-slave
+spec:
+  replicas: 2
+  selector:
+    matchLabels:
+      app: agnhost
+      role: slave
+      tier: backend
+  template:
+    metadata:
+      labels:
+        app: agnhost
+        role: slave
+        tier: backend
+    spec:
+      containers:
+      - name: slave
+        image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8
+        args: [ "guestbook", "--slaveof", "agnhost-master", "--http-port", "6379" ]
+        resources:
+          requests:
+            cpu: 100m
+            memory: 100Mi
+        ports:
+        - containerPort: 6379
 
-• [SLOW TEST:93.036 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should rollback without unnecessary restarts [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance]","total":280,"completed":150,"skipped":2508,"failed":0}
-SSSSSSSSSSS
-------------------------------
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Subpath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:13:34.402: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename subpath
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-5036
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
-STEP: Setting up data
-[It] should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod pod-subpath-test-configmap-zq6l
-STEP: Creating a pod to test atomic-volume-subpath
-Aug 30 17:13:34.627: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-zq6l" in namespace "subpath-5036" to be "success or failure"
-Aug 30 17:13:34.636: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Pending", Reason="", readiness=false. Elapsed: 8.678941ms
-Aug 30 17:13:36.644: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Running", Reason="", readiness=true. Elapsed: 2.016952632s
-Aug 30 17:13:38.663: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Running", Reason="", readiness=true. Elapsed: 4.036120864s
-Aug 30 17:13:40.669: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Running", Reason="", readiness=true. Elapsed: 6.041687425s
-Aug 30 17:13:42.675: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Running", Reason="", readiness=true. Elapsed: 8.048260073s
-Aug 30 17:13:44.681: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Running", Reason="", readiness=true. Elapsed: 10.054519263s
-Aug 30 17:13:46.687: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Running", Reason="", readiness=true. Elapsed: 12.060296521s
-Aug 30 17:13:48.693: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Running", Reason="", readiness=true. Elapsed: 14.065941811s
-Aug 30 17:13:50.699: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Running", Reason="", readiness=true. Elapsed: 16.07193876s
-Aug 30 17:13:52.705: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Running", Reason="", readiness=true. Elapsed: 18.077899193s
-Aug 30 17:13:54.711: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Running", Reason="", readiness=true. Elapsed: 20.083738302s
-Aug 30 17:13:56.716: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Running", Reason="", readiness=true. Elapsed: 22.089390209s
-Aug 30 17:13:58.722: INFO: Pod "pod-subpath-test-configmap-zq6l": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.095037378s
-STEP: Saw pod success
-Aug 30 17:13:58.722: INFO: Pod "pod-subpath-test-configmap-zq6l" satisfied condition "success or failure"
-Aug 30 17:13:58.727: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-subpath-test-configmap-zq6l container test-container-subpath-configmap-zq6l: 
-STEP: delete the pod
-Aug 30 17:13:58.797: INFO: Waiting for pod pod-subpath-test-configmap-zq6l to disappear
-Aug 30 17:13:58.802: INFO: Pod pod-subpath-test-configmap-zq6l no longer exists
-STEP: Deleting pod pod-subpath-test-configmap-zq6l
-Aug 30 17:13:58.802: INFO: Deleting pod "pod-subpath-test-configmap-zq6l" in namespace "subpath-5036"
-[AfterEach] [sig-storage] Subpath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:13:58.809: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "subpath-5036" for this suite.
+Feb 12 10:30:09.672: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-2761'
+Feb 12 10:30:09.836: INFO: stderr: ""
+Feb 12 10:30:09.837: INFO: stdout: "deployment.apps/agnhost-slave created\n"
+STEP: validating guestbook app
+Feb 12 10:30:09.837: INFO: Waiting for all frontend pods to be Running.
+Feb 12 10:30:14.887: INFO: Waiting for frontend to serve content.
+Feb 12 10:30:14.985: INFO: Trying to add a new entry to the guestbook.
+Feb 12 10:30:15.118: INFO: Verifying that added entry can be retrieved.
+STEP: using delete to clean up resources
+Feb 12 10:30:15.166: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete --grace-period=0 --force -f - --namespace=kubectl-2761'
+Feb 12 10:30:15.277: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Feb 12 10:30:15.277: INFO: stdout: "service \"agnhost-slave\" force deleted\n"
+STEP: using delete to clean up resources
+Feb 12 10:30:15.277: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete --grace-period=0 --force -f - --namespace=kubectl-2761'
+Feb 12 10:30:15.394: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Feb 12 10:30:15.394: INFO: stdout: "service \"agnhost-master\" force deleted\n"
+STEP: using delete to clean up resources
+Feb 12 10:30:15.394: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete --grace-period=0 --force -f - --namespace=kubectl-2761'
+Feb 12 10:30:15.512: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Feb 12 10:30:15.512: INFO: stdout: "service \"frontend\" force deleted\n"
+STEP: using delete to clean up resources
+Feb 12 10:30:15.512: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete --grace-period=0 --force -f - --namespace=kubectl-2761'
+Feb 12 10:30:15.627: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Feb 12 10:30:15.627: INFO: stdout: "deployment.apps \"frontend\" force deleted\n"
+STEP: using delete to clean up resources
+Feb 12 10:30:15.628: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete --grace-period=0 --force -f - --namespace=kubectl-2761'
+Feb 12 10:30:15.731: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Feb 12 10:30:15.731: INFO: stdout: "deployment.apps \"agnhost-master\" force deleted\n"
+STEP: using delete to clean up resources
+Feb 12 10:30:15.734: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete --grace-period=0 --force -f - --namespace=kubectl-2761'
+Feb 12 10:30:15.835: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Feb 12 10:30:15.835: INFO: stdout: "deployment.apps \"agnhost-slave\" force deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:15.837: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-2761" for this suite.
 
-• [SLOW TEST:24.422 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
-  Atomic writer volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
-    should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]","total":280,"completed":151,"skipped":2519,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-storage] Projected configMap 
-  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:13:58.824: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3244
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name projected-configmap-test-volume-b4dee1d3-bfb6-4ee3-a653-214c28232fb9
-STEP: Creating a pod to test consume configMaps
-Aug 30 17:13:59.016: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-7b738ac2-fa26-4e00-9131-c8d53cf7a28e" in namespace "projected-3244" to be "success or failure"
-Aug 30 17:13:59.022: INFO: Pod "pod-projected-configmaps-7b738ac2-fa26-4e00-9131-c8d53cf7a28e": Phase="Pending", Reason="", readiness=false. Elapsed: 6.166681ms
-Aug 30 17:14:01.029: INFO: Pod "pod-projected-configmaps-7b738ac2-fa26-4e00-9131-c8d53cf7a28e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.01291461s
-STEP: Saw pod success
-Aug 30 17:14:01.029: INFO: Pod "pod-projected-configmaps-7b738ac2-fa26-4e00-9131-c8d53cf7a28e" satisfied condition "success or failure"
-Aug 30 17:14:01.034: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-projected-configmaps-7b738ac2-fa26-4e00-9131-c8d53cf7a28e container projected-configmap-volume-test: 
-STEP: delete the pod
-Aug 30 17:14:01.109: INFO: Waiting for pod pod-projected-configmaps-7b738ac2-fa26-4e00-9131-c8d53cf7a28e to disappear
-Aug 30 17:14:01.114: INFO: Pod pod-projected-configmaps-7b738ac2-fa26-4e00-9131-c8d53cf7a28e no longer exists
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:14:01.114: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-3244" for this suite.
-•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","total":280,"completed":152,"skipped":2543,"failed":0}
-SSSSSSSSS
+• [SLOW TEST:7.413 seconds]
+[sig-cli] Kubectl client
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Guestbook application
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:381
+    should create and stop a working application  [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:14:01.129: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-8957
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-map-3cc2fcc6-092b-47a8-8928-65bc4fa397b5
-STEP: Creating a pod to test consume secrets
-Aug 30 17:14:01.319: INFO: Waiting up to 5m0s for pod "pod-secrets-22a85b26-b08b-4472-bf08-b9c559bfc7a4" in namespace "secrets-8957" to be "success or failure"
-Aug 30 17:14:01.327: INFO: Pod "pod-secrets-22a85b26-b08b-4472-bf08-b9c559bfc7a4": Phase="Pending", Reason="", readiness=false. Elapsed: 7.705686ms
-Aug 30 17:14:03.333: INFO: Pod "pod-secrets-22a85b26-b08b-4472-bf08-b9c559bfc7a4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.013768796s
-STEP: Saw pod success
-Aug 30 17:14:03.333: INFO: Pod "pod-secrets-22a85b26-b08b-4472-bf08-b9c559bfc7a4" satisfied condition "success or failure"
-Aug 30 17:14:03.338: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-secrets-22a85b26-b08b-4472-bf08-b9c559bfc7a4 container secret-volume-test: 
-STEP: delete the pod
-Aug 30 17:14:03.409: INFO: Waiting for pod pod-secrets-22a85b26-b08b-4472-bf08-b9c559bfc7a4 to disappear
-Aug 30 17:14:03.414: INFO: Pod pod-secrets-22a85b26-b08b-4472-bf08-b9c559bfc7a4 no longer exists
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:14:03.414: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-8957" for this suite.
-•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":153,"skipped":2552,"failed":0}
-SSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-cli] Kubectl client Guestbook application should create and stop a working application  [Conformance]","total":280,"completed":133,"skipped":2042,"failed":0}
+SSS
 ------------------------------
-[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] 
-  custom resource defaulting for requests and from storage works  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] ResourceQuota 
+  should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:14:03.428: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename custom-resource-definition
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-2513
+Feb 12 10:30:15.856: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename resourcequota
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-7478
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] custom resource defaulting for requests and from storage works  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:14:03.590: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:14:04.805: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "custom-resource-definition-2513" for this suite.
-•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works  [Conformance]","total":280,"completed":154,"skipped":2571,"failed":0}
-SSSSS
+[It] should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Counting existing ResourceQuota
+STEP: Creating a ResourceQuota
+STEP: Ensuring resource quota status is calculated
+[AfterEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:23.034: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-7478" for this suite.
+
+• [SLOW TEST:7.200 seconds]
+[sig-api-machinery] ResourceQuota
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-network] Proxy version v1 
-  should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] version v1
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:14:04.845: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename proxy
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-9598
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:14:05.121: INFO: (0) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 45.650215ms)
-Aug 30 17:14:05.171: INFO: (1) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 49.26616ms)
-Aug 30 17:14:05.181: INFO: (2) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 10.485906ms)
-Aug 30 17:14:05.266: INFO: (3) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 84.735259ms)
-Aug 30 17:14:05.276: INFO: (4) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 10.209102ms)
-Aug 30 17:14:05.324: INFO: (5) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 47.961764ms)
-Aug 30 17:14:05.334: INFO: (6) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 9.846851ms)
-Aug 30 17:14:05.343: INFO: (7) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 8.728188ms)
-Aug 30 17:14:05.351: INFO: (8) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 8.534128ms)
-Aug 30 17:14:05.360: INFO: (9) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 8.518062ms)
-Aug 30 17:14:05.368: INFO: (10) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 8.650015ms)
-Aug 30 17:14:05.377: INFO: (11) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 8.084091ms)
-Aug 30 17:14:05.385: INFO: (12) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 8.498956ms)
-Aug 30 17:14:05.393: INFO: (13) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 8.098914ms)
-Aug 30 17:14:05.401: INFO: (14) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 7.923056ms)
-Aug 30 17:14:05.517: INFO: (15) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 115.609524ms)
-Aug 30 17:14:05.561: INFO: (16) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 44.013959ms)
-Aug 30 17:14:05.608: INFO: (17) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 47.327446ms)
-Aug 30 17:14:05.618: INFO: (18) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 9.548809ms)
-Aug 30 17:14:05.665: INFO: (19) /api/v1/nodes/adoring-wozniak-54dcfd79fc-948mf:10250/proxy/logs/: 
-alternatives.log
-apt/
-... (200; 47.195692ms)
-[AfterEach] version v1
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:14:05.665: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "proxy-9598" for this suite.
-•{"msg":"PASSED [sig-network] Proxy version v1 should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]","total":280,"completed":155,"skipped":2576,"failed":0}
-SSS
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]","total":280,"completed":134,"skipped":2045,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Pods 
-  should support remote command execution over websockets [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-cli] Kubectl client Kubectl cluster-info 
+  should check if Kubernetes master services is included in cluster-info  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:14:05.680: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-377
+Feb 12 10:30:23.056: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-997
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
-[It] should support remote command execution over websockets [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:14:05.846: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: creating the pod
-STEP: submitting the pod to kubernetes
-[AfterEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:14:08.080: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-377" for this suite.
-•{"msg":"PASSED [k8s.io] Pods should support remote command execution over websockets [NodeConformance] [Conformance]","total":280,"completed":156,"skipped":2579,"failed":0}
-S
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[It] should check if Kubernetes master services is included in cluster-info  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: validating cluster-info
+Feb 12 10:30:23.248: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 cluster-info'
+Feb 12 10:30:23.331: INFO: stderr: ""
+Feb 12 10:30:23.331: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://10.240.16.1:443\x1b[0m\n\x1b[0;32mkube-dns\x1b[0m is running at \x1b[0;33mhttps://10.240.16.1:443/api/v1/namespaces/kube-system/services/kube-dns:dns-tcp/proxy\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:23.331: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-997" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes master services is included in cluster-info  [Conformance]","total":280,"completed":135,"skipped":2067,"failed":0}
+SSSS
 ------------------------------
 [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
   works for CRD with validation schema [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:14:08.095: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:30:23.352: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename crd-publish-openapi
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-4157
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-4073
 STEP: Waiting for a default service account to be provisioned in namespace
 [It] works for CRD with validation schema [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:14:08.258: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:30:23.530: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: client-side validation (kubectl create and apply) allows request with known and required properties
-Aug 30 17:14:11.113: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-4157 create -f -'
-Aug 30 17:14:11.548: INFO: stderr: ""
-Aug 30 17:14:11.548: INFO: stdout: "e2e-test-crd-publish-openapi-2881-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n"
-Aug 30 17:14:11.548: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-4157 delete e2e-test-crd-publish-openapi-2881-crds test-foo'
-Aug 30 17:14:11.623: INFO: stderr: ""
-Aug 30 17:14:11.623: INFO: stdout: "e2e-test-crd-publish-openapi-2881-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n"
-Aug 30 17:14:11.623: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-4157 apply -f -'
-Aug 30 17:14:11.776: INFO: stderr: ""
-Aug 30 17:14:11.776: INFO: stdout: "e2e-test-crd-publish-openapi-2881-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n"
-Aug 30 17:14:11.777: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-4157 delete e2e-test-crd-publish-openapi-2881-crds test-foo'
-Aug 30 17:14:11.854: INFO: stderr: ""
-Aug 30 17:14:11.854: INFO: stdout: "e2e-test-crd-publish-openapi-2881-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n"
+Feb 12 10:30:26.398: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-4073 create -f -'
+Feb 12 10:30:26.867: INFO: stderr: ""
+Feb 12 10:30:26.867: INFO: stdout: "e2e-test-crd-publish-openapi-2484-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n"
+Feb 12 10:30:26.867: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-4073 delete e2e-test-crd-publish-openapi-2484-crds test-foo'
+Feb 12 10:30:26.961: INFO: stderr: ""
+Feb 12 10:30:26.961: INFO: stdout: "e2e-test-crd-publish-openapi-2484-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n"
+Feb 12 10:30:26.961: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-4073 apply -f -'
+Feb 12 10:30:27.126: INFO: stderr: ""
+Feb 12 10:30:27.126: INFO: stdout: "e2e-test-crd-publish-openapi-2484-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n"
+Feb 12 10:30:27.127: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-4073 delete e2e-test-crd-publish-openapi-2484-crds test-foo'
+Feb 12 10:30:27.226: INFO: stderr: ""
+Feb 12 10:30:27.226: INFO: stdout: "e2e-test-crd-publish-openapi-2484-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n"
 STEP: client-side validation (kubectl create and apply) rejects request with unknown properties when disallowed by the schema
-Aug 30 17:14:11.854: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-4157 create -f -'
-Aug 30 17:14:12.076: INFO: rc: 1
-Aug 30 17:14:12.076: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-4157 apply -f -'
-Aug 30 17:14:12.278: INFO: rc: 1
+Feb 12 10:30:27.227: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-4073 create -f -'
+Feb 12 10:30:27.365: INFO: rc: 1
+Feb 12 10:30:27.366: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-4073 apply -f -'
+Feb 12 10:30:27.587: INFO: rc: 1
 STEP: client-side validation (kubectl create and apply) rejects request without required properties
-Aug 30 17:14:12.278: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-4157 create -f -'
-Aug 30 17:14:12.408: INFO: rc: 1
-Aug 30 17:14:12.408: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-4157 apply -f -'
-Aug 30 17:14:12.620: INFO: rc: 1
+Feb 12 10:30:27.587: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-4073 create -f -'
+Feb 12 10:30:27.725: INFO: rc: 1
+Feb 12 10:30:27.725: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 --namespace=crd-publish-openapi-4073 apply -f -'
+Feb 12 10:30:27.877: INFO: rc: 1
 STEP: kubectl explain works to explain CR properties
-Aug 30 17:14:12.620: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 explain e2e-test-crd-publish-openapi-2881-crds'
-Aug 30 17:14:12.753: INFO: stderr: ""
-Aug 30 17:14:12.753: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-2881-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nDESCRIPTION:\n     Foo CRD for Testing\n\nFIELDS:\n   apiVersion\t\n     APIVersion defines the versioned schema of this representation of an\n     object. Servers should convert recognized schemas to the latest internal\n     value, and may reject unrecognized values. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n   kind\t\n     Kind is a string value representing the REST resource this object\n     represents. Servers may infer this from the endpoint the client submits\n     requests to. Cannot be updated. In CamelCase. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n   metadata\t\n     Standard object's metadata. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   spec\t\n     Specification of Foo\n\n   status\t\n     Status of Foo\n\n"
+Feb 12 10:30:27.877: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 explain e2e-test-crd-publish-openapi-2484-crds'
+Feb 12 10:30:28.025: INFO: stderr: ""
+Feb 12 10:30:28.025: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-2484-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nDESCRIPTION:\n     Foo CRD for Testing\n\nFIELDS:\n   apiVersion\t\n     APIVersion defines the versioned schema of this representation of an\n     object. Servers should convert recognized schemas to the latest internal\n     value, and may reject unrecognized values. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n   kind\t\n     Kind is a string value representing the REST resource this object\n     represents. Servers may infer this from the endpoint the client submits\n     requests to. Cannot be updated. In CamelCase. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n   metadata\t\n     Standard object's metadata. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   spec\t\n     Specification of Foo\n\n   status\t\n     Status of Foo\n\n"
 STEP: kubectl explain works to explain CR properties recursively
-Aug 30 17:14:12.753: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 explain e2e-test-crd-publish-openapi-2881-crds.metadata'
-Aug 30 17:14:12.889: INFO: stderr: ""
-Aug 30 17:14:12.889: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-2881-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: metadata \n\nDESCRIPTION:\n     Standard object's metadata. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n     ObjectMeta is metadata that all persisted resources must have, which\n     includes all objects users must create.\n\nFIELDS:\n   annotations\t\n     Annotations is an unstructured key value map stored with a resource that\n     may be set by external tools to store and retrieve arbitrary metadata. They\n     are not queryable and should be preserved when modifying objects. More\n     info: http://kubernetes.io/docs/user-guide/annotations\n\n   clusterName\t\n     The name of the cluster which the object belongs to. This is used to\n     distinguish resources with same name and namespace in different clusters.\n     This field is not set anywhere right now and apiserver is going to ignore\n     it if set in create or update request.\n\n   creationTimestamp\t\n     CreationTimestamp is a timestamp representing the server time when this\n     object was created. It is not guaranteed to be set in happens-before order\n     across separate operations. Clients may not set this value. It is\n     represented in RFC3339 form and is in UTC. Populated by the system.\n     Read-only. Null for lists. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   deletionGracePeriodSeconds\t\n     Number of seconds allowed for this object to gracefully terminate before it\n     will be removed from the system. Only set when deletionTimestamp is also\n     set. May only be shortened. Read-only.\n\n   deletionTimestamp\t\n     DeletionTimestamp is RFC 3339 date and time at which this resource will be\n     deleted. This field is set by the server when a graceful deletion is\n     requested by the user, and is not directly settable by a client. The\n     resource is expected to be deleted (no longer visible from resource lists,\n     and not reachable by name) after the time in this field, once the\n     finalizers list is empty. As long as the finalizers list contains items,\n     deletion is blocked. Once the deletionTimestamp is set, this value may not\n     be unset or be set further into the future, although it may be shortened or\n     the resource may be deleted prior to this time. For example, a user may\n     request that a pod is deleted in 30 seconds. The Kubelet will react by\n     sending a graceful termination signal to the containers in the pod. After\n     that 30 seconds, the Kubelet will send a hard termination signal (SIGKILL)\n     to the container and after cleanup, remove the pod from the API. In the\n     presence of network partitions, this object may still exist after this\n     timestamp, until an administrator or automated process can determine the\n     resource is fully terminated. If not set, graceful deletion of the object\n     has not been requested. Populated by the system when a graceful deletion is\n     requested. Read-only. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   finalizers\t<[]string>\n     Must be empty before the object is deleted from the registry. Each entry is\n     an identifier for the responsible component that will remove the entry from\n     the list. If the deletionTimestamp of the object is non-nil, entries in\n     this list can only be removed. Finalizers may be processed and removed in\n     any order. Order is NOT enforced because it introduces significant risk of\n     stuck finalizers. finalizers is a shared field, any actor with permission\n     can reorder it. If the finalizer list is processed in order, then this can\n     lead to a situation in which the component responsible for the first\n     finalizer in the list is waiting for a signal (field value, external\n     system, or other) produced by a component responsible for a finalizer later\n     in the list, resulting in a deadlock. Without enforced ordering finalizers\n     are free to order amongst themselves and are not vulnerable to ordering\n     changes in the list.\n\n   generateName\t\n     GenerateName is an optional prefix, used by the server, to generate a\n     unique name ONLY IF the Name field has not been provided. If this field is\n     used, the name returned to the client will be different than the name\n     passed. This value will also be combined with a unique suffix. The provided\n     value has the same validation rules as the Name field, and may be truncated\n     by the length of the suffix required to make the value unique on the\n     server. If this field is specified and the generated name exists, the\n     server will NOT return a 409 - instead, it will either return 201 Created\n     or 500 with Reason ServerTimeout indicating a unique name could not be\n     found in the time allotted, and the client should retry (optionally after\n     the time indicated in the Retry-After header). Applied only if Name is not\n     specified. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency\n\n   generation\t\n     A sequence number representing a specific generation of the desired state.\n     Populated by the system. Read-only.\n\n   labels\t\n     Map of string keys and values that can be used to organize and categorize\n     (scope and select) objects. May match selectors of replication controllers\n     and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n   managedFields\t<[]Object>\n     ManagedFields maps workflow-id and version to the set of fields that are\n     managed by that workflow. This is mostly for internal housekeeping, and\n     users typically shouldn't need to set or understand this field. A workflow\n     can be the user's name, a controller's name, or the name of a specific\n     apply path like \"ci-cd\". The set of fields is always in the version that\n     the workflow used when modifying the object.\n\n   name\t\n     Name must be unique within a namespace. Is required when creating\n     resources, although some resources may allow a client to request the\n     generation of an appropriate name automatically. Name is primarily intended\n     for creation idempotence and configuration definition. Cannot be updated.\n     More info: http://kubernetes.io/docs/user-guide/identifiers#names\n\n   namespace\t\n     Namespace defines the space within each name must be unique. An empty\n     namespace is equivalent to the \"default\" namespace, but \"default\" is the\n     canonical representation. Not all objects are required to be scoped to a\n     namespace - the value of this field for those objects will be empty. Must\n     be a DNS_LABEL. Cannot be updated. More info:\n     http://kubernetes.io/docs/user-guide/namespaces\n\n   ownerReferences\t<[]Object>\n     List of objects depended by this object. If ALL objects in the list have\n     been deleted, this object will be garbage collected. If this object is\n     managed by a controller, then an entry in this list will point to this\n     controller, with the controller field set to true. There cannot be more\n     than one managing controller.\n\n   resourceVersion\t\n     An opaque value that represents the internal version of this object that\n     can be used by clients to determine when objects have changed. May be used\n     for optimistic concurrency, change detection, and the watch operation on a\n     resource or set of resources. Clients must treat these values as opaque and\n     passed unmodified back to the server. They may only be valid for a\n     particular resource or set of resources. Populated by the system.\n     Read-only. Value must be treated as opaque by clients and . More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency\n\n   selfLink\t\n     SelfLink is a URL representing this object. Populated by the system.\n     Read-only. DEPRECATED Kubernetes will stop propagating this field in 1.20\n     release and the field is planned to be removed in 1.21 release.\n\n   uid\t\n     UID is the unique in time and space value for this object. It is typically\n     generated by the server on successful creation of a resource and is not\n     allowed to change on PUT operations. Populated by the system. Read-only.\n     More info: http://kubernetes.io/docs/user-guide/identifiers#uids\n\n"
-Aug 30 17:14:12.889: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 explain e2e-test-crd-publish-openapi-2881-crds.spec'
-Aug 30 17:14:13.100: INFO: stderr: ""
-Aug 30 17:14:13.100: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-2881-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: spec \n\nDESCRIPTION:\n     Specification of Foo\n\nFIELDS:\n   bars\t<[]Object>\n     List of Bars and their specs.\n\n"
-Aug 30 17:14:13.100: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 explain e2e-test-crd-publish-openapi-2881-crds.spec.bars'
-Aug 30 17:14:13.232: INFO: stderr: ""
-Aug 30 17:14:13.232: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-2881-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: bars <[]Object>\n\nDESCRIPTION:\n     List of Bars and their specs.\n\nFIELDS:\n   age\t\n     Age of Bar.\n\n   bazs\t<[]string>\n     List of Bazs.\n\n   name\t -required-\n     Name of Bar.\n\n"
-STEP: kubectl explain works to return error when explain is called on property that doesn't exist
-Aug 30 17:14:13.232: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 explain e2e-test-crd-publish-openapi-2881-crds.spec.bars2'
-Aug 30 17:14:13.467: INFO: rc: 1
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:14:16.801: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-4157" for this suite.
-
-• [SLOW TEST:8.725 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for CRD with validation schema [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance]","total":280,"completed":157,"skipped":2580,"failed":0}
-SS
-------------------------------
-[k8s.io] Pods 
-  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:14:16.820: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-4238
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
-[It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-STEP: submitting the pod to kubernetes
-STEP: verifying the pod is in kubernetes
-STEP: updating the pod
-Aug 30 17:14:19.534: INFO: Successfully updated pod "pod-update-activedeadlineseconds-226b01c7-5b86-4a3b-9a1b-d77eb8642a3c"
-Aug 30 17:14:19.534: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-226b01c7-5b86-4a3b-9a1b-d77eb8642a3c" in namespace "pods-4238" to be "terminated due to deadline exceeded"
-Aug 30 17:14:19.538: INFO: Pod "pod-update-activedeadlineseconds-226b01c7-5b86-4a3b-9a1b-d77eb8642a3c": Phase="Running", Reason="", readiness=true. Elapsed: 4.086459ms
-Aug 30 17:14:21.544: INFO: Pod "pod-update-activedeadlineseconds-226b01c7-5b86-4a3b-9a1b-d77eb8642a3c": Phase="Running", Reason="", readiness=true. Elapsed: 2.010589436s
-Aug 30 17:14:23.550: INFO: Pod "pod-update-activedeadlineseconds-226b01c7-5b86-4a3b-9a1b-d77eb8642a3c": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 4.01605853s
-Aug 30 17:14:23.550: INFO: Pod "pod-update-activedeadlineseconds-226b01c7-5b86-4a3b-9a1b-d77eb8642a3c" satisfied condition "terminated due to deadline exceeded"
-[AfterEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:14:23.550: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-4238" for this suite.
-
-• [SLOW TEST:6.745 seconds]
-[k8s.io] Pods
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [k8s.io] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]","total":280,"completed":158,"skipped":2582,"failed":0}
-SSSSSSS
-------------------------------
-[sig-network] DNS 
-  should provide DNS for pods for Subdomain [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:14:23.565: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename dns
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-3557
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide DNS for pods for Subdomain [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a test headless service
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-3557.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service-2.dns-3557.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-3557.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service-2.dns-3557.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3557.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
-
-STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-3557.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service-2.dns-3557.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-3557.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service-2.dns-3557.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-3557.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
-
-STEP: creating a pod to probe DNS
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Aug 30 17:14:25.918: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local from pod dns-3557/dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69: the server could not find the requested resource (get pods dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69)
-Aug 30 17:14:25.965: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local from pod dns-3557/dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69: the server could not find the requested resource (get pods dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69)
-Aug 30 17:14:25.975: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-3557.svc.cluster.local from pod dns-3557/dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69: the server could not find the requested resource (get pods dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69)
-Aug 30 17:14:25.985: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-3557.svc.cluster.local from pod dns-3557/dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69: the server could not find the requested resource (get pods dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69)
-Aug 30 17:14:26.164: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local from pod dns-3557/dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69: the server could not find the requested resource (get pods dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69)
-Aug 30 17:14:26.173: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local from pod dns-3557/dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69: the server could not find the requested resource (get pods dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69)
-Aug 30 17:14:26.182: INFO: Unable to read jessie_udp@dns-test-service-2.dns-3557.svc.cluster.local from pod dns-3557/dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69: the server could not find the requested resource (get pods dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69)
-Aug 30 17:14:26.191: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-3557.svc.cluster.local from pod dns-3557/dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69: the server could not find the requested resource (get pods dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69)
-Aug 30 17:14:26.287: INFO: Lookups using dns-3557/dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local wheezy_udp@dns-test-service-2.dns-3557.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-3557.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-3557.svc.cluster.local jessie_udp@dns-test-service-2.dns-3557.svc.cluster.local jessie_tcp@dns-test-service-2.dns-3557.svc.cluster.local]
-
-Aug 30 17:14:32.256: INFO: DNS probes using dns-3557/dns-test-6bf33241-fa47-457f-98e2-2137b84d0b69 succeeded
-
-STEP: deleting the pod
-STEP: deleting the test headless service
-[AfterEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:14:32.298: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-3557" for this suite.
-
-• [SLOW TEST:8.754 seconds]
-[sig-network] DNS
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should provide DNS for pods for Subdomain [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Subdomain [Conformance]","total":280,"completed":159,"skipped":2589,"failed":0}
-SS
+Feb 12 10:30:28.026: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 explain e2e-test-crd-publish-openapi-2484-crds.metadata'
+Feb 12 10:30:28.188: INFO: stderr: ""
+Feb 12 10:30:28.188: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-2484-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: metadata \n\nDESCRIPTION:\n     Standard object's metadata. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n     ObjectMeta is metadata that all persisted resources must have, which\n     includes all objects users must create.\n\nFIELDS:\n   annotations\t\n     Annotations is an unstructured key value map stored with a resource that\n     may be set by external tools to store and retrieve arbitrary metadata. They\n     are not queryable and should be preserved when modifying objects. More\n     info: http://kubernetes.io/docs/user-guide/annotations\n\n   clusterName\t\n     The name of the cluster which the object belongs to. This is used to\n     distinguish resources with same name and namespace in different clusters.\n     This field is not set anywhere right now and apiserver is going to ignore\n     it if set in create or update request.\n\n   creationTimestamp\t\n     CreationTimestamp is a timestamp representing the server time when this\n     object was created. It is not guaranteed to be set in happens-before order\n     across separate operations. Clients may not set this value. It is\n     represented in RFC3339 form and is in UTC. Populated by the system.\n     Read-only. Null for lists. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   deletionGracePeriodSeconds\t\n     Number of seconds allowed for this object to gracefully terminate before it\n     will be removed from the system. Only set when deletionTimestamp is also\n     set. May only be shortened. Read-only.\n\n   deletionTimestamp\t\n     DeletionTimestamp is RFC 3339 date and time at which this resource will be\n     deleted. This field is set by the server when a graceful deletion is\n     requested by the user, and is not directly settable by a client. The\n     resource is expected to be deleted (no longer visible from resource lists,\n     and not reachable by name) after the time in this field, once the\n     finalizers list is empty. As long as the finalizers list contains items,\n     deletion is blocked. Once the deletionTimestamp is set, this value may not\n     be unset or be set further into the future, although it may be shortened or\n     the resource may be deleted prior to this time. For example, a user may\n     request that a pod is deleted in 30 seconds. The Kubelet will react by\n     sending a graceful termination signal to the containers in the pod. After\n     that 30 seconds, the Kubelet will send a hard termination signal (SIGKILL)\n     to the container and after cleanup, remove the pod from the API. In the\n     presence of network partitions, this object may still exist after this\n     timestamp, until an administrator or automated process can determine the\n     resource is fully terminated. If not set, graceful deletion of the object\n     has not been requested. Populated by the system when a graceful deletion is\n     requested. Read-only. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n   finalizers\t<[]string>\n     Must be empty before the object is deleted from the registry. Each entry is\n     an identifier for the responsible component that will remove the entry from\n     the list. If the deletionTimestamp of the object is non-nil, entries in\n     this list can only be removed. Finalizers may be processed and removed in\n     any order. Order is NOT enforced because it introduces significant risk of\n     stuck finalizers. finalizers is a shared field, any actor with permission\n     can reorder it. If the finalizer list is processed in order, then this can\n     lead to a situation in which the component responsible for the first\n     finalizer in the list is waiting for a signal (field value, external\n     system, or other) produced by a component responsible for a finalizer later\n     in the list, resulting in a deadlock. Without enforced ordering finalizers\n     are free to order amongst themselves and are not vulnerable to ordering\n     changes in the list.\n\n   generateName\t\n     GenerateName is an optional prefix, used by the server, to generate a\n     unique name ONLY IF the Name field has not been provided. If this field is\n     used, the name returned to the client will be different than the name\n     passed. This value will also be combined with a unique suffix. The provided\n     value has the same validation rules as the Name field, and may be truncated\n     by the length of the suffix required to make the value unique on the\n     server. If this field is specified and the generated name exists, the\n     server will NOT return a 409 - instead, it will either return 201 Created\n     or 500 with Reason ServerTimeout indicating a unique name could not be\n     found in the time allotted, and the client should retry (optionally after\n     the time indicated in the Retry-After header). Applied only if Name is not\n     specified. More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency\n\n   generation\t\n     A sequence number representing a specific generation of the desired state.\n     Populated by the system. Read-only.\n\n   labels\t\n     Map of string keys and values that can be used to organize and categorize\n     (scope and select) objects. May match selectors of replication controllers\n     and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n   managedFields\t<[]Object>\n     ManagedFields maps workflow-id and version to the set of fields that are\n     managed by that workflow. This is mostly for internal housekeeping, and\n     users typically shouldn't need to set or understand this field. A workflow\n     can be the user's name, a controller's name, or the name of a specific\n     apply path like \"ci-cd\". The set of fields is always in the version that\n     the workflow used when modifying the object.\n\n   name\t\n     Name must be unique within a namespace. Is required when creating\n     resources, although some resources may allow a client to request the\n     generation of an appropriate name automatically. Name is primarily intended\n     for creation idempotence and configuration definition. Cannot be updated.\n     More info: http://kubernetes.io/docs/user-guide/identifiers#names\n\n   namespace\t\n     Namespace defines the space within each name must be unique. An empty\n     namespace is equivalent to the \"default\" namespace, but \"default\" is the\n     canonical representation. Not all objects are required to be scoped to a\n     namespace - the value of this field for those objects will be empty. Must\n     be a DNS_LABEL. Cannot be updated. More info:\n     http://kubernetes.io/docs/user-guide/namespaces\n\n   ownerReferences\t<[]Object>\n     List of objects depended by this object. If ALL objects in the list have\n     been deleted, this object will be garbage collected. If this object is\n     managed by a controller, then an entry in this list will point to this\n     controller, with the controller field set to true. There cannot be more\n     than one managing controller.\n\n   resourceVersion\t\n     An opaque value that represents the internal version of this object that\n     can be used by clients to determine when objects have changed. May be used\n     for optimistic concurrency, change detection, and the watch operation on a\n     resource or set of resources. Clients must treat these values as opaque and\n     passed unmodified back to the server. They may only be valid for a\n     particular resource or set of resources. Populated by the system.\n     Read-only. Value must be treated as opaque by clients and . More info:\n     https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency\n\n   selfLink\t\n     SelfLink is a URL representing this object. Populated by the system.\n     Read-only. DEPRECATED Kubernetes will stop propagating this field in 1.20\n     release and the field is planned to be removed in 1.21 release.\n\n   uid\t\n     UID is the unique in time and space value for this object. It is typically\n     generated by the server on successful creation of a resource and is not\n     allowed to change on PUT operations. Populated by the system. Read-only.\n     More info: http://kubernetes.io/docs/user-guide/identifiers#uids\n\n"
+Feb 12 10:30:28.188: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 explain e2e-test-crd-publish-openapi-2484-crds.spec'
+Feb 12 10:30:28.469: INFO: stderr: ""
+Feb 12 10:30:28.469: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-2484-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: spec \n\nDESCRIPTION:\n     Specification of Foo\n\nFIELDS:\n   bars\t<[]Object>\n     List of Bars and their specs.\n\n"
+Feb 12 10:30:28.469: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 explain e2e-test-crd-publish-openapi-2484-crds.spec.bars'
+Feb 12 10:30:28.701: INFO: stderr: ""
+Feb 12 10:30:28.701: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-2484-crd\nVERSION:  crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: bars <[]Object>\n\nDESCRIPTION:\n     List of Bars and their specs.\n\nFIELDS:\n   age\t\n     Age of Bar.\n\n   bazs\t<[]string>\n     List of Bazs.\n\n   name\t -required-\n     Name of Bar.\n\n"
+STEP: kubectl explain works to return error when explain is called on property that doesn't exist
+Feb 12 10:30:28.701: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 explain e2e-test-crd-publish-openapi-2484-crds.spec.bars2'
+Feb 12 10:30:28.930: INFO: rc: 1
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:31.790: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-4073" for this suite.
+
+• [SLOW TEST:8.460 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  works for CRD with validation schema [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] Kubelet when scheduling a busybox command in a pod 
-  should print the output to logs [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance]","total":280,"completed":136,"skipped":2071,"failed":0}
+[sig-storage] EmptyDir volumes 
+  volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:14:32.320: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubelet-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-5652
+Feb 12 10:30:31.815: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-9756
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
-[It] should print the output to logs [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:14:36.539: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubelet-test-5652" for this suite.
-•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance]","total":280,"completed":160,"skipped":2591,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSS
+[It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir volume type on tmpfs
+Feb 12 10:30:31.993: INFO: Waiting up to 5m0s for pod "pod-e803ce46-a0f4-443e-840a-503f01b6006d" in namespace "emptydir-9756" to be "success or failure"
+Feb 12 10:30:32.000: INFO: Pod "pod-e803ce46-a0f4-443e-840a-503f01b6006d": Phase="Pending", Reason="", readiness=false. Elapsed: 6.800728ms
+Feb 12 10:30:34.008: INFO: Pod "pod-e803ce46-a0f4-443e-840a-503f01b6006d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014157635s
+STEP: Saw pod success
+Feb 12 10:30:34.008: INFO: Pod "pod-e803ce46-a0f4-443e-840a-503f01b6006d" satisfied condition "success or failure"
+Feb 12 10:30:34.014: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-e803ce46-a0f4-443e-840a-503f01b6006d container test-container: 
+STEP: delete the pod
+Feb 12 10:30:34.057: INFO: Waiting for pod pod-e803ce46-a0f4-443e-840a-503f01b6006d to disappear
+Feb 12 10:30:34.063: INFO: Pod pod-e803ce46-a0f4-443e-840a-503f01b6006d no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:34.063: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-9756" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":137,"skipped":2071,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
-  should execute poststart http hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should honor timeout [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:14:36.557: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-lifecycle-hook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-5022
+Feb 12 10:30:34.087: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-7680
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] when create a pod with lifecycle hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
-STEP: create the container to handle the HTTPGet hook request.
-[It] should execute poststart http hook properly [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the pod with lifecycle hook
-STEP: check poststart hook
-STEP: delete the pod with lifecycle hook
-Aug 30 17:14:42.830: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Aug 30 17:14:42.835: INFO: Pod pod-with-poststart-http-hook still exists
-Aug 30 17:14:44.835: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Aug 30 17:14:44.841: INFO: Pod pod-with-poststart-http-hook still exists
-Aug 30 17:14:46.835: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Aug 30 17:14:46.841: INFO: Pod pod-with-poststart-http-hook still exists
-Aug 30 17:14:48.835: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Aug 30 17:14:48.841: INFO: Pod pod-with-poststart-http-hook still exists
-Aug 30 17:14:50.835: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Aug 30 17:14:50.845: INFO: Pod pod-with-poststart-http-hook still exists
-Aug 30 17:14:52.835: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Aug 30 17:14:52.841: INFO: Pod pod-with-poststart-http-hook still exists
-Aug 30 17:14:54.835: INFO: Waiting for pod pod-with-poststart-http-hook to disappear
-Aug 30 17:14:54.841: INFO: Pod pod-with-poststart-http-hook no longer exists
-[AfterEach] [k8s.io] Container Lifecycle Hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:14:54.841: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-lifecycle-hook-5022" for this suite.
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Feb 12 10:30:34.580: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Feb 12 10:30:37.674: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should honor timeout [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Setting timeout (1s) shorter than webhook latency (5s)
+STEP: Registering slow webhook via the AdmissionRegistration API
+STEP: Request fails when timeout (1s) is shorter than slow webhook latency (5s)
+STEP: Having no error when timeout is shorter than webhook latency and failure policy is ignore
+STEP: Registering slow webhook via the AdmissionRegistration API
+STEP: Having no error when timeout is longer than webhook latency
+STEP: Registering slow webhook via the AdmissionRegistration API
+Feb 12 10:30:39.948: INFO: Waiting for webhook configuration to be ready...
+STEP: Having no error when timeout is empty (defaulted to 10s in v1)
+STEP: Registering slow webhook via the AdmissionRegistration API
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:50.343: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-7680" for this suite.
+STEP: Destroying namespace "webhook-7680-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:18.300 seconds]
-[k8s.io] Container Lifecycle Hook
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  when create a pod with lifecycle hook
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
-    should execute poststart http hook properly [NodeConformance] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:16.371 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should honor timeout [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance]","total":280,"completed":161,"skipped":2614,"failed":0}
-SSSSSSSS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance]","total":280,"completed":138,"skipped":2098,"failed":0}
+SSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide container's memory limit [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] EmptyDir volumes 
+  should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:14:54.857: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7422
+Feb 12 10:30:50.458: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-7930
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should provide container's memory limit [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 17:14:55.035: INFO: Waiting up to 5m0s for pod "downwardapi-volume-49773cf5-70e4-4b5c-9792-e05720c5c08f" in namespace "projected-7422" to be "success or failure"
-Aug 30 17:14:55.039: INFO: Pod "downwardapi-volume-49773cf5-70e4-4b5c-9792-e05720c5c08f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.095563ms
-Aug 30 17:14:57.046: INFO: Pod "downwardapi-volume-49773cf5-70e4-4b5c-9792-e05720c5c08f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.01154145s
+[It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir 0644 on node default medium
+Feb 12 10:30:50.647: INFO: Waiting up to 5m0s for pod "pod-09bf1fa7-cd88-40a6-856d-26792dbc0360" in namespace "emptydir-7930" to be "success or failure"
+Feb 12 10:30:50.655: INFO: Pod "pod-09bf1fa7-cd88-40a6-856d-26792dbc0360": Phase="Pending", Reason="", readiness=false. Elapsed: 8.732306ms
+Feb 12 10:30:52.661: INFO: Pod "pod-09bf1fa7-cd88-40a6-856d-26792dbc0360": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014097357s
 STEP: Saw pod success
-Aug 30 17:14:57.047: INFO: Pod "downwardapi-volume-49773cf5-70e4-4b5c-9792-e05720c5c08f" satisfied condition "success or failure"
-Aug 30 17:14:57.052: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-49773cf5-70e4-4b5c-9792-e05720c5c08f container client-container: 
+Feb 12 10:30:52.661: INFO: Pod "pod-09bf1fa7-cd88-40a6-856d-26792dbc0360" satisfied condition "success or failure"
+Feb 12 10:30:52.667: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-09bf1fa7-cd88-40a6-856d-26792dbc0360 container test-container: 
 STEP: delete the pod
-Aug 30 17:14:57.122: INFO: Waiting for pod downwardapi-volume-49773cf5-70e4-4b5c-9792-e05720c5c08f to disappear
-Aug 30 17:14:57.126: INFO: Pod downwardapi-volume-49773cf5-70e4-4b5c-9792-e05720c5c08f no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:14:57.126: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-7422" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance]","total":280,"completed":162,"skipped":2622,"failed":0}
-SSSSSSSSSSSS
+Feb 12 10:30:52.704: INFO: Waiting for pod pod-09bf1fa7-cd88-40a6-856d-26792dbc0360 to disappear
+Feb 12 10:30:52.709: INFO: Pod pod-09bf1fa7-cd88-40a6-856d-26792dbc0360 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:52.710: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-7930" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":139,"skipped":2101,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Container Runtime blackbox test on terminated container 
-  should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] Secrets 
+  should fail to create secret due to empty secret key [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:14:57.143: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-runtime
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-4092
+Feb 12 10:30:52.731: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-5265
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the container
-STEP: wait for the container to reach Failed
-STEP: get the container status
-STEP: the container should be terminated
-STEP: the termination message should be set
-Aug 30 17:14:59.350: INFO: Expected: &{DONE} to match Container's Termination Message: DONE --
-STEP: delete the container
-[AfterEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:14:59.373: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-runtime-4092" for this suite.
-•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":280,"completed":163,"skipped":2634,"failed":0}
-SSSSSSSSSSSSSSSS
+[It] should fail to create secret due to empty secret key [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating projection with secret that has name secret-emptykey-test-3cb95ba0-c9ac-4f18-b903-b837bdfa81fe
+[AfterEach] [sig-api-machinery] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:52.896: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-5265" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Secrets should fail to create secret due to empty secret key [Conformance]","total":280,"completed":140,"skipped":2160,"failed":0}
+SSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected secret 
-  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:14:59.388: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-4736
+Feb 12 10:30:52.911: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-5535
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name projected-secret-test-eb3ab438-75b4-4390-8934-11f10a0e5841
-STEP: Creating a pod to test consume secrets
-Aug 30 17:14:59.579: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-3c49fdb1-6205-4e8e-99d4-db4f7dd0c8e3" in namespace "projected-4736" to be "success or failure"
-Aug 30 17:14:59.587: INFO: Pod "pod-projected-secrets-3c49fdb1-6205-4e8e-99d4-db4f7dd0c8e3": Phase="Pending", Reason="", readiness=false. Elapsed: 8.196416ms
-Aug 30 17:15:01.593: INFO: Pod "pod-projected-secrets-3c49fdb1-6205-4e8e-99d4-db4f7dd0c8e3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014686588s
+[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name configmap-test-volume-16949945-f378-438c-8db7-a22319e3a1fb
+STEP: Creating a pod to test consume configMaps
+Feb 12 10:30:53.104: INFO: Waiting up to 5m0s for pod "pod-configmaps-53f544f5-cccc-4f4d-914e-e8312cc9ab8e" in namespace "configmap-5535" to be "success or failure"
+Feb 12 10:30:53.119: INFO: Pod "pod-configmaps-53f544f5-cccc-4f4d-914e-e8312cc9ab8e": Phase="Pending", Reason="", readiness=false. Elapsed: 14.551357ms
+Feb 12 10:30:55.125: INFO: Pod "pod-configmaps-53f544f5-cccc-4f4d-914e-e8312cc9ab8e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.020905916s
 STEP: Saw pod success
-Aug 30 17:15:01.594: INFO: Pod "pod-projected-secrets-3c49fdb1-6205-4e8e-99d4-db4f7dd0c8e3" satisfied condition "success or failure"
-Aug 30 17:15:01.599: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-projected-secrets-3c49fdb1-6205-4e8e-99d4-db4f7dd0c8e3 container secret-volume-test: 
+Feb 12 10:30:55.125: INFO: Pod "pod-configmaps-53f544f5-cccc-4f4d-914e-e8312cc9ab8e" satisfied condition "success or failure"
+Feb 12 10:30:55.131: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-configmaps-53f544f5-cccc-4f4d-914e-e8312cc9ab8e container configmap-volume-test: 
 STEP: delete the pod
-Aug 30 17:15:01.666: INFO: Waiting for pod pod-projected-secrets-3c49fdb1-6205-4e8e-99d4-db4f7dd0c8e3 to disappear
-Aug 30 17:15:01.671: INFO: Pod pod-projected-secrets-3c49fdb1-6205-4e8e-99d4-db4f7dd0c8e3 no longer exists
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:15:01.671: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-4736" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":280,"completed":164,"skipped":2650,"failed":0}
-SSSSSSSS
+Feb 12 10:30:55.174: INFO: Waiting for pod pod-configmaps-53f544f5-cccc-4f4d-914e-e8312cc9ab8e to disappear
+Feb 12 10:30:55.179: INFO: Pod pod-configmaps-53f544f5-cccc-4f4d-914e-e8312cc9ab8e no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:55.179: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-5535" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":141,"skipped":2180,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Kubelet when scheduling a busybox command that always fails in a pod 
-  should be possible to delete [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-cli] Kubectl client Kubectl version 
+  should check is all data is printed  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:15:01.686: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubelet-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-3737
+Feb 12 10:30:55.209: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7169
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
-[BeforeEach] when scheduling a busybox command that always fails in a pod
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:81
-[It] should be possible to delete [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[AfterEach] [k8s.io] Kubelet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:15:01.876: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubelet-test-3737" for this suite.
-•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance]","total":280,"completed":165,"skipped":2658,"failed":0}
-SSSSSS
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[It] should check is all data is printed  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:30:55.387: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 version'
+Feb 12 10:30:55.460: INFO: stderr: ""
+Feb 12 10:30:55.460: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"17\", GitVersion:\"v1.17.16\", GitCommit:\"d88fadbd65c5e8bde22630d251766a634c7613b0\", GitTreeState:\"clean\", BuildDate:\"2020-12-18T12:15:37Z\", GoVersion:\"go1.13.15\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"17\", GitVersion:\"v1.17.16\", GitCommit:\"d88fadbd65c5e8bde22630d251766a634c7613b0\", GitTreeState:\"clean\", BuildDate:\"2020-12-18T12:07:12Z\", GoVersion:\"go1.13.15\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:55.460: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-7169" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl version should check is all data is printed  [Conformance]","total":280,"completed":142,"skipped":2276,"failed":0}
+SSSSSSSSSSSS
 ------------------------------
-[sig-apps] Deployment 
-  deployment should support proportional scaling [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] ResourceQuota 
+  should be able to update and delete ResourceQuota. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:15:01.889: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename deployment
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-6481
+Feb 12 10:30:55.476: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename resourcequota
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-4066
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
-[It] deployment should support proportional scaling [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:15:02.059: INFO: Creating deployment "webserver-deployment"
-Aug 30 17:15:02.068: INFO: Waiting for observed generation 1
-Aug 30 17:15:04.081: INFO: Waiting for all required pods to come up
-Aug 30 17:15:04.088: INFO: Pod name httpd: Found 10 pods out of 10
-STEP: ensuring each pod is running
-Aug 30 17:15:06.108: INFO: Waiting for deployment "webserver-deployment" to complete
-Aug 30 17:15:06.119: INFO: Updating deployment "webserver-deployment" with a non-existent image
-Aug 30 17:15:06.133: INFO: Updating deployment webserver-deployment
-Aug 30 17:15:06.133: INFO: Waiting for observed generation 2
-Aug 30 17:15:08.143: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8
-Aug 30 17:15:08.148: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8
-Aug 30 17:15:08.152: INFO: Waiting for the first rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas
-Aug 30 17:15:08.167: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0
-Aug 30 17:15:08.167: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5
-Aug 30 17:15:08.172: INFO: Waiting for the second rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas
-Aug 30 17:15:08.182: INFO: Verifying that deployment "webserver-deployment" has minimum required number of available replicas
-Aug 30 17:15:08.182: INFO: Scaling up the deployment "webserver-deployment" from 10 to 30
-Aug 30 17:15:08.194: INFO: Updating deployment webserver-deployment
-Aug 30 17:15:08.194: INFO: Waiting for the replicasets of deployment "webserver-deployment" to have desired number of replicas
-Aug 30 17:15:08.209: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20
-Aug 30 17:15:08.219: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
-Aug 30 17:15:10.245: INFO: Deployment "webserver-deployment":
-&Deployment{ObjectMeta:{webserver-deployment  deployment-6481 /apis/apps/v1/namespaces/deployment-6481/deployments/webserver-deployment 680bcbb3-2940-48c8-b5d7-6eb06ee84d2a 19541 3 2020-08-30 17:15:02 +0000 UTC   map[name:httpd] map[deployment.kubernetes.io/revision:2] [] []  []},Spec:DeploymentSpec{Replicas:*30,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:httpd] map[] [] []  []} {[] [] [{httpd webserver:404 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc00486bb18  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:33,UpdatedReplicas:13,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2020-08-30 17:15:08 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "webserver-deployment-c7997dcc8" is progressing.,LastUpdateTime:2020-08-30 17:15:08 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,},},ReadyReplicas:8,CollisionCount:nil,},}
-
-Aug 30 17:15:10.251: INFO: New ReplicaSet "webserver-deployment-c7997dcc8" of Deployment "webserver-deployment":
-&ReplicaSet{ObjectMeta:{webserver-deployment-c7997dcc8  deployment-6481 /apis/apps/v1/namespaces/deployment-6481/replicasets/webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 19519 3 2020-08-30 17:15:06 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment webserver-deployment 680bcbb3-2940-48c8-b5d7-6eb06ee84d2a 0xc004728497 0xc004728498}] []  []},Spec:ReplicaSetSpec{Replicas:*13,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: c7997dcc8,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [] []  []} {[] [] [{httpd webserver:404 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc004728518  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:13,FullyLabeledReplicas:13,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
-Aug 30 17:15:10.251: INFO: All old ReplicaSets of Deployment "webserver-deployment":
-Aug 30 17:15:10.251: INFO: &ReplicaSet{ObjectMeta:{webserver-deployment-595b5b9587  deployment-6481 /apis/apps/v1/namespaces/deployment-6481/replicasets/webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 19529 3 2020-08-30 17:15:02 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment webserver-deployment 680bcbb3-2940-48c8-b5d7-6eb06ee84d2a 0xc004728347 0xc004728348}] []  []},Spec:ReplicaSetSpec{Replicas:*20,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 595b5b9587,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc004728418  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:8,AvailableReplicas:8,Conditions:[]ReplicaSetCondition{},},}
-Aug 30 17:15:10.265: INFO: Pod "webserver-deployment-595b5b9587-2g2p9" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-2g2p9 webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-2g2p9 7bcb6784-a737-4312-a50e-ec5cb789b496 19339 0 2020-08-30 17:15:02 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.1.66/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc004728da7 0xc004728da8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:172.25.1.66,StartTime:2020-08-30 17:15:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-08-30 17:15:04 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://9c8424db3be642029a8b23da2eb37fab730a5f3cf21cbf4340de6a4656f9a560,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.66,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.265: INFO: Pod "webserver-deployment-595b5b9587-4vq75" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-4vq75 webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-4vq75 ed576391-a1e1-4733-9b35-3293e1eab4cb 19352 0 2020-08-30 17:15:02 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.0.153/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc004729030 0xc004729031}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:172.25.0.153,StartTime:2020-08-30 17:15:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-08-30 17:15:04 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://fa5e294c068129f3eeb713027e1bd4dd72b281b0369211f731f4226049ee7c0e,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.0.153,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.266: INFO: Pod "webserver-deployment-595b5b9587-5qkvb" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-5qkvb webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-5qkvb 342ad11b-59e9-484c-b9ce-c2d77b263a63 19576 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.1.70/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0047292b7 0xc0047292b8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.266: INFO: Pod "webserver-deployment-595b5b9587-6dd57" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-6dd57 webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-6dd57 59f5ba7c-6537-4f78-862e-e08717b1b197 19355 0 2020-08-30 17:15:02 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.1.65/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0047294a7 0xc0047294a8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:172.25.1.65,StartTime:2020-08-30 17:15:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-08-30 17:15:03 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://f4bc01eafdf7c2dd8cd3fca5c3b95de547fed3a27c6057baf9b300efb875e057,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.65,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.266: INFO: Pod "webserver-deployment-595b5b9587-6rvh6" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-6rvh6 webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-6rvh6 edf7dc56-d928-4ac5-a9e7-cd31aa9f822a 19587 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.0.158/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0047296d0 0xc0047296d1}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.267: INFO: Pod "webserver-deployment-595b5b9587-8qkjq" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-8qkjq webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-8qkjq e1ab339f-ea03-489e-8e55-3a1476198eef 19583 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0047299b7 0xc0047299b8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.267: INFO: Pod "webserver-deployment-595b5b9587-gqnzf" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-gqnzf webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-gqnzf a3a07b9c-ffc6-4b6a-bc56-e485ee093684 19584 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.1.71/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc004729c77 0xc004729c78}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.267: INFO: Pod "webserver-deployment-595b5b9587-jp6wd" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-jp6wd webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-jp6wd 2d4dfd70-98f2-4d06-978c-84530b6a6faf 19549 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc004729f07 0xc004729f08}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.268: INFO: Pod "webserver-deployment-595b5b9587-lkh5h" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-lkh5h webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-lkh5h 6071dcd6-14aa-47d4-bdad-0fbc421fc353 19524 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0046f0117 0xc0046f0118}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.269: INFO: Pod "webserver-deployment-595b5b9587-lrg65" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-lrg65 webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-lrg65 0e89445c-7b40-44a1-b2fe-74e93049b4aa 19546 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0046f02b0 0xc0046f02b1}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.270: INFO: Pod "webserver-deployment-595b5b9587-n5pds" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-n5pds webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-n5pds 1c2052f1-6343-464b-9dfc-1b8feadf9bb0 19304 0 2020-08-30 17:15:02 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.1.64/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0046f04a7 0xc0046f04a8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:172.25.1.64,StartTime:2020-08-30 17:15:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-08-30 17:15:03 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://d4a31967f40d671692ce7e803ab24b40ce5f1c23bb9455398d62c382d08b0cf6,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.64,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.270: INFO: Pod "webserver-deployment-595b5b9587-q9qz8" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-q9qz8 webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-q9qz8 9c9fc6ef-d3fa-4c57-9b1b-f81013588e0a 19296 0 2020-08-30 17:15:02 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.1.63/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0046f06c0 0xc0046f06c1}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:172.25.1.63,StartTime:2020-08-30 17:15:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-08-30 17:15:03 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://49a9df326d03568e2a7f93d96b7d1508fe84d42adc48e286b8e9392e6e2faf8d,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.63,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.270: INFO: Pod "webserver-deployment-595b5b9587-qd2hl" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-qd2hl webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-qd2hl bd542e9b-144f-431d-94b7-5f1cb14eeff2 19348 0 2020-08-30 17:15:02 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.0.150/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0046f0880 0xc0046f0881}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:172.25.0.150,StartTime:2020-08-30 17:15:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-08-30 17:15:03 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://9a26a92d98685efaab0b7cd93db1e9c0f20fc3cc074c474b0d2e797cb7a625a7,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.0.150,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.270: INFO: Pod "webserver-deployment-595b5b9587-rph88" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-rph88 webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-rph88 0d7401ac-6352-4eda-bdea-cbe7c2851f2f 19606 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.1.73/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0046f0a17 0xc0046f0a18}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.270: INFO: Pod "webserver-deployment-595b5b9587-svsp8" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-svsp8 webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-svsp8 2d103d51-e74e-4db5-a92b-8e086e96dc98 19611 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.0.160/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0046f0c47 0xc0046f0c48}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.271: INFO: Pod "webserver-deployment-595b5b9587-tmslx" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-tmslx webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-tmslx d8a8ab71-e7cb-4c45-865f-624dded4991d 19346 0 2020-08-30 17:15:02 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.0.152/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0046f0f37 0xc0046f0f38}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:172.25.0.152,StartTime:2020-08-30 17:15:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-08-30 17:15:04 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://fe56b2fa5b49b869818a2abeb39b7c41d71ae26f206572cc1e2b7f79fe8a72cc,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.0.152,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.271: INFO: Pod "webserver-deployment-595b5b9587-vw4rc" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-vw4rc webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-vw4rc 668aa491-cd19-4d23-8a84-6051a0c3aca3 19578 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.0.157/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0046f11c7 0xc0046f11c8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.271: INFO: Pod "webserver-deployment-595b5b9587-vz75q" is available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-vz75q webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-vz75q 0e5b59a0-526d-42a5-aa8b-de26d2495886 19335 0 2020-08-30 17:15:02 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.1.67/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0046f1387 0xc0046f1388}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:02 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:172.25.1.67,StartTime:2020-08-30 17:15:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-08-30 17:15:04 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://ea2c33b947cabcaab164130f56f497797b9f4d9b08f5fda50957b3ccd962e540,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.67,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.271: INFO: Pod "webserver-deployment-595b5b9587-xvhrk" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-xvhrk webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-xvhrk fad69d70-8ba4-4338-b02f-e1148b5fe6b2 19599 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0046f15f0 0xc0046f15f1}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.271: INFO: Pod "webserver-deployment-595b5b9587-zdbf4" is not available:
-&Pod{ObjectMeta:{webserver-deployment-595b5b9587-zdbf4 webserver-deployment-595b5b9587- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-595b5b9587-zdbf4 0e8491c9-9f70-4764-9048-1604c7f1e3ce 19581 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 52d387b8-02db-40c1-a5cd-f22a2165744c 0xc0046f1817 0xc0046f1818}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.272: INFO: Pod "webserver-deployment-c7997dcc8-4hkzf" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-4hkzf webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-4hkzf 5c866929-b11c-41f9-b994-5c8cf44540b8 19453 0 2020-08-30 17:15:06 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.0.155/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046f19e7 0xc0046f19e8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 17:15:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.272: INFO: Pod "webserver-deployment-c7997dcc8-5bbk7" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-5bbk7 webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-5bbk7 4b3c9258-4351-4044-b7b4-13926db068f7 19594 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.1.72/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046f1c07 0xc0046f1c08}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.272: INFO: Pod "webserver-deployment-c7997dcc8-6gr78" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-6gr78 webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-6gr78 ccae3208-34c5-4bc9-ac0d-60a25daa5b9d 19604 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.0.159/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046f1e47 0xc0046f1e48}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.272: INFO: Pod "webserver-deployment-c7997dcc8-b8lth" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-b8lth webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-b8lth e15886eb-c8f8-4dc3-948a-7cdfd8ee9aaf 19456 0 2020-08-30 17:15:06 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.0.156/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046b8077 0xc0046b8078}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 17:15:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.272: INFO: Pod "webserver-deployment-c7997dcc8-fz996" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-fz996 webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-fz996 6417c552-c23d-41b5-901a-3b9f8191ed65 19548 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046b8277 0xc0046b8278}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.272: INFO: Pod "webserver-deployment-c7997dcc8-g2t2g" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-g2t2g webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-g2t2g 89ee1bb6-de37-4acc-ae09-9f57cf3eedf0 19608 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.1.74/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046b8477 0xc0046b8478}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.273: INFO: Pod "webserver-deployment-c7997dcc8-g75bp" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-g75bp webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-g75bp fd498bf2-b1e4-45fa-ad90-227155676e36 19612 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.1.75/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046b86c7 0xc0046b86c8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.273: INFO: Pod "webserver-deployment-c7997dcc8-gzqcn" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-gzqcn webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-gzqcn 70a264c0-3acc-4fcf-be57-da566256f8c5 19507 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046b8897 0xc0046b8898}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.273: INFO: Pod "webserver-deployment-c7997dcc8-kjtdz" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-kjtdz webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-kjtdz 185ee828-93ac-4db8-9b35-125336d52a3f 19450 0 2020-08-30 17:15:06 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.1.69/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046b8b47 0xc0046b8b48}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:,StartTime:2020-08-30 17:15:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.273: INFO: Pod "webserver-deployment-c7997dcc8-lbxl5" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-lbxl5 webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-lbxl5 01fb03d6-fde4-444c-80c8-439f917617d2 19449 0 2020-08-30 17:15:06 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.0.154/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046b8df7 0xc0046b8df8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 17:15:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.273: INFO: Pod "webserver-deployment-c7997dcc8-lfz5s" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-lfz5s webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-lfz5s f0f19c92-9f4b-416e-ba86-077e8028e0ec 19603 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046b8fc7 0xc0046b8fc8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:,StartTime:2020-08-30 17:15:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.274: INFO: Pod "webserver-deployment-c7997dcc8-p2fj7" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-p2fj7 webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-p2fj7 91b476e6-ac25-4032-be3b-487e12419e81 19528 0 2020-08-30 17:15:08 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046b9287 0xc0046b9288}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
-Aug 30 17:15:10.274: INFO: Pod "webserver-deployment-c7997dcc8-x9vzj" is not available:
-&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-x9vzj webserver-deployment-c7997dcc8- deployment-6481 /api/v1/namespaces/deployment-6481/pods/webserver-deployment-c7997dcc8-x9vzj 8efa2591-1b09-4bed-b08d-e3b1fc4a92bd 19615 0 2020-08-30 17:15:06 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.1.68/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 5d470b43-d410-4c14-808f-a19570286625 0xc0046b9430 0xc0046b9431}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-42hvn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-42hvn,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-42hvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-6rshr,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:15:06 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.153.64,PodIP:172.25.1.68,StartTime:2020-08-30 17:15:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = Error response from daemon: pull access denied for webserver, repository does not exist or may require 'docker login',},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.68,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
-[AfterEach] [sig-apps] Deployment
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:15:10.274: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "deployment-6481" for this suite.
-
-• [SLOW TEST:8.400 seconds]
-[sig-apps] Deployment
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  deployment should support proportional scaling [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[It] should be able to update and delete ResourceQuota. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a ResourceQuota
+STEP: Getting a ResourceQuota
+STEP: Updating a ResourceQuota
+STEP: Verifying a ResourceQuota was modified
+STEP: Deleting a ResourceQuota
+STEP: Verifying the deleted ResourceQuota
+[AfterEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:55.696: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-4066" for this suite.
+•{"msg":"PASSED [sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance]","total":280,"completed":143,"skipped":2288,"failed":0}
+SS
 ------------------------------
-{"msg":"PASSED [sig-apps] Deployment deployment should support proportional scaling [Conformance]","total":280,"completed":166,"skipped":2664,"failed":0}
+[k8s.io] Variable Expansion 
+  should allow composing env vars into new env vars [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Variable Expansion
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:30:55.713: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename var-expansion
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-1004
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should allow composing env vars into new env vars [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test env composition
+Feb 12 10:30:55.890: INFO: Waiting up to 5m0s for pod "var-expansion-8996eb2d-39bc-4dff-b5ed-6d334cc9f7b8" in namespace "var-expansion-1004" to be "success or failure"
+Feb 12 10:30:55.902: INFO: Pod "var-expansion-8996eb2d-39bc-4dff-b5ed-6d334cc9f7b8": Phase="Pending", Reason="", readiness=false. Elapsed: 11.434225ms
+Feb 12 10:30:57.907: INFO: Pod "var-expansion-8996eb2d-39bc-4dff-b5ed-6d334cc9f7b8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017048651s
+Feb 12 10:30:59.919: INFO: Pod "var-expansion-8996eb2d-39bc-4dff-b5ed-6d334cc9f7b8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.028836433s
+STEP: Saw pod success
+Feb 12 10:30:59.919: INFO: Pod "var-expansion-8996eb2d-39bc-4dff-b5ed-6d334cc9f7b8" satisfied condition "success or failure"
+Feb 12 10:30:59.926: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod var-expansion-8996eb2d-39bc-4dff-b5ed-6d334cc9f7b8 container dapi-container: 
+STEP: delete the pod
+Feb 12 10:30:59.966: INFO: Waiting for pod var-expansion-8996eb2d-39bc-4dff-b5ed-6d334cc9f7b8 to disappear
+Feb 12 10:30:59.972: INFO: Pod var-expansion-8996eb2d-39bc-4dff-b5ed-6d334cc9f7b8 no longer exists
+[AfterEach] [k8s.io] Variable Expansion
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:30:59.972: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "var-expansion-1004" for this suite.
+•{"msg":"PASSED [k8s.io] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance]","total":280,"completed":144,"skipped":2290,"failed":0}
 SSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with secret pod [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Subpath
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:15:10.290: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7447
+Feb 12 10:31:00.000: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-4690
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 17:15:10.470: INFO: Waiting up to 5m0s for pod "downwardapi-volume-853f89c9-9240-4ab8-b5de-754d9ea9b2e9" in namespace "projected-7447" to be "success or failure"
-Aug 30 17:15:10.476: INFO: Pod "downwardapi-volume-853f89c9-9240-4ab8-b5de-754d9ea9b2e9": Phase="Pending", Reason="", readiness=false. Elapsed: 6.498804ms
-Aug 30 17:15:12.486: INFO: Pod "downwardapi-volume-853f89c9-9240-4ab8-b5de-754d9ea9b2e9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015973423s
-Aug 30 17:15:14.491: INFO: Pod "downwardapi-volume-853f89c9-9240-4ab8-b5de-754d9ea9b2e9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.020914463s
-Aug 30 17:15:16.497: INFO: Pod "downwardapi-volume-853f89c9-9240-4ab8-b5de-754d9ea9b2e9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.027098723s
+[BeforeEach] Atomic writer volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
+STEP: Setting up data
+[It] should support subpaths with secret pod [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating pod pod-subpath-test-secret-p9hc
+STEP: Creating a pod to test atomic-volume-subpath
+Feb 12 10:31:00.207: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-p9hc" in namespace "subpath-4690" to be "success or failure"
+Feb 12 10:31:00.216: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Pending", Reason="", readiness=false. Elapsed: 8.165258ms
+Feb 12 10:31:02.225: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Running", Reason="", readiness=true. Elapsed: 2.017639193s
+Feb 12 10:31:04.233: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Running", Reason="", readiness=true. Elapsed: 4.025250226s
+Feb 12 10:31:06.241: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Running", Reason="", readiness=true. Elapsed: 6.033114035s
+Feb 12 10:31:08.247: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Running", Reason="", readiness=true. Elapsed: 8.039774156s
+Feb 12 10:31:10.253: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Running", Reason="", readiness=true. Elapsed: 10.045735391s
+Feb 12 10:31:12.260: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Running", Reason="", readiness=true. Elapsed: 12.052749479s
+Feb 12 10:31:14.266: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Running", Reason="", readiness=true. Elapsed: 14.058820422s
+Feb 12 10:31:16.273: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Running", Reason="", readiness=true. Elapsed: 16.065221285s
+Feb 12 10:31:18.284: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Running", Reason="", readiness=true. Elapsed: 18.076175094s
+Feb 12 10:31:20.290: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Running", Reason="", readiness=true. Elapsed: 20.082066817s
+Feb 12 10:31:22.295: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Running", Reason="", readiness=true. Elapsed: 22.087368347s
+Feb 12 10:31:24.301: INFO: Pod "pod-subpath-test-secret-p9hc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.093377564s
 STEP: Saw pod success
-Aug 30 17:15:16.497: INFO: Pod "downwardapi-volume-853f89c9-9240-4ab8-b5de-754d9ea9b2e9" satisfied condition "success or failure"
-Aug 30 17:15:16.501: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-6rshr pod downwardapi-volume-853f89c9-9240-4ab8-b5de-754d9ea9b2e9 container client-container: 
+Feb 12 10:31:24.301: INFO: Pod "pod-subpath-test-secret-p9hc" satisfied condition "success or failure"
+Feb 12 10:31:24.308: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-subpath-test-secret-p9hc container test-container-subpath-secret-p9hc: 
 STEP: delete the pod
-Aug 30 17:15:16.569: INFO: Waiting for pod downwardapi-volume-853f89c9-9240-4ab8-b5de-754d9ea9b2e9 to disappear
-Aug 30 17:15:16.575: INFO: Pod downwardapi-volume-853f89c9-9240-4ab8-b5de-754d9ea9b2e9 no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:15:16.575: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-7447" for this suite.
+Feb 12 10:31:24.348: INFO: Waiting for pod pod-subpath-test-secret-p9hc to disappear
+Feb 12 10:31:24.353: INFO: Pod pod-subpath-test-secret-p9hc no longer exists
+STEP: Deleting pod pod-subpath-test-secret-p9hc
+Feb 12 10:31:24.353: INFO: Deleting pod "pod-subpath-test-secret-p9hc" in namespace "subpath-4690"
+[AfterEach] [sig-storage] Subpath
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:31:24.358: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-4690" for this suite.
 
-• [SLOW TEST:6.298 seconds]
-[sig-storage] Projected downwardAPI
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:34
-  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:24.377 seconds]
+[sig-storage] Subpath
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
+  Atomic writer volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
+    should support subpaths with secret pod [LinuxOnly] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":167,"skipped":2688,"failed":0}
+{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance]","total":280,"completed":145,"skipped":2314,"failed":0}
 SSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] ResourceQuota 
-  should verify ResourceQuota with terminating scopes. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  should perform canary updates and phased rolling updates of template modifications [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:15:16.589: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename resourcequota
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-9218
+Feb 12 10:31:24.377: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-4726
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should verify ResourceQuota with terminating scopes. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a ResourceQuota with terminating scope
-STEP: Ensuring ResourceQuota status is calculated
-STEP: Creating a ResourceQuota with not terminating scope
-STEP: Ensuring ResourceQuota status is calculated
-STEP: Creating a long running pod
-STEP: Ensuring resource quota with not terminating scope captures the pod usage
-STEP: Ensuring resource quota with terminating scope ignored the pod usage
-STEP: Deleting the pod
-STEP: Ensuring resource quota status released the pod usage
-STEP: Creating a terminating pod
-STEP: Ensuring resource quota with terminating scope captures the pod usage
-STEP: Ensuring resource quota with not terminating scope ignored the pod usage
-STEP: Deleting the pod
-STEP: Ensuring resource quota status released the pod usage
-[AfterEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:15:32.911: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-9218" for this suite.
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
+STEP: Creating service test in namespace statefulset-4726
+[It] should perform canary updates and phased rolling updates of template modifications [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a new StatefulSet
+Feb 12 10:31:24.561: INFO: Found 0 stateful pods, waiting for 3
+Feb 12 10:31:34.569: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
+Feb 12 10:31:34.569: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
+Feb 12 10:31:34.569: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Updating stateful set template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine
+Feb 12 10:31:34.615: INFO: Updating stateful set ss2
+STEP: Creating a new revision
+STEP: Not applying an update when the partition is greater than the number of replicas
+STEP: Performing a canary update
+Feb 12 10:31:44.664: INFO: Updating stateful set ss2
+Feb 12 10:31:44.678: INFO: Waiting for Pod statefulset-4726/ss2-2 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
+STEP: Restoring Pods to the correct revision when they are deleted
+Feb 12 10:31:54.735: INFO: Found 2 stateful pods, waiting for 3
+Feb 12 10:32:04.745: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
+Feb 12 10:32:04.745: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
+Feb 12 10:32:04.745: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
+STEP: Performing a phased rolling update
+Feb 12 10:32:04.779: INFO: Updating stateful set ss2
+Feb 12 10:32:04.793: INFO: Waiting for Pod statefulset-4726/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
+Feb 12 10:32:14.830: INFO: Updating stateful set ss2
+Feb 12 10:32:14.841: INFO: Waiting for StatefulSet statefulset-4726/ss2 to complete update
+Feb 12 10:32:14.841: INFO: Waiting for Pod statefulset-4726/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
+Feb 12 10:32:24.853: INFO: Waiting for StatefulSet statefulset-4726/ss2 to complete update
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
+Feb 12 10:32:34.854: INFO: Deleting all statefulset in ns statefulset-4726
+Feb 12 10:32:34.860: INFO: Scaling statefulset ss2 to 0
+Feb 12 10:32:44.886: INFO: Waiting for statefulset status.replicas updated to 0
+Feb 12 10:32:44.892: INFO: Deleting statefulset ss2
+[AfterEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:32:44.916: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-4726" for this suite.
+
+• [SLOW TEST:80.556 seconds]
+[sig-apps] StatefulSet
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+    should perform canary updates and phased rolling updates of template modifications [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance]","total":280,"completed":146,"skipped":2340,"failed":0}
+SSSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client Update Demo 
+  should scale a replication controller  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:32:44.937: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-5526
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[BeforeEach] Update Demo
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:325
+[It] should scale a replication controller  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating a replication controller
+Feb 12 10:32:45.096: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-5526'
+Feb 12 10:32:45.338: INFO: stderr: ""
+Feb 12 10:32:45.338: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Feb 12 10:32:45.338: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5526'
+Feb 12 10:32:45.425: INFO: stderr: ""
+Feb 12 10:32:45.426: INFO: stdout: "update-demo-nautilus-2j42t update-demo-nautilus-tfbcr "
+Feb 12 10:32:45.426: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-2j42t -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5526'
+Feb 12 10:32:45.501: INFO: stderr: ""
+Feb 12 10:32:45.501: INFO: stdout: ""
+Feb 12 10:32:45.501: INFO: update-demo-nautilus-2j42t is created but not running
+Feb 12 10:32:50.501: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5526'
+Feb 12 10:32:50.577: INFO: stderr: ""
+Feb 12 10:32:50.577: INFO: stdout: "update-demo-nautilus-2j42t update-demo-nautilus-tfbcr "
+Feb 12 10:32:50.577: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-2j42t -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5526'
+Feb 12 10:32:50.663: INFO: stderr: ""
+Feb 12 10:32:50.663: INFO: stdout: "true"
+Feb 12 10:32:50.663: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-2j42t -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5526'
+Feb 12 10:32:50.735: INFO: stderr: ""
+Feb 12 10:32:50.735: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Feb 12 10:32:50.735: INFO: validating pod update-demo-nautilus-2j42t
+Feb 12 10:32:50.830: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Feb 12 10:32:50.830: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Feb 12 10:32:50.830: INFO: update-demo-nautilus-2j42t is verified up and running
+Feb 12 10:32:50.830: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-tfbcr -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5526'
+Feb 12 10:32:50.903: INFO: stderr: ""
+Feb 12 10:32:50.903: INFO: stdout: "true"
+Feb 12 10:32:50.903: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-tfbcr -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5526'
+Feb 12 10:32:50.977: INFO: stderr: ""
+Feb 12 10:32:50.977: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Feb 12 10:32:50.977: INFO: validating pod update-demo-nautilus-tfbcr
+Feb 12 10:32:51.073: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Feb 12 10:32:51.073: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Feb 12 10:32:51.073: INFO: update-demo-nautilus-tfbcr is verified up and running
+STEP: scaling down the replication controller
+Feb 12 10:32:51.074: INFO: scanned /root for discovery docs: 
+Feb 12 10:32:51.074: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=kubectl-5526'
+Feb 12 10:32:52.186: INFO: stderr: ""
+Feb 12 10:32:52.186: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Feb 12 10:32:52.186: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5526'
+Feb 12 10:32:52.285: INFO: stderr: ""
+Feb 12 10:32:52.285: INFO: stdout: "update-demo-nautilus-2j42t update-demo-nautilus-tfbcr "
+STEP: Replicas for name=update-demo: expected=1 actual=2
+Feb 12 10:32:57.285: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5526'
+Feb 12 10:32:57.390: INFO: stderr: ""
+Feb 12 10:32:57.390: INFO: stdout: "update-demo-nautilus-2j42t update-demo-nautilus-tfbcr "
+STEP: Replicas for name=update-demo: expected=1 actual=2
+Feb 12 10:33:02.390: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5526'
+Feb 12 10:33:02.484: INFO: stderr: ""
+Feb 12 10:33:02.484: INFO: stdout: "update-demo-nautilus-2j42t update-demo-nautilus-tfbcr "
+STEP: Replicas for name=update-demo: expected=1 actual=2
+Feb 12 10:33:07.485: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5526'
+Feb 12 10:33:07.561: INFO: stderr: ""
+Feb 12 10:33:07.561: INFO: stdout: "update-demo-nautilus-tfbcr "
+Feb 12 10:33:07.561: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-tfbcr -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5526'
+Feb 12 10:33:07.630: INFO: stderr: ""
+Feb 12 10:33:07.630: INFO: stdout: "true"
+Feb 12 10:33:07.630: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-tfbcr -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5526'
+Feb 12 10:33:07.700: INFO: stderr: ""
+Feb 12 10:33:07.700: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Feb 12 10:33:07.700: INFO: validating pod update-demo-nautilus-tfbcr
+Feb 12 10:33:07.712: INFO: got data: {
+  "image": "nautilus.jpg"
+}
 
-• [SLOW TEST:16.336 seconds]
-[sig-api-machinery] ResourceQuota
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should verify ResourceQuota with terminating scopes. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance]","total":280,"completed":168,"skipped":2714,"failed":0}
-S
+Feb 12 10:33:07.712: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Feb 12 10:33:07.712: INFO: update-demo-nautilus-tfbcr is verified up and running
+STEP: scaling up the replication controller
+Feb 12 10:33:07.714: INFO: scanned /root for discovery docs: 
+Feb 12 10:33:07.714: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=kubectl-5526'
+Feb 12 10:33:08.813: INFO: stderr: ""
+Feb 12 10:33:08.813: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Feb 12 10:33:08.813: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5526'
+Feb 12 10:33:08.887: INFO: stderr: ""
+Feb 12 10:33:08.887: INFO: stdout: "update-demo-nautilus-ll9m5 update-demo-nautilus-tfbcr "
+Feb 12 10:33:08.887: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-ll9m5 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5526'
+Feb 12 10:33:08.969: INFO: stderr: ""
+Feb 12 10:33:08.969: INFO: stdout: ""
+Feb 12 10:33:08.969: INFO: update-demo-nautilus-ll9m5 is created but not running
+Feb 12 10:33:13.969: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-5526'
+Feb 12 10:33:14.049: INFO: stderr: ""
+Feb 12 10:33:14.049: INFO: stdout: "update-demo-nautilus-ll9m5 update-demo-nautilus-tfbcr "
+Feb 12 10:33:14.049: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-ll9m5 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5526'
+Feb 12 10:33:14.119: INFO: stderr: ""
+Feb 12 10:33:14.119: INFO: stdout: "true"
+Feb 12 10:33:14.120: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-ll9m5 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5526'
+Feb 12 10:33:14.215: INFO: stderr: ""
+Feb 12 10:33:14.215: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Feb 12 10:33:14.215: INFO: validating pod update-demo-nautilus-ll9m5
+Feb 12 10:33:14.290: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Feb 12 10:33:14.290: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Feb 12 10:33:14.290: INFO: update-demo-nautilus-ll9m5 is verified up and running
+Feb 12 10:33:14.290: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-tfbcr -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-5526'
+Feb 12 10:33:14.360: INFO: stderr: ""
+Feb 12 10:33:14.360: INFO: stdout: "true"
+Feb 12 10:33:14.360: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-tfbcr -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-5526'
+Feb 12 10:33:14.432: INFO: stderr: ""
+Feb 12 10:33:14.432: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Feb 12 10:33:14.433: INFO: validating pod update-demo-nautilus-tfbcr
+Feb 12 10:33:14.444: INFO: got data: {
+  "image": "nautilus.jpg"
+}
+
+Feb 12 10:33:14.444: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Feb 12 10:33:14.444: INFO: update-demo-nautilus-tfbcr is verified up and running
+STEP: using delete to clean up resources
+Feb 12 10:33:14.444: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete --grace-period=0 --force -f - --namespace=kubectl-5526'
+Feb 12 10:33:14.534: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Feb 12 10:33:14.534: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n"
+Feb 12 10:33:14.534: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-5526'
+Feb 12 10:33:14.614: INFO: stderr: "No resources found in kubectl-5526 namespace.\n"
+Feb 12 10:33:14.614: INFO: stdout: ""
+Feb 12 10:33:14.614: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -l name=update-demo --namespace=kubectl-5526 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Feb 12 10:33:14.689: INFO: stderr: ""
+Feb 12 10:33:14.689: INFO: stdout: "update-demo-nautilus-ll9m5\nupdate-demo-nautilus-tfbcr\n"
+Feb 12 10:33:15.189: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-5526'
+Feb 12 10:33:15.334: INFO: stderr: "No resources found in kubectl-5526 namespace.\n"
+Feb 12 10:33:15.335: INFO: stdout: ""
+Feb 12 10:33:15.335: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -l name=update-demo --namespace=kubectl-5526 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Feb 12 10:33:15.444: INFO: stderr: ""
+Feb 12 10:33:15.444: INFO: stdout: ""
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:33:15.444: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-5526" for this suite.
+
+• [SLOW TEST:30.536 seconds]
+[sig-cli] Kubectl client
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Update Demo
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:323
+    should scale a replication controller  [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:15:32.926: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-9947
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-cd80d43a-e83a-412e-a087-e171cd919837
-STEP: Creating a pod to test consume secrets
-Aug 30 17:15:33.104: INFO: Waiting up to 5m0s for pod "pod-secrets-2a311052-57ed-4d6e-b05c-84381ef6c902" in namespace "secrets-9947" to be "success or failure"
-Aug 30 17:15:33.110: INFO: Pod "pod-secrets-2a311052-57ed-4d6e-b05c-84381ef6c902": Phase="Pending", Reason="", readiness=false. Elapsed: 6.221694ms
-Aug 30 17:15:35.117: INFO: Pod "pod-secrets-2a311052-57ed-4d6e-b05c-84381ef6c902": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012770344s
-STEP: Saw pod success
-Aug 30 17:15:35.117: INFO: Pod "pod-secrets-2a311052-57ed-4d6e-b05c-84381ef6c902" satisfied condition "success or failure"
-Aug 30 17:15:35.123: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-secrets-2a311052-57ed-4d6e-b05c-84381ef6c902 container secret-volume-test: 
-STEP: delete the pod
-Aug 30 17:15:35.195: INFO: Waiting for pod pod-secrets-2a311052-57ed-4d6e-b05c-84381ef6c902 to disappear
-Aug 30 17:15:35.199: INFO: Pod pod-secrets-2a311052-57ed-4d6e-b05c-84381ef6c902 no longer exists
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:15:35.199: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-9947" for this suite.
-•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":169,"skipped":2715,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-cli] Kubectl client Update Demo should scale a replication controller  [Conformance]","total":280,"completed":147,"skipped":2354,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Projected downwardAPI 
+  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:15:35.213: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-2895
+Feb 12 10:33:15.473: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-4718
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0777 on node default medium
-Aug 30 17:15:35.388: INFO: Waiting up to 5m0s for pod "pod-1c473627-b6a5-4d6e-801b-bffa346ca5af" in namespace "emptydir-2895" to be "success or failure"
-Aug 30 17:15:35.401: INFO: Pod "pod-1c473627-b6a5-4d6e-801b-bffa346ca5af": Phase="Pending", Reason="", readiness=false. Elapsed: 13.458977ms
-Aug 30 17:15:37.407: INFO: Pod "pod-1c473627-b6a5-4d6e-801b-bffa346ca5af": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.019230287s
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Feb 12 10:33:15.658: INFO: Waiting up to 5m0s for pod "downwardapi-volume-47ddee72-6dec-40ee-b786-4452e60c7140" in namespace "projected-4718" to be "success or failure"
+Feb 12 10:33:15.676: INFO: Pod "downwardapi-volume-47ddee72-6dec-40ee-b786-4452e60c7140": Phase="Pending", Reason="", readiness=false. Elapsed: 18.361513ms
+Feb 12 10:33:17.682: INFO: Pod "downwardapi-volume-47ddee72-6dec-40ee-b786-4452e60c7140": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024371012s
+Feb 12 10:33:19.689: INFO: Pod "downwardapi-volume-47ddee72-6dec-40ee-b786-4452e60c7140": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.031037304s
 STEP: Saw pod success
-Aug 30 17:15:37.407: INFO: Pod "pod-1c473627-b6a5-4d6e-801b-bffa346ca5af" satisfied condition "success or failure"
-Aug 30 17:15:37.411: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-1c473627-b6a5-4d6e-801b-bffa346ca5af container test-container: 
+Feb 12 10:33:19.689: INFO: Pod "downwardapi-volume-47ddee72-6dec-40ee-b786-4452e60c7140" satisfied condition "success or failure"
+Feb 12 10:33:19.694: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-47ddee72-6dec-40ee-b786-4452e60c7140 container client-container: 
 STEP: delete the pod
-Aug 30 17:15:37.491: INFO: Waiting for pod pod-1c473627-b6a5-4d6e-801b-bffa346ca5af to disappear
-Aug 30 17:15:37.496: INFO: Pod pod-1c473627-b6a5-4d6e-801b-bffa346ca5af no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:15:37.496: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-2895" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":170,"skipped":2772,"failed":0}
-
-------------------------------
-[sig-network] Proxy version v1 
-  should proxy through a service and a pod  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] version v1
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:15:37.510: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename proxy
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-4269
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should proxy through a service and a pod  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: starting an echo server on multiple ports
-STEP: creating replication controller proxy-service-9wvsq in namespace proxy-4269
-I0830 17:15:37.697014      23 runners.go:189] Created replication controller with name: proxy-service-9wvsq, namespace: proxy-4269, replica count: 1
-I0830 17:15:38.747528      23 runners.go:189] proxy-service-9wvsq Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-I0830 17:15:39.747955      23 runners.go:189] proxy-service-9wvsq Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady 
-I0830 17:15:40.748170      23 runners.go:189] proxy-service-9wvsq Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-Aug 30 17:15:40.755: INFO: setup took 3.081448192s, starting test cases
-STEP: running 16 cases, 20 attempts per case, 320 total attempts
-Aug 30 17:15:40.811: INFO: (0) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 55.292595ms)
-Aug 30 17:15:40.811: INFO: (0) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:1080/proxy/: test<... (200; 55.461107ms)
-Aug 30 17:15:40.811: INFO: (0) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 55.175117ms)
-Aug 30 17:15:40.811: INFO: (0) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 55.531876ms)
-Aug 30 17:15:40.811: INFO: (0) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 55.529326ms)
-Aug 30 17:15:40.811: INFO: (0) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 55.415929ms)
-Aug 30 17:15:40.811: INFO: (0) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 55.031416ms)
-Aug 30 17:15:40.811: INFO: (0) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 55.319023ms)
-Aug 30 17:15:40.811: INFO: (0) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 55.378748ms)
-Aug 30 17:15:40.811: INFO: (0) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 55.409143ms)
-Aug 30 17:15:40.811: INFO: (0) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 55.508125ms)
-Aug 30 17:15:40.817: INFO: (0) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 61.908249ms)
-Aug 30 17:15:40.817: INFO: (0) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 61.544929ms)
-Aug 30 17:15:40.818: INFO: (0) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 62.118673ms)
-Aug 30 17:15:40.818: INFO: (0) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 62.263628ms)
-Aug 30 17:15:40.820: INFO: (0) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test<... (200; 12.684967ms)
-Aug 30 17:15:40.833: INFO: (1) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 12.567932ms)
-Aug 30 17:15:40.833: INFO: (1) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 12.626477ms)
-Aug 30 17:15:40.835: INFO: (1) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 14.043488ms)
-Aug 30 17:15:40.837: INFO: (1) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 16.332829ms)
-Aug 30 17:15:40.837: INFO: (1) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 16.747056ms)
-Aug 30 17:15:40.837: INFO: (1) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test (200; 14.301493ms)
-Aug 30 17:15:40.856: INFO: (2) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 13.837758ms)
-Aug 30 17:15:40.856: INFO: (2) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 14.085813ms)
-Aug 30 17:15:40.856: INFO: (2) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 14.70654ms)
-Aug 30 17:15:40.858: INFO: (2) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 16.883174ms)
-Aug 30 17:15:40.859: INFO: (2) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 16.956622ms)
-Aug 30 17:15:40.859: INFO: (2) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 16.867176ms)
-Aug 30 17:15:40.859: INFO: (2) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test<... (200; 17.020736ms)
-Aug 30 17:15:40.859: INFO: (2) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 17.422567ms)
-Aug 30 17:15:40.859: INFO: (2) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 18.033669ms)
-Aug 30 17:15:40.859: INFO: (2) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 17.38063ms)
-Aug 30 17:15:40.859: INFO: (2) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 17.640874ms)
-Aug 30 17:15:40.863: INFO: (2) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 21.370564ms)
-Aug 30 17:15:40.871: INFO: (3) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 8.2623ms)
-Aug 30 17:15:40.873: INFO: (3) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 10.147238ms)
-Aug 30 17:15:40.873: INFO: (3) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 10.14417ms)
-Aug 30 17:15:40.873: INFO: (3) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 10.244407ms)
-Aug 30 17:15:40.874: INFO: (3) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 10.871874ms)
-Aug 30 17:15:40.874: INFO: (3) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 10.953524ms)
-Aug 30 17:15:40.874: INFO: (3) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 11.238809ms)
-Aug 30 17:15:40.874: INFO: (3) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test<... (200; 10.944493ms)
-Aug 30 17:15:40.874: INFO: (3) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 11.279ms)
-Aug 30 17:15:40.876: INFO: (3) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 12.722382ms)
-Aug 30 17:15:40.876: INFO: (3) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 12.951222ms)
-Aug 30 17:15:40.877: INFO: (3) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 13.907972ms)
-Aug 30 17:15:40.878: INFO: (3) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 15.145606ms)
-Aug 30 17:15:40.878: INFO: (3) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 15.101726ms)
-Aug 30 17:15:40.879: INFO: (3) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 15.682346ms)
-Aug 30 17:15:40.890: INFO: (4) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:1080/proxy/: test<... (200; 11.506743ms)
-Aug 30 17:15:40.890: INFO: (4) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 11.00304ms)
-Aug 30 17:15:40.890: INFO: (4) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 11.60588ms)
-Aug 30 17:15:40.890: INFO: (4) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 10.74146ms)
-Aug 30 17:15:40.890: INFO: (4) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 11.343776ms)
-Aug 30 17:15:40.890: INFO: (4) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 10.9826ms)
-Aug 30 17:15:40.890: INFO: (4) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 11.287717ms)
-Aug 30 17:15:40.890: INFO: (4) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 11.759739ms)
-Aug 30 17:15:40.890: INFO: (4) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: ... (200; 12.106247ms)
-Aug 30 17:15:40.961: INFO: (5) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 11.929259ms)
-Aug 30 17:15:40.961: INFO: (5) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 12.019159ms)
-Aug 30 17:15:40.963: INFO: (5) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:1080/proxy/: test<... (200; 13.89746ms)
-Aug 30 17:15:40.963: INFO: (5) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 13.871851ms)
-Aug 30 17:15:40.963: INFO: (5) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 14.062104ms)
-Aug 30 17:15:40.965: INFO: (5) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 15.836325ms)
-Aug 30 17:15:40.967: INFO: (5) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 17.658701ms)
-Aug 30 17:15:40.967: INFO: (5) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 17.619377ms)
-Aug 30 17:15:40.967: INFO: (5) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 17.684458ms)
-Aug 30 17:15:40.967: INFO: (5) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 17.741951ms)
-Aug 30 17:15:40.977: INFO: (6) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 9.845138ms)
-Aug 30 17:15:40.977: INFO: (6) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 10.045084ms)
-Aug 30 17:15:40.977: INFO: (6) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 10.004217ms)
-Aug 30 17:15:40.977: INFO: (6) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 10.060649ms)
-Aug 30 17:15:40.979: INFO: (6) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test<... (200; 12.999898ms)
-Aug 30 17:15:40.980: INFO: (6) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 13.267485ms)
-Aug 30 17:15:40.980: INFO: (6) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 13.381566ms)
-Aug 30 17:15:40.980: INFO: (6) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 13.620613ms)
-Aug 30 17:15:40.982: INFO: (6) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 15.474571ms)
-Aug 30 17:15:40.983: INFO: (6) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 15.647318ms)
-Aug 30 17:15:40.984: INFO: (6) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 17.285855ms)
-Aug 30 17:15:40.984: INFO: (6) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 17.293668ms)
-Aug 30 17:15:40.984: INFO: (6) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 17.289365ms)
-Aug 30 17:15:40.984: INFO: (6) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 17.346754ms)
-Aug 30 17:15:40.995: INFO: (7) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test<... (200; 10.367459ms)
-Aug 30 17:15:40.995: INFO: (7) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 10.397277ms)
-Aug 30 17:15:40.995: INFO: (7) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 10.343953ms)
-Aug 30 17:15:40.995: INFO: (7) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 10.943902ms)
-Aug 30 17:15:40.996: INFO: (7) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 11.424027ms)
-Aug 30 17:15:40.997: INFO: (7) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 12.074236ms)
-Aug 30 17:15:40.997: INFO: (7) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 11.905012ms)
-Aug 30 17:15:40.997: INFO: (7) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 11.93122ms)
-Aug 30 17:15:40.997: INFO: (7) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 12.376721ms)
-Aug 30 17:15:41.049: INFO: (7) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 64.566214ms)
-Aug 30 17:15:41.049: INFO: (7) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 64.300314ms)
-Aug 30 17:15:41.051: INFO: (7) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 66.271094ms)
-Aug 30 17:15:41.051: INFO: (7) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 66.456563ms)
-Aug 30 17:15:41.052: INFO: (7) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 66.890005ms)
-Aug 30 17:15:41.066: INFO: (8) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 14.676118ms)
-Aug 30 17:15:41.066: INFO: (8) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 14.500391ms)
-Aug 30 17:15:41.066: INFO: (8) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 14.147284ms)
-Aug 30 17:15:41.066: INFO: (8) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 14.023738ms)
-Aug 30 17:15:41.066: INFO: (8) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 14.430358ms)
-Aug 30 17:15:41.066: INFO: (8) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 14.66756ms)
-Aug 30 17:15:41.066: INFO: (8) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 14.82776ms)
-Aug 30 17:15:41.066: INFO: (8) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test (200; 14.253334ms)
-Aug 30 17:15:41.066: INFO: (8) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:1080/proxy/: test<... (200; 14.528493ms)
-Aug 30 17:15:41.066: INFO: (8) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 14.206794ms)
-Aug 30 17:15:41.067: INFO: (8) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 15.242471ms)
-Aug 30 17:15:41.070: INFO: (8) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 17.796402ms)
-Aug 30 17:15:41.070: INFO: (8) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 17.937886ms)
-Aug 30 17:15:41.070: INFO: (8) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 17.881218ms)
-Aug 30 17:15:41.072: INFO: (8) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 19.074874ms)
-Aug 30 17:15:41.081: INFO: (9) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 8.754789ms)
-Aug 30 17:15:41.081: INFO: (9) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 8.885706ms)
-Aug 30 17:15:41.081: INFO: (9) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test<... (200; 9.275028ms)
-Aug 30 17:15:41.082: INFO: (9) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 9.932788ms)
-Aug 30 17:15:41.082: INFO: (9) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 10.035001ms)
-Aug 30 17:15:41.082: INFO: (9) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 10.910486ms)
-Aug 30 17:15:41.082: INFO: (9) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 10.607868ms)
-Aug 30 17:15:41.151: INFO: (9) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 79.071785ms)
-Aug 30 17:15:41.151: INFO: (9) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 79.31495ms)
-Aug 30 17:15:41.151: INFO: (9) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 79.0615ms)
-Aug 30 17:15:41.151: INFO: (9) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 78.982617ms)
-Aug 30 17:15:41.151: INFO: (9) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 79.642246ms)
-Aug 30 17:15:41.153: INFO: (9) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 80.326231ms)
-Aug 30 17:15:41.164: INFO: (10) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 10.830333ms)
-Aug 30 17:15:41.164: INFO: (10) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 10.834875ms)
-Aug 30 17:15:41.164: INFO: (10) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 10.701947ms)
-Aug 30 17:15:41.164: INFO: (10) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 10.437597ms)
-Aug 30 17:15:41.164: INFO: (10) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 10.387226ms)
-Aug 30 17:15:41.164: INFO: (10) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test<... (200; 10.64367ms)
-Aug 30 17:15:41.164: INFO: (10) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 10.970383ms)
-Aug 30 17:15:41.164: INFO: (10) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 11.000652ms)
-Aug 30 17:15:41.166: INFO: (10) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 12.646118ms)
-Aug 30 17:15:41.166: INFO: (10) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 12.699324ms)
-Aug 30 17:15:41.167: INFO: (10) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 13.818686ms)
-Aug 30 17:15:41.167: INFO: (10) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 13.86985ms)
-Aug 30 17:15:41.167: INFO: (10) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 13.725175ms)
-Aug 30 17:15:41.167: INFO: (10) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 13.586659ms)
-Aug 30 17:15:41.167: INFO: (10) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 14.461871ms)
-Aug 30 17:15:41.178: INFO: (11) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 10.906257ms)
-Aug 30 17:15:41.179: INFO: (11) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 11.086138ms)
-Aug 30 17:15:41.179: INFO: (11) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 11.264594ms)
-Aug 30 17:15:41.179: INFO: (11) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 10.856153ms)
-Aug 30 17:15:41.179: INFO: (11) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:1080/proxy/: test<... (200; 11.016539ms)
-Aug 30 17:15:41.180: INFO: (11) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 12.468631ms)
-Aug 30 17:15:41.180: INFO: (11) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: ... (200; 12.258149ms)
-Aug 30 17:15:41.180: INFO: (11) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 12.850746ms)
-Aug 30 17:15:41.180: INFO: (11) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 12.826147ms)
-Aug 30 17:15:41.182: INFO: (11) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 14.287773ms)
-Aug 30 17:15:41.182: INFO: (11) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 14.43057ms)
-Aug 30 17:15:41.182: INFO: (11) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 13.974868ms)
-Aug 30 17:15:41.182: INFO: (11) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 14.011938ms)
-Aug 30 17:15:41.184: INFO: (11) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 15.931239ms)
-Aug 30 17:15:41.196: INFO: (12) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 11.57295ms)
-Aug 30 17:15:41.196: INFO: (12) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 12.152778ms)
-Aug 30 17:15:41.196: INFO: (12) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 12.097499ms)
-Aug 30 17:15:41.196: INFO: (12) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:1080/proxy/: test<... (200; 12.002259ms)
-Aug 30 17:15:41.196: INFO: (12) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 11.800768ms)
-Aug 30 17:15:41.249: INFO: (12) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 65.244225ms)
-Aug 30 17:15:41.249: INFO: (12) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 64.399893ms)
-Aug 30 17:15:41.249: INFO: (12) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 64.351797ms)
-Aug 30 17:15:41.249: INFO: (12) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test (200; 12.858228ms)
-Aug 30 17:15:41.269: INFO: (13) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 12.942005ms)
-Aug 30 17:15:41.269: INFO: (13) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: ... (200; 12.852064ms)
-Aug 30 17:15:41.269: INFO: (13) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:1080/proxy/: test<... (200; 12.900512ms)
-Aug 30 17:15:41.269: INFO: (13) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 13.160968ms)
-Aug 30 17:15:41.269: INFO: (13) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 13.065116ms)
-Aug 30 17:15:41.273: INFO: (13) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 17.164618ms)
-Aug 30 17:15:41.273: INFO: (13) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 17.135606ms)
-Aug 30 17:15:41.273: INFO: (13) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 17.231616ms)
-Aug 30 17:15:41.273: INFO: (13) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 17.095982ms)
-Aug 30 17:15:41.273: INFO: (13) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 17.18489ms)
-Aug 30 17:15:41.283: INFO: (14) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 9.756512ms)
-Aug 30 17:15:41.283: INFO: (14) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 9.754309ms)
-Aug 30 17:15:41.283: INFO: (14) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 9.794106ms)
-Aug 30 17:15:41.283: INFO: (14) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 9.762324ms)
-Aug 30 17:15:41.283: INFO: (14) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 9.91277ms)
-Aug 30 17:15:41.283: INFO: (14) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 9.781502ms)
-Aug 30 17:15:41.283: INFO: (14) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 9.895165ms)
-Aug 30 17:15:41.284: INFO: (14) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: ... (200; 10.965605ms)
-Aug 30 17:15:41.284: INFO: (14) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:1080/proxy/: test<... (200; 10.802395ms)
-Aug 30 17:15:41.292: INFO: (14) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 18.777379ms)
-Aug 30 17:15:41.293: INFO: (14) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 19.105807ms)
-Aug 30 17:15:41.293: INFO: (14) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 19.13377ms)
-Aug 30 17:15:41.293: INFO: (14) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 19.303629ms)
-Aug 30 17:15:41.293: INFO: (14) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 19.376015ms)
-Aug 30 17:15:41.293: INFO: (14) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 19.530142ms)
-Aug 30 17:15:41.353: INFO: (15) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test<... (200; 59.356348ms)
-Aug 30 17:15:41.353: INFO: (15) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 58.885708ms)
-Aug 30 17:15:41.353: INFO: (15) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 59.22126ms)
-Aug 30 17:15:41.353: INFO: (15) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 59.840846ms)
-Aug 30 17:15:41.353: INFO: (15) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 59.212405ms)
-Aug 30 17:15:41.357: INFO: (15) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 63.651302ms)
-Aug 30 17:15:41.357: INFO: (15) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 63.002227ms)
-Aug 30 17:15:41.357: INFO: (15) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 63.456629ms)
-Aug 30 17:15:41.357: INFO: (15) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 63.923251ms)
-Aug 30 17:15:41.369: INFO: (16) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 11.928258ms)
-Aug 30 17:15:41.369: INFO: (16) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test<... (200; 11.964792ms)
-Aug 30 17:15:41.369: INFO: (16) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 11.981988ms)
-Aug 30 17:15:41.369: INFO: (16) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 12.102815ms)
-Aug 30 17:15:41.369: INFO: (16) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 12.222212ms)
-Aug 30 17:15:41.370: INFO: (16) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 12.702972ms)
-Aug 30 17:15:41.373: INFO: (16) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 16.036069ms)
-Aug 30 17:15:41.373: INFO: (16) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 16.004771ms)
-Aug 30 17:15:41.373: INFO: (16) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 16.00063ms)
-Aug 30 17:15:41.373: INFO: (16) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 15.979915ms)
-Aug 30 17:15:41.373: INFO: (16) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 16.114104ms)
-Aug 30 17:15:41.374: INFO: (16) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 16.419849ms)
-Aug 30 17:15:41.374: INFO: (16) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 16.398462ms)
-Aug 30 17:15:41.378: INFO: (16) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 20.78876ms)
-Aug 30 17:15:41.389: INFO: (17) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 11.19479ms)
-Aug 30 17:15:41.389: INFO: (17) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:1080/proxy/: test<... (200; 11.317557ms)
-Aug 30 17:15:41.389: INFO: (17) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 11.324859ms)
-Aug 30 17:15:41.390: INFO: (17) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 11.685834ms)
-Aug 30 17:15:41.390: INFO: (17) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 11.82859ms)
-Aug 30 17:15:41.390: INFO: (17) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 11.800443ms)
-Aug 30 17:15:41.390: INFO: (17) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 11.638098ms)
-Aug 30 17:15:41.390: INFO: (17) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 11.702824ms)
-Aug 30 17:15:41.390: INFO: (17) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test (200; 11.591151ms)
-Aug 30 17:15:41.467: INFO: (18) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 11.580405ms)
-Aug 30 17:15:41.467: INFO: (18) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 11.665908ms)
-Aug 30 17:15:41.467: INFO: (18) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: ... (200; 11.890983ms)
-Aug 30 17:15:41.468: INFO: (18) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 12.272112ms)
-Aug 30 17:15:41.468: INFO: (18) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 12.329538ms)
-Aug 30 17:15:41.469: INFO: (18) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 13.269398ms)
-Aug 30 17:15:41.471: INFO: (18) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 14.707156ms)
-Aug 30 17:15:41.471: INFO: (18) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 14.66879ms)
-Aug 30 17:15:41.471: INFO: (18) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:1080/proxy/: test<... (200; 14.6562ms)
-Aug 30 17:15:41.471: INFO: (18) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 14.964936ms)
-Aug 30 17:15:41.471: INFO: (18) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 15.130802ms)
-Aug 30 17:15:41.471: INFO: (18) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 15.015835ms)
-Aug 30 17:15:41.471: INFO: (18) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 15.087768ms)
-Aug 30 17:15:41.482: INFO: (19) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7/proxy/: test (200; 10.817148ms)
-Aug 30 17:15:41.483: INFO: (19) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:460/proxy/: tls baz (200; 11.509774ms)
-Aug 30 17:15:41.483: INFO: (19) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:443/proxy/: test<... (200; 11.58009ms)
-Aug 30 17:15:41.483: INFO: (19) /api/v1/namespaces/proxy-4269/pods/https:proxy-service-9wvsq-2m8h7:462/proxy/: tls qux (200; 11.512016ms)
-Aug 30 17:15:41.483: INFO: (19) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 11.483573ms)
-Aug 30 17:15:41.483: INFO: (19) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 12.117451ms)
-Aug 30 17:15:41.483: INFO: (19) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:1080/proxy/: ... (200; 12.336066ms)
-Aug 30 17:15:41.483: INFO: (19) /api/v1/namespaces/proxy-4269/pods/proxy-service-9wvsq-2m8h7:162/proxy/: bar (200; 12.232334ms)
-Aug 30 17:15:41.483: INFO: (19) /api/v1/namespaces/proxy-4269/pods/http:proxy-service-9wvsq-2m8h7:160/proxy/: foo (200; 12.312142ms)
-Aug 30 17:15:41.484: INFO: (19) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname1/proxy/: tls baz (200; 13.097417ms)
-Aug 30 17:15:41.484: INFO: (19) /api/v1/namespaces/proxy-4269/services/https:proxy-service-9wvsq:tlsportname2/proxy/: tls qux (200; 13.129519ms)
-Aug 30 17:15:41.486: INFO: (19) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname1/proxy/: foo (200; 15.301113ms)
-Aug 30 17:15:41.486: INFO: (19) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname2/proxy/: bar (200; 15.338158ms)
-Aug 30 17:15:41.486: INFO: (19) /api/v1/namespaces/proxy-4269/services/http:proxy-service-9wvsq:portname1/proxy/: foo (200; 15.417041ms)
-Aug 30 17:15:41.486: INFO: (19) /api/v1/namespaces/proxy-4269/services/proxy-service-9wvsq:portname2/proxy/: bar (200; 15.492965ms)
-STEP: deleting ReplicationController proxy-service-9wvsq in namespace proxy-4269, will wait for the garbage collector to delete the pods
-Aug 30 17:15:41.553: INFO: Deleting ReplicationController proxy-service-9wvsq took: 11.810098ms
-Aug 30 17:15:42.054: INFO: Terminating ReplicationController proxy-service-9wvsq pods took: 500.631149ms
-[AfterEach] version v1
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:15:54.354: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "proxy-4269" for this suite.
-
-• [SLOW TEST:16.858 seconds]
-[sig-network] Proxy
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  version v1
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:57
-    should proxy through a service and a pod  [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] Proxy version v1 should proxy through a service and a pod  [Conformance]","total":280,"completed":171,"skipped":2772,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Feb 12 10:33:19.740: INFO: Waiting for pod downwardapi-volume-47ddee72-6dec-40ee-b786-4452e60c7140 to disappear
+Feb 12 10:33:19.745: INFO: Pod downwardapi-volume-47ddee72-6dec-40ee-b786-4452e60c7140 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:33:19.746: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-4718" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":148,"skipped":2378,"failed":0}
+SSSS
 ------------------------------
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should be able to deny custom resource creation, update and deletion [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  listing mutating webhooks should work [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:15:54.369: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:33:19.766: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-3020
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-9682
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
 STEP: Setting up server cert
 STEP: Create role binding to let webhook read extension-apiserver-authentication
 STEP: Deploying the webhook pod
 STEP: Wait for the deployment to be ready
-Aug 30 17:15:55.232: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Aug 30 17:15:57.255: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404555, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404555, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404555, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404555, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+Feb 12 10:33:20.406: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:33:22.423: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722800, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722800, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722800, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748722800, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
 STEP: Deploying the webhook service
 STEP: Verifying the service has paired with the endpoint
-Aug 30 17:16:00.277: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should be able to deny custom resource creation, update and deletion [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:16:00.283: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Registering the custom resource webhook via the AdmissionRegistration API
-STEP: Creating a custom resource that should be denied by the webhook
-STEP: Creating a custom resource whose deletion would be denied by the webhook
-STEP: Updating the custom resource with disallowed data should be denied
-STEP: Deleting the custom resource should be denied
-STEP: Remove the offending key and value from the custom resource data
-STEP: Deleting the updated custom resource should be successful
+Feb 12 10:33:25.442: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] listing mutating webhooks should work [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Listing all of the created validation webhooks
+STEP: Creating a configMap that should be mutated
+STEP: Deleting the collection of validation webhooks
+STEP: Creating a configMap that should not be mutated
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:16:01.994: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-3020" for this suite.
-STEP: Destroying namespace "webhook-3020-markers" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:33:26.013: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-9682" for this suite.
+STEP: Destroying namespace "webhook-9682-markers" for this suite.
 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:7.714 seconds]
+• [SLOW TEST:6.369 seconds]
 [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should be able to deny custom resource creation, update and deletion [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  listing mutating webhooks should work [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance]","total":280,"completed":172,"skipped":2803,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance]","total":280,"completed":149,"skipped":2382,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition 
-  creating/deleting custom resource definition objects works  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] InitContainer [NodeConformance] 
+  should invoke init containers on a RestartAlways pod [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:16:02.083: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename custom-resource-definition
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-5030
+Feb 12 10:33:26.136: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename init-container
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-3399
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] creating/deleting custom resource definition objects works  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:16:02.261: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:16:03.300: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "custom-resource-definition-5030" for this suite.
-•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works  [Conformance]","total":280,"completed":173,"skipped":2841,"failed":0}
-SSSSSSSS
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
+[It] should invoke init containers on a RestartAlways pod [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating the pod
+Feb 12 10:33:26.325: INFO: PodSpec: initContainers in spec.initContainers
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:33:30.525: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "init-container-3399" for this suite.
+•{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance]","total":280,"completed":150,"skipped":2418,"failed":0}
+SSS
 ------------------------------
-[sig-api-machinery] Garbage collector 
-  should delete RS created by deployment when not orphaning [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] DNS 
+  should provide DNS for services  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:16:03.354: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-5080
+Feb 12 10:33:30.551: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename dns
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-5026
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should delete RS created by deployment when not orphaning [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the deployment
-STEP: Wait for the Deployment to create new ReplicaSet
-STEP: delete the deployment
-STEP: wait for all rs to be garbage collected
-STEP: expected 0 rs, got 1 rs
-STEP: expected 0 pods, got 2 pods
-STEP: Gathering metrics
-W0830 17:16:04.641775      23 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-Aug 30 17:16:04.641: INFO: For apiserver_request_total:
-For apiserver_request_latency_seconds:
-For apiserver_init_events_total:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
+[It] should provide DNS for services  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a test headless service
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-5026.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-5026.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-5026.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-5026.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-5026.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-5026.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-5026.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-5026.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-5026.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-5026.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-5026.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-5026.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5026.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 219.21.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.21.219_udp@PTR;check="$$(dig +tcp +noall +answer +search 219.21.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.21.219_tcp@PTR;sleep 1; done
 
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:16:04.641: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-5080" for this suite.
-•{"msg":"PASSED [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance]","total":280,"completed":174,"skipped":2849,"failed":0}
-SSSSSSSSSSSS
+STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-5026.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-5026.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-5026.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-5026.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-5026.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-5026.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-5026.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-5026.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-5026.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-5026.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-5026.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-5026.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5026.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 219.21.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.21.219_udp@PTR;check="$$(dig +tcp +noall +answer +search 219.21.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.21.219_tcp@PTR;sleep 1; done
+
+STEP: creating a pod to probe DNS
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Feb 12 10:33:32.896: INFO: Unable to read wheezy_udp@dns-test-service.dns-5026.svc.cluster.local from pod dns-5026/dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263: the server could not find the requested resource (get pods dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263)
+Feb 12 10:33:32.940: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5026.svc.cluster.local from pod dns-5026/dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263: the server could not find the requested resource (get pods dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263)
+Feb 12 10:33:32.948: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-5026.svc.cluster.local from pod dns-5026/dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263: the server could not find the requested resource (get pods dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263)
+Feb 12 10:33:32.958: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-5026.svc.cluster.local from pod dns-5026/dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263: the server could not find the requested resource (get pods dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263)
+Feb 12 10:33:33.484: INFO: Unable to read jessie_udp@dns-test-service.dns-5026.svc.cluster.local from pod dns-5026/dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263: the server could not find the requested resource (get pods dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263)
+Feb 12 10:33:33.494: INFO: Unable to read jessie_tcp@dns-test-service.dns-5026.svc.cluster.local from pod dns-5026/dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263: the server could not find the requested resource (get pods dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263)
+Feb 12 10:33:33.504: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-5026.svc.cluster.local from pod dns-5026/dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263: the server could not find the requested resource (get pods dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263)
+Feb 12 10:33:33.514: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-5026.svc.cluster.local from pod dns-5026/dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263: the server could not find the requested resource (get pods dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263)
+Feb 12 10:33:34.005: INFO: Lookups using dns-5026/dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263 failed for: [wheezy_udp@dns-test-service.dns-5026.svc.cluster.local wheezy_tcp@dns-test-service.dns-5026.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-5026.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-5026.svc.cluster.local jessie_udp@dns-test-service.dns-5026.svc.cluster.local jessie_tcp@dns-test-service.dns-5026.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-5026.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-5026.svc.cluster.local]
+
+Feb 12 10:33:40.504: INFO: DNS probes using dns-5026/dns-test-bb26aa49-c159-4c9c-9d58-1d868b03f263 succeeded
+
+STEP: deleting the pod
+STEP: deleting the test service
+STEP: deleting the test headless service
+[AfterEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:33:40.606: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-5026" for this suite.
+
+• [SLOW TEST:10.074 seconds]
+[sig-network] DNS
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should provide DNS for services  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] Container Runtime blackbox test on terminated container 
-  should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-network] DNS should provide DNS for services  [Conformance]","total":280,"completed":151,"skipped":2421,"failed":0}
+SSSSSSSSSSS
+------------------------------
+[k8s.io] Kubelet when scheduling a busybox command in a pod 
+  should print the output to logs [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:16:04.660: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-runtime
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-8032
+Feb 12 10:33:40.634: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-9347
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the container
-STEP: wait for the container to reach Succeeded
-STEP: get the container status
-STEP: the container should be terminated
-STEP: the termination message should be set
-Aug 30 17:16:06.853: INFO: Expected: &{DONE} to match Container's Termination Message: DONE --
-STEP: delete the container
-[AfterEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:16:06.881: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-runtime-8032" for this suite.
-•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]","total":280,"completed":175,"skipped":2861,"failed":0}
-SSSSS
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should print the output to logs [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[AfterEach] [k8s.io] Kubelet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:33:44.843: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-9347" for this suite.
+•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance]","total":280,"completed":152,"skipped":2432,"failed":0}
+SSSSSSS
 ------------------------------
-[k8s.io] Probing container 
-  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] Servers with support for Table transformation 
+  should return a 406 for a backend which does not implement metadata [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Servers with support for Table transformation
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:16:06.897: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-8778
+Feb 12 10:33:44.862: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename tables
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in tables-6454
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod busybox-2343cc3c-147f-4370-a70c-11198a14fc41 in namespace container-probe-8778
-Aug 30 17:16:09.083: INFO: Started pod busybox-2343cc3c-147f-4370-a70c-11198a14fc41 in namespace container-probe-8778
-STEP: checking the pod's current state and verifying that restartCount is present
-Aug 30 17:16:09.087: INFO: Initial restart count of pod busybox-2343cc3c-147f-4370-a70c-11198a14fc41 is 0
-Aug 30 17:17:03.255: INFO: Restart count of pod container-probe-8778/busybox-2343cc3c-147f-4370-a70c-11198a14fc41 is now 1 (54.168115482s elapsed)
-STEP: deleting the pod
-[AfterEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:17:03.278: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-8778" for this suite.
-
-• [SLOW TEST:56.399 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Servers with support for Table transformation
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/table_conversion.go:46
+[It] should return a 406 for a backend which does not implement metadata [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[AfterEach] [sig-api-machinery] Servers with support for Table transformation
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:33:45.027: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "tables-6454" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance]","total":280,"completed":153,"skipped":2439,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-{"msg":"PASSED [k8s.io] Probing container should be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":280,"completed":176,"skipped":2866,"failed":0}
-SSSS
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:33:45.046: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-1199
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name configmap-test-volume-map-97c736c5-b9c3-48b3-92cf-b3d58322505c
+STEP: Creating a pod to test consume configMaps
+Feb 12 10:33:45.227: INFO: Waiting up to 5m0s for pod "pod-configmaps-1c07b377-03df-4622-b14e-86e8bc16142f" in namespace "configmap-1199" to be "success or failure"
+Feb 12 10:33:45.235: INFO: Pod "pod-configmaps-1c07b377-03df-4622-b14e-86e8bc16142f": Phase="Pending", Reason="", readiness=false. Elapsed: 7.731359ms
+Feb 12 10:33:47.242: INFO: Pod "pod-configmaps-1c07b377-03df-4622-b14e-86e8bc16142f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014681516s
+STEP: Saw pod success
+Feb 12 10:33:47.242: INFO: Pod "pod-configmaps-1c07b377-03df-4622-b14e-86e8bc16142f" satisfied condition "success or failure"
+Feb 12 10:33:47.248: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-configmaps-1c07b377-03df-4622-b14e-86e8bc16142f container configmap-volume-test: 
+STEP: delete the pod
+Feb 12 10:33:47.284: INFO: Waiting for pod pod-configmaps-1c07b377-03df-4622-b14e-86e8bc16142f to disappear
+Feb 12 10:33:47.290: INFO: Pod pod-configmaps-1c07b377-03df-4622-b14e-86e8bc16142f no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:33:47.291: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-1199" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":154,"skipped":2484,"failed":0}
+SSSSSSSSS
 ------------------------------
-[sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial] 
-  evicts pods with minTolerationSeconds [Disruptive] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] Garbage collector 
+  should not be blocked by dependency circle [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:17:03.296: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename taint-multiple-pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in taint-multiple-pods-5282
+Feb 12 10:33:47.315: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-7984
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/taints.go:344
-Aug 30 17:17:03.476: INFO: Waiting up to 1m0s for all nodes to be ready
-Aug 30 17:18:03.505: INFO: Waiting for terminating namespaces to be deleted...
-[It] evicts pods with minTolerationSeconds [Disruptive] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:18:03.510: INFO: Starting informer...
-STEP: Starting pods...
-Aug 30 17:18:03.736: INFO: Pod1 is running on adoring-wozniak-54dcfd79fc-948mf. Tainting Node
-Aug 30 17:18:05.964: INFO: Pod2 is running on adoring-wozniak-54dcfd79fc-948mf. Tainting Node
-STEP: Trying to apply a taint on the Node
-STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute
-STEP: Waiting for Pod1 and Pod2 to be deleted
-Aug 30 17:18:12.588: INFO: Noticed Pod "taint-eviction-b1" gets evicted.
-Aug 30 17:18:32.734: INFO: Noticed Pod "taint-eviction-b2" gets evicted.
-STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute
-[AfterEach] [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:18:32.760: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "taint-multiple-pods-5282" for this suite.
+[It] should not be blocked by dependency circle [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:33:47.524: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"9a9ef66c-cfe0-4a21-8255-c5c529f24658", Controller:(*bool)(0xc003a09ba6), BlockOwnerDeletion:(*bool)(0xc003a09ba7)}}
+Feb 12 10:33:47.534: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"1c74a43e-f945-49ca-a59c-f3d2edbd53f9", Controller:(*bool)(0xc0039c132e), BlockOwnerDeletion:(*bool)(0xc0039c132f)}}
+Feb 12 10:33:47.543: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"eac38642-a208-4a59-8666-d421fe6acbe3", Controller:(*bool)(0xc003a09d66), BlockOwnerDeletion:(*bool)(0xc003a09d67)}}
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:33:52.564: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-7984" for this suite.
 
-• [SLOW TEST:89.489 seconds]
-[sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
-  evicts pods with minTolerationSeconds [Disruptive] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:5.267 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should not be blocked by dependency circle [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial] evicts pods with minTolerationSeconds [Disruptive] [Conformance]","total":280,"completed":177,"skipped":2870,"failed":0}
-SSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]","total":280,"completed":155,"skipped":2493,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Secrets 
+  should be consumable via the environment [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:33:52.584: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-1235
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable via the environment [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating secret secrets-1235/secret-test-d1c55995-dad1-4f15-a4f1-b8f5b674386c
+STEP: Creating a pod to test consume secrets
+Feb 12 10:33:52.771: INFO: Waiting up to 5m0s for pod "pod-configmaps-ad919486-9c20-444b-b24b-0ec0113100e4" in namespace "secrets-1235" to be "success or failure"
+Feb 12 10:33:52.782: INFO: Pod "pod-configmaps-ad919486-9c20-444b-b24b-0ec0113100e4": Phase="Pending", Reason="", readiness=false. Elapsed: 10.284257ms
+Feb 12 10:33:54.789: INFO: Pod "pod-configmaps-ad919486-9c20-444b-b24b-0ec0113100e4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.017230902s
+STEP: Saw pod success
+Feb 12 10:33:54.789: INFO: Pod "pod-configmaps-ad919486-9c20-444b-b24b-0ec0113100e4" satisfied condition "success or failure"
+Feb 12 10:33:54.795: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-configmaps-ad919486-9c20-444b-b24b-0ec0113100e4 container env-test: 
+STEP: delete the pod
+Feb 12 10:33:54.835: INFO: Waiting for pod pod-configmaps-ad919486-9c20-444b-b24b-0ec0113100e4 to disappear
+Feb 12 10:33:54.840: INFO: Pod pod-configmaps-ad919486-9c20-444b-b24b-0ec0113100e4 no longer exists
+[AfterEach] [sig-api-machinery] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:33:54.840: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-1235" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Secrets should be consumable via the environment [NodeConformance] [Conformance]","total":280,"completed":156,"skipped":2547,"failed":0}
+SSSSSSSSS
+------------------------------
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:33:54.858: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7315
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name projected-configmap-test-volume-4a2bb7b4-cd04-43b3-98f3-fb0593968f2f
+STEP: Creating a pod to test consume configMaps
+Feb 12 10:33:55.037: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-a153d6b1-414e-4350-8ac4-c3554d078263" in namespace "projected-7315" to be "success or failure"
+Feb 12 10:33:55.044: INFO: Pod "pod-projected-configmaps-a153d6b1-414e-4350-8ac4-c3554d078263": Phase="Pending", Reason="", readiness=false. Elapsed: 6.871663ms
+Feb 12 10:33:57.053: INFO: Pod "pod-projected-configmaps-a153d6b1-414e-4350-8ac4-c3554d078263": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016025429s
+Feb 12 10:33:59.060: INFO: Pod "pod-projected-configmaps-a153d6b1-414e-4350-8ac4-c3554d078263": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022691146s
+STEP: Saw pod success
+Feb 12 10:33:59.060: INFO: Pod "pod-projected-configmaps-a153d6b1-414e-4350-8ac4-c3554d078263" satisfied condition "success or failure"
+Feb 12 10:33:59.066: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-projected-configmaps-a153d6b1-414e-4350-8ac4-c3554d078263 container projected-configmap-volume-test: 
+STEP: delete the pod
+Feb 12 10:33:59.102: INFO: Waiting for pod pod-projected-configmaps-a153d6b1-414e-4350-8ac4-c3554d078263 to disappear
+Feb 12 10:33:59.107: INFO: Pod pod-projected-configmaps-a153d6b1-414e-4350-8ac4-c3554d078263 no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:33:59.107: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-7315" for this suite.
+•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","total":280,"completed":157,"skipped":2556,"failed":0}
+SSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Variable Expansion 
+  should allow substituting values in a container's args [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Variable Expansion
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:18:32.788: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-6642
+Feb 12 10:33:59.127: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename var-expansion
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-2133
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-map-1c3b662a-1664-4e34-859d-f0795133201d
-STEP: Creating a pod to test consume secrets
-Aug 30 17:18:32.966: INFO: Waiting up to 5m0s for pod "pod-secrets-a4ae5a9c-c38f-4ed8-ad7a-4d1ce45e5ae6" in namespace "secrets-6642" to be "success or failure"
-Aug 30 17:18:32.971: INFO: Pod "pod-secrets-a4ae5a9c-c38f-4ed8-ad7a-4d1ce45e5ae6": Phase="Pending", Reason="", readiness=false. Elapsed: 4.66063ms
-Aug 30 17:18:34.976: INFO: Pod "pod-secrets-a4ae5a9c-c38f-4ed8-ad7a-4d1ce45e5ae6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.009804427s
-Aug 30 17:18:36.982: INFO: Pod "pod-secrets-a4ae5a9c-c38f-4ed8-ad7a-4d1ce45e5ae6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015769934s
+[It] should allow substituting values in a container's args [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test substitution in container's args
+Feb 12 10:33:59.304: INFO: Waiting up to 5m0s for pod "var-expansion-b0b7baf2-8908-43e7-a1a8-952a68d42d0a" in namespace "var-expansion-2133" to be "success or failure"
+Feb 12 10:33:59.309: INFO: Pod "var-expansion-b0b7baf2-8908-43e7-a1a8-952a68d42d0a": Phase="Pending", Reason="", readiness=false. Elapsed: 5.088379ms
+Feb 12 10:34:01.316: INFO: Pod "var-expansion-b0b7baf2-8908-43e7-a1a8-952a68d42d0a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011826954s
 STEP: Saw pod success
-Aug 30 17:18:36.982: INFO: Pod "pod-secrets-a4ae5a9c-c38f-4ed8-ad7a-4d1ce45e5ae6" satisfied condition "success or failure"
-Aug 30 17:18:36.987: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-secrets-a4ae5a9c-c38f-4ed8-ad7a-4d1ce45e5ae6 container secret-volume-test: 
+Feb 12 10:34:01.316: INFO: Pod "var-expansion-b0b7baf2-8908-43e7-a1a8-952a68d42d0a" satisfied condition "success or failure"
+Feb 12 10:34:01.321: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-dbhjj pod var-expansion-b0b7baf2-8908-43e7-a1a8-952a68d42d0a container dapi-container: 
 STEP: delete the pod
-Aug 30 17:18:37.055: INFO: Waiting for pod pod-secrets-a4ae5a9c-c38f-4ed8-ad7a-4d1ce45e5ae6 to disappear
-Aug 30 17:18:37.060: INFO: Pod pod-secrets-a4ae5a9c-c38f-4ed8-ad7a-4d1ce45e5ae6 no longer exists
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:18:37.060: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-6642" for this suite.
-•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":178,"skipped":2890,"failed":0}
-SSSSS
+Feb 12 10:34:01.406: INFO: Waiting for pod var-expansion-b0b7baf2-8908-43e7-a1a8-952a68d42d0a to disappear
+Feb 12 10:34:01.412: INFO: Pod var-expansion-b0b7baf2-8908-43e7-a1a8-952a68d42d0a no longer exists
+[AfterEach] [k8s.io] Variable Expansion
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:34:01.412: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "var-expansion-2133" for this suite.
+•{"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance]","total":280,"completed":158,"skipped":2572,"failed":0}
+SSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Garbage collector 
-  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Pods 
+  should get a host IP [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:18:37.075: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-9319
+Feb 12 10:34:01.433: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-969
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the deployment
-STEP: Wait for the Deployment to create new ReplicaSet
-STEP: delete the deployment
-STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the rs
-STEP: Gathering metrics
-Aug 30 17:19:07.801: INFO: For apiserver_request_total:
-For apiserver_request_latency_seconds:
-For apiserver_init_events_total:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
-
-W0830 17:19:07.801010      23 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:19:07.801: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-9319" for this suite.
-
-• [SLOW TEST:30.742 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]","total":280,"completed":179,"skipped":2895,"failed":0}
-SS
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
+[It] should get a host IP [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating pod
+Feb 12 10:34:03.640: INFO: Pod pod-hostip-d6fef3f7-a2d6-45e3-bb37-9efca4dd1d9f has hostIP: 138.68.97.197
+[AfterEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:34:03.640: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-969" for this suite.
+•{"msg":"PASSED [k8s.io] Pods should get a host IP [NodeConformance] [Conformance]","total":280,"completed":159,"skipped":2587,"failed":0}
+SSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-network] Services 
-  should be able to change the type from ExternalName to NodePort [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Kubelet when scheduling a busybox Pod with hostAliases 
+  should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:19:07.820: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename services
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-7509
+Feb 12 10:34:03.661: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-8709
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should be able to change the type from ExternalName to NodePort [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a service externalname-service with the type=ExternalName in namespace services-7509
-STEP: changing the ExternalName service to type=NodePort
-STEP: creating replication controller externalname-service in namespace services-7509
-I0830 17:19:08.038754      23 runners.go:189] Created replication controller with name: externalname-service, namespace: services-7509, replica count: 2
-Aug 30 17:19:11.089: INFO: Creating new exec pod
-I0830 17:19:11.089228      23 runners.go:189] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-Aug 30 17:19:14.117: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=services-7509 execpodhsdxw -- /bin/sh -x -c nc -zv -t -w 2 externalname-service 80'
-Aug 30 17:19:14.657: INFO: stderr: "+ nc -zv -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n"
-Aug 30 17:19:14.657: INFO: stdout: ""
-Aug 30 17:19:14.658: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=services-7509 execpodhsdxw -- /bin/sh -x -c nc -zv -t -w 2 10.240.23.168 80'
-Aug 30 17:19:15.221: INFO: stderr: "+ nc -zv -t -w 2 10.240.23.168 80\nConnection to 10.240.23.168 80 port [tcp/http] succeeded!\n"
-Aug 30 17:19:15.221: INFO: stdout: ""
-Aug 30 17:19:15.221: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=services-7509 execpodhsdxw -- /bin/sh -x -c nc -zv -t -w 2 46.101.153.64 30694'
-Aug 30 17:19:15.748: INFO: stderr: "+ nc -zv -t -w 2 46.101.153.64 30694\nConnection to 46.101.153.64 30694 port [tcp/30694] succeeded!\n"
-Aug 30 17:19:15.748: INFO: stdout: ""
-Aug 30 17:19:15.748: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=services-7509 execpodhsdxw -- /bin/sh -x -c nc -zv -t -w 2 46.101.135.210 30694'
-Aug 30 17:19:16.306: INFO: stderr: "+ nc -zv -t -w 2 46.101.135.210 30694\nConnection to 46.101.135.210 30694 port [tcp/30694] succeeded!\n"
-Aug 30 17:19:16.306: INFO: stdout: ""
-Aug 30 17:19:16.306: INFO: Cleaning up the ExternalName to NodePort test service
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:19:16.340: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-7509" for this suite.
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
-
-• [SLOW TEST:8.535 seconds]
-[sig-network] Services
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should be able to change the type from ExternalName to NodePort [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance]","total":280,"completed":180,"skipped":2897,"failed":0}
-S
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[AfterEach] [k8s.io] Kubelet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:34:05.867: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-8709" for this suite.
+•{"msg":"PASSED [k8s.io] Kubelet when scheduling a busybox Pod with hostAliases should write entries to /etc/hosts [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":160,"skipped":2607,"failed":0}
+SSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Watchers 
-  should receive events on concurrent watches in same order [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-cli] Kubectl client Kubectl replace 
+  should update a single-container pod's image  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:19:16.354: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename watch
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-72
+Feb 12 10:34:05.891: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-1968
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should receive events on concurrent watches in same order [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: starting a background goroutine to produce watch events
-STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order
-[AfterEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:19:21.779: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "watch-72" for this suite.
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[BeforeEach] Kubectl replace
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1796
+[It] should update a single-container pod's image  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: running the image docker.io/library/httpd:2.4.38-alpine
+Feb 12 10:34:06.067: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 run e2e-test-httpd-pod --generator=run-pod/v1 --image=docker.io/library/httpd:2.4.38-alpine --labels=run=e2e-test-httpd-pod --namespace=kubectl-1968'
+Feb 12 10:34:06.149: INFO: stderr: ""
+Feb 12 10:34:06.150: INFO: stdout: "pod/e2e-test-httpd-pod created\n"
+STEP: verifying the pod e2e-test-httpd-pod is running
+STEP: verifying the pod e2e-test-httpd-pod was created
+Feb 12 10:34:11.200: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pod e2e-test-httpd-pod --namespace=kubectl-1968 -o json'
+Feb 12 10:34:11.273: INFO: stderr: ""
+Feb 12 10:34:11.273: INFO: stdout: "{\n    \"apiVersion\": \"v1\",\n    \"kind\": \"Pod\",\n    \"metadata\": {\n        \"annotations\": {\n            \"cni.projectcalico.org/podIP\": \"172.25.0.35/32\"\n        },\n        \"creationTimestamp\": \"2021-02-12T10:34:05Z\",\n        \"labels\": {\n            \"run\": \"e2e-test-httpd-pod\"\n        },\n        \"name\": \"e2e-test-httpd-pod\",\n        \"namespace\": \"kubectl-1968\",\n        \"resourceVersion\": \"24923\",\n        \"selfLink\": \"/api/v1/namespaces/kubectl-1968/pods/e2e-test-httpd-pod\",\n        \"uid\": \"05d4a3b0-c44d-4582-9973-92f2119e69b3\"\n    },\n    \"spec\": {\n        \"containers\": [\n            {\n                \"image\": \"docker.io/library/httpd:2.4.38-alpine\",\n                \"imagePullPolicy\": \"IfNotPresent\",\n                \"name\": \"e2e-test-httpd-pod\",\n                \"resources\": {},\n                \"terminationMessagePath\": \"/dev/termination-log\",\n                \"terminationMessagePolicy\": \"File\",\n                \"volumeMounts\": [\n                    {\n                        \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n                        \"name\": \"default-token-xqsj4\",\n                        \"readOnly\": true\n                    }\n                ]\n            }\n        ],\n        \"dnsPolicy\": \"ClusterFirst\",\n        \"enableServiceLinks\": true,\n        \"nodeName\": \"nostalgic-fermat-cdc5d8777-r6snm\",\n        \"priority\": 0,\n        \"restartPolicy\": \"Always\",\n        \"schedulerName\": \"default-scheduler\",\n        \"securityContext\": {},\n        \"serviceAccount\": \"default\",\n        \"serviceAccountName\": \"default\",\n        \"terminationGracePeriodSeconds\": 30,\n        \"tolerations\": [\n            {\n                \"effect\": \"NoExecute\",\n                \"key\": \"node.kubernetes.io/not-ready\",\n                \"operator\": \"Exists\",\n                \"tolerationSeconds\": 300\n            },\n            {\n                \"effect\": \"NoExecute\",\n                \"key\": \"node.kubernetes.io/unreachable\",\n                \"operator\": \"Exists\",\n                \"tolerationSeconds\": 300\n            }\n        ],\n        \"volumes\": [\n            {\n                \"name\": \"default-token-xqsj4\",\n                \"secret\": {\n                    \"defaultMode\": 420,\n                    \"secretName\": \"default-token-xqsj4\"\n                }\n            }\n        ]\n    },\n    \"status\": {\n        \"conditions\": [\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2021-02-12T10:34:05Z\",\n                \"status\": \"True\",\n                \"type\": \"Initialized\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2021-02-12T10:34:07Z\",\n                \"status\": \"True\",\n                \"type\": \"Ready\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2021-02-12T10:34:07Z\",\n                \"status\": \"True\",\n                \"type\": \"ContainersReady\"\n            },\n            {\n                \"lastProbeTime\": null,\n                \"lastTransitionTime\": \"2021-02-12T10:34:05Z\",\n                \"status\": \"True\",\n                \"type\": \"PodScheduled\"\n            }\n        ],\n        \"containerStatuses\": [\n            {\n                \"containerID\": \"docker://7f76abdddc9b877bb90b26f2f3792c37041def25c8017c2ed90a711765612bf2\",\n                \"image\": \"httpd:2.4.38-alpine\",\n                \"imageID\": \"docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060\",\n                \"lastState\": {},\n                \"name\": \"e2e-test-httpd-pod\",\n                \"ready\": true,\n                \"restartCount\": 0,\n                \"started\": true,\n                \"state\": {\n                    \"running\": {\n                        \"startedAt\": \"2021-02-12T10:34:07Z\"\n                    }\n                }\n            }\n        ],\n        \"hostIP\": \"139.59.138.127\",\n        \"phase\": \"Running\",\n        \"podIP\": \"172.25.0.35\",\n        \"podIPs\": [\n            {\n                \"ip\": \"172.25.0.35\"\n            }\n        ],\n        \"qosClass\": \"BestEffort\",\n        \"startTime\": \"2021-02-12T10:34:05Z\"\n    }\n}\n"
+STEP: replace the image in the pod
+Feb 12 10:34:11.273: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 replace -f - --namespace=kubectl-1968'
+Feb 12 10:34:11.553: INFO: stderr: ""
+Feb 12 10:34:11.553: INFO: stdout: "pod/e2e-test-httpd-pod replaced\n"
+STEP: verifying the pod e2e-test-httpd-pod has the right image docker.io/library/busybox:1.29
+[AfterEach] Kubectl replace
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1801
+Feb 12 10:34:11.579: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete pods e2e-test-httpd-pod --namespace=kubectl-1968'
+Feb 12 10:34:13.133: INFO: stderr: ""
+Feb 12 10:34:13.133: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:34:13.133: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-1968" for this suite.
 
-• [SLOW TEST:5.516 seconds]
-[sig-api-machinery] Watchers
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should receive events on concurrent watches in same order [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:7.258 seconds]
+[sig-cli] Kubectl client
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Kubectl replace
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1792
+    should update a single-container pod's image  [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance]","total":280,"completed":181,"skipped":2898,"failed":0}
-[sig-network] DNS 
-  should provide DNS for the cluster  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-cli] Kubectl client Kubectl replace should update a single-container pod's image  [Conformance]","total":280,"completed":161,"skipped":2619,"failed":0}
+SSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client Update Demo 
+  should do a rolling update of a replication controller  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:19:21.871: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename dns
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-7569
+Feb 12 10:34:13.150: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-1416
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide DNS for the cluster  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-7569.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[BeforeEach] Update Demo
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:325
+[It] should do a rolling update of a replication controller  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating the initial replication controller
+Feb 12 10:34:13.307: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-1416'
+Feb 12 10:34:13.467: INFO: stderr: ""
+Feb 12 10:34:13.467: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Feb 12 10:34:13.467: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1416'
+Feb 12 10:34:13.547: INFO: stderr: ""
+Feb 12 10:34:13.547: INFO: stdout: "update-demo-nautilus-crnh8 update-demo-nautilus-dcqld "
+Feb 12 10:34:13.547: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-crnh8 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1416'
+Feb 12 10:34:13.624: INFO: stderr: ""
+Feb 12 10:34:13.624: INFO: stdout: ""
+Feb 12 10:34:13.624: INFO: update-demo-nautilus-crnh8 is created but not running
+Feb 12 10:34:18.624: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1416'
+Feb 12 10:34:18.705: INFO: stderr: ""
+Feb 12 10:34:18.705: INFO: stdout: "update-demo-nautilus-crnh8 update-demo-nautilus-dcqld "
+Feb 12 10:34:18.705: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-crnh8 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1416'
+Feb 12 10:34:18.777: INFO: stderr: ""
+Feb 12 10:34:18.777: INFO: stdout: "true"
+Feb 12 10:34:18.777: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-crnh8 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1416'
+Feb 12 10:34:18.866: INFO: stderr: ""
+Feb 12 10:34:18.866: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Feb 12 10:34:18.866: INFO: validating pod update-demo-nautilus-crnh8
+Feb 12 10:34:18.965: INFO: got data: {
+  "image": "nautilus.jpg"
+}
 
-STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-7569.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+Feb 12 10:34:18.965: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Feb 12 10:34:18.965: INFO: update-demo-nautilus-crnh8 is verified up and running
+Feb 12 10:34:18.965: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-dcqld -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1416'
+Feb 12 10:34:19.038: INFO: stderr: ""
+Feb 12 10:34:19.038: INFO: stdout: "true"
+Feb 12 10:34:19.038: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-dcqld -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1416'
+Feb 12 10:34:19.110: INFO: stderr: ""
+Feb 12 10:34:19.110: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Feb 12 10:34:19.110: INFO: validating pod update-demo-nautilus-dcqld
+Feb 12 10:34:19.206: INFO: got data: {
+  "image": "nautilus.jpg"
+}
 
-STEP: creating a pod to probe DNS
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Aug 30 17:19:32.772: INFO: DNS probes using dns-7569/dns-test-2416c6ae-f8d1-4eb8-bb34-1b604d707dcf succeeded
+Feb 12 10:34:19.206: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Feb 12 10:34:19.206: INFO: update-demo-nautilus-dcqld is verified up and running
+STEP: rolling-update to new replication controller
+Feb 12 10:34:19.208: INFO: scanned /root for discovery docs: 
+Feb 12 10:34:19.208: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 rolling-update update-demo-nautilus --update-period=1s -f - --namespace=kubectl-1416'
+Feb 12 10:34:41.748: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n"
+Feb 12 10:34:41.748: INFO: stdout: "Created update-demo-kitten\nScaling up update-demo-kitten from 0 to 2, scaling down update-demo-nautilus from 2 to 0 (keep 2 pods available, don't exceed 3 pods)\nScaling update-demo-kitten up to 1\nScaling update-demo-nautilus down to 1\nScaling update-demo-kitten up to 2\nScaling update-demo-nautilus down to 0\nUpdate succeeded. Deleting old controller: update-demo-nautilus\nRenaming update-demo-kitten to update-demo-nautilus\nreplicationcontroller/update-demo-nautilus rolling updated\n"
+STEP: waiting for all containers in name=update-demo pods to come up.
+Feb 12 10:34:41.749: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-1416'
+Feb 12 10:34:41.874: INFO: stderr: ""
+Feb 12 10:34:41.874: INFO: stdout: "update-demo-kitten-dv4db update-demo-kitten-zm4np "
+Feb 12 10:34:41.874: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-kitten-dv4db -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1416'
+Feb 12 10:34:41.963: INFO: stderr: ""
+Feb 12 10:34:41.963: INFO: stdout: "true"
+Feb 12 10:34:41.963: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-kitten-dv4db -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1416'
+Feb 12 10:34:42.040: INFO: stderr: ""
+Feb 12 10:34:42.040: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
+Feb 12 10:34:42.040: INFO: validating pod update-demo-kitten-dv4db
+Feb 12 10:34:42.142: INFO: got data: {
+  "image": "kitten.jpg"
+}
 
-STEP: deleting the pod
-[AfterEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:19:32.790: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-7569" for this suite.
+Feb 12 10:34:42.142: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
+Feb 12 10:34:42.142: INFO: update-demo-kitten-dv4db is verified up and running
+Feb 12 10:34:42.142: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-kitten-zm4np -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-1416'
+Feb 12 10:34:42.278: INFO: stderr: ""
+Feb 12 10:34:42.278: INFO: stdout: "true"
+Feb 12 10:34:42.278: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-kitten-zm4np -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-1416'
+Feb 12 10:34:42.369: INFO: stderr: ""
+Feb 12 10:34:42.369: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/kitten:1.0"
+Feb 12 10:34:42.369: INFO: validating pod update-demo-kitten-zm4np
+Feb 12 10:34:42.466: INFO: got data: {
+  "image": "kitten.jpg"
+}
 
-• [SLOW TEST:10.933 seconds]
-[sig-network] DNS
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should provide DNS for the cluster  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:34:42.466: INFO: Unmarshalled json jpg/img => {kitten.jpg} , expecting kitten.jpg .
+Feb 12 10:34:42.466: INFO: update-demo-kitten-zm4np is verified up and running
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:34:42.466: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-1416" for this suite.
+
+• [SLOW TEST:29.334 seconds]
+[sig-cli] Kubectl client
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+  Update Demo
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:323
+    should do a rolling update of a replication controller  [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] DNS should provide DNS for the cluster  [Conformance]","total":280,"completed":182,"skipped":2898,"failed":0}
-SSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-cli] Kubectl client Update Demo should do a rolling update of a replication controller  [Conformance]","total":280,"completed":162,"skipped":2640,"failed":0}
+SS
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  listing mutating webhooks should work [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] Watchers 
+  should be able to restart watching from the last resource version observed by the previous watch [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Watchers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:19:32.805: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-2355
+Feb 12 10:34:42.485: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename watch
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-6945
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Aug 30 17:19:33.271: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Aug 30 17:19:35.287: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404773, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404773, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404773, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404773, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Aug 30 17:19:38.306: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] listing mutating webhooks should work [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Listing all of the created validation webhooks
-STEP: Creating a configMap that should be mutated
-STEP: Deleting the collection of validation webhooks
-STEP: Creating a configMap that should not be mutated
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:19:38.816: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-2355" for this suite.
-STEP: Destroying namespace "webhook-2355-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
-
-• [SLOW TEST:6.109 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  listing mutating webhooks should work [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[It] should be able to restart watching from the last resource version observed by the previous watch [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating a watch on configmaps
+STEP: creating a new configmap
+STEP: modifying the configmap once
+STEP: closing the watch once it receives two notifications
+Feb 12 10:34:42.669: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed  watch-6945 /api/v1/namespaces/watch-6945/configmaps/e2e-watch-test-watch-closed 586c0a19-d062-416a-ada8-01fd710b3b68 25250 0 2021-02-12 10:34:42 +0000 UTC   map[watch-this-configmap:watch-closed-and-restarted] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Feb 12 10:34:42.670: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed  watch-6945 /api/v1/namespaces/watch-6945/configmaps/e2e-watch-test-watch-closed 586c0a19-d062-416a-ada8-01fd710b3b68 25251 0 2021-02-12 10:34:42 +0000 UTC   map[watch-this-configmap:watch-closed-and-restarted] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+STEP: modifying the configmap a second time, while the watch is closed
+STEP: creating a new watch on configmaps from the last resource version observed by the first watch
+STEP: deleting the configmap
+STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed
+Feb 12 10:34:42.700: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed  watch-6945 /api/v1/namespaces/watch-6945/configmaps/e2e-watch-test-watch-closed 586c0a19-d062-416a-ada8-01fd710b3b68 25252 0 2021-02-12 10:34:42 +0000 UTC   map[watch-this-configmap:watch-closed-and-restarted] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Feb 12 10:34:42.701: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed  watch-6945 /api/v1/namespaces/watch-6945/configmaps/e2e-watch-test-watch-closed 586c0a19-d062-416a-ada8-01fd710b3b68 25253 0 2021-02-12 10:34:42 +0000 UTC   map[watch-this-configmap:watch-closed-and-restarted] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:34:42.701: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-6945" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance]","total":280,"completed":163,"skipped":2642,"failed":0}
+SSSSSSSSSS
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance]","total":280,"completed":183,"skipped":2913,"failed":0}
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  should perform canary updates and phased rolling updates of template modifications [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Security Context when creating containers with AllowPrivilegeEscalation 
+  should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Security Context
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:19:38.916: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename statefulset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-766
+Feb 12 10:34:42.718: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename security-context-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in security-context-test-5799
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
-STEP: Creating service test in namespace statefulset-766
-[It] should perform canary updates and phased rolling updates of template modifications [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a new StatefulSet
-Aug 30 17:19:39.096: INFO: Found 0 stateful pods, waiting for 3
-Aug 30 17:19:49.104: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
-Aug 30 17:19:49.104: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
-Aug 30 17:19:49.104: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Updating stateful set template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine
-Aug 30 17:19:49.148: INFO: Updating stateful set ss2
-STEP: Creating a new revision
-STEP: Not applying an update when the partition is greater than the number of replicas
-STEP: Performing a canary update
-Aug 30 17:19:59.194: INFO: Updating stateful set ss2
-Aug 30 17:19:59.206: INFO: Waiting for Pod statefulset-766/ss2-2 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
-STEP: Restoring Pods to the correct revision when they are deleted
-Aug 30 17:20:09.314: INFO: Found 2 stateful pods, waiting for 3
-Aug 30 17:20:19.323: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
-Aug 30 17:20:19.323: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
-Aug 30 17:20:19.323: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Performing a phased rolling update
-Aug 30 17:20:19.358: INFO: Updating stateful set ss2
-Aug 30 17:20:19.371: INFO: Waiting for Pod statefulset-766/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
-Aug 30 17:20:29.388: INFO: Waiting for Pod statefulset-766/ss2-1 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
-Aug 30 17:20:39.406: INFO: Updating stateful set ss2
-Aug 30 17:20:39.418: INFO: Waiting for StatefulSet statefulset-766/ss2 to complete update
-Aug 30 17:20:39.418: INFO: Waiting for Pod statefulset-766/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
-Aug 30 17:20:49.432: INFO: Deleting all statefulset in ns statefulset-766
-Aug 30 17:20:49.437: INFO: Scaling statefulset ss2 to 0
-Aug 30 17:20:59.460: INFO: Waiting for statefulset status.replicas updated to 0
-Aug 30 17:20:59.467: INFO: Deleting statefulset ss2
-[AfterEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:20:59.489: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-766" for this suite.
-
-• [SLOW TEST:80.588 seconds]
-[sig-apps] StatefulSet
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-    should perform canary updates and phased rolling updates of template modifications [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance]","total":280,"completed":184,"skipped":2913,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] [k8s.io] Security Context
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39
+[It] should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:34:42.895: INFO: Waiting up to 5m0s for pod "alpine-nnp-false-6dc0f890-7c82-44e5-bce1-cdc7946e77b9" in namespace "security-context-test-5799" to be "success or failure"
+Feb 12 10:34:42.902: INFO: Pod "alpine-nnp-false-6dc0f890-7c82-44e5-bce1-cdc7946e77b9": Phase="Pending", Reason="", readiness=false. Elapsed: 7.013279ms
+Feb 12 10:34:44.908: INFO: Pod "alpine-nnp-false-6dc0f890-7c82-44e5-bce1-cdc7946e77b9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012846021s
+Feb 12 10:34:46.913: INFO: Pod "alpine-nnp-false-6dc0f890-7c82-44e5-bce1-cdc7946e77b9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018504585s
+Feb 12 10:34:46.914: INFO: Pod "alpine-nnp-false-6dc0f890-7c82-44e5-bce1-cdc7946e77b9" satisfied condition "success or failure"
+[AfterEach] [k8s.io] Security Context
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:34:46.934: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "security-context-test-5799" for this suite.
+•{"msg":"PASSED [k8s.io] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":164,"skipped":2652,"failed":0}
+SSSSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should provide container's cpu request [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] ConfigMap 
+  should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:20:59.507: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-4304
+Feb 12 10:34:46.953: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-1695
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should provide container's cpu request [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 17:20:59.695: INFO: Waiting up to 5m0s for pod "downwardapi-volume-bf1468dd-22bd-419f-8159-f41d23ae3c63" in namespace "projected-4304" to be "success or failure"
-Aug 30 17:20:59.699: INFO: Pod "downwardapi-volume-bf1468dd-22bd-419f-8159-f41d23ae3c63": Phase="Pending", Reason="", readiness=false. Elapsed: 4.289494ms
-Aug 30 17:21:01.705: INFO: Pod "downwardapi-volume-bf1468dd-22bd-419f-8159-f41d23ae3c63": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010243467s
+[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name configmap-test-volume-23f221e5-a2a5-4f80-a8e5-458781f96479
+STEP: Creating a pod to test consume configMaps
+Feb 12 10:34:47.147: INFO: Waiting up to 5m0s for pod "pod-configmaps-060148fd-1675-4958-bfdb-19b5606475e5" in namespace "configmap-1695" to be "success or failure"
+Feb 12 10:34:47.153: INFO: Pod "pod-configmaps-060148fd-1675-4958-bfdb-19b5606475e5": Phase="Pending", Reason="", readiness=false. Elapsed: 6.265138ms
+Feb 12 10:34:49.160: INFO: Pod "pod-configmaps-060148fd-1675-4958-bfdb-19b5606475e5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012349909s
+Feb 12 10:34:51.165: INFO: Pod "pod-configmaps-060148fd-1675-4958-bfdb-19b5606475e5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017790588s
 STEP: Saw pod success
-Aug 30 17:21:01.705: INFO: Pod "downwardapi-volume-bf1468dd-22bd-419f-8159-f41d23ae3c63" satisfied condition "success or failure"
-Aug 30 17:21:01.710: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-bf1468dd-22bd-419f-8159-f41d23ae3c63 container client-container: 
+Feb 12 10:34:51.165: INFO: Pod "pod-configmaps-060148fd-1675-4958-bfdb-19b5606475e5" satisfied condition "success or failure"
+Feb 12 10:34:51.171: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-configmaps-060148fd-1675-4958-bfdb-19b5606475e5 container configmap-volume-test: 
 STEP: delete the pod
-Aug 30 17:21:01.778: INFO: Waiting for pod downwardapi-volume-bf1468dd-22bd-419f-8159-f41d23ae3c63 to disappear
-Aug 30 17:21:01.782: INFO: Pod downwardapi-volume-bf1468dd-22bd-419f-8159-f41d23ae3c63 no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:01.783: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-4304" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance]","total":280,"completed":185,"skipped":2948,"failed":0}
-SSS
+Feb 12 10:34:51.207: INFO: Waiting for pod pod-configmaps-060148fd-1675-4958-bfdb-19b5606475e5 to disappear
+Feb 12 10:34:51.212: INFO: Pod pod-configmaps-060148fd-1675-4958-bfdb-19b5606475e5 no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:34:51.212: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-1695" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","total":280,"completed":165,"skipped":2657,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition 
-  getting/updating/patching custom resource definition status sub-resource works  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Probing container 
+  should have monotonically increasing restart count [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:21:01.800: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename custom-resource-definition
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-5789
+Feb 12 10:34:51.234: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-2508
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] getting/updating/patching custom resource definition status sub-resource works  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:21:01.962: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:02.525: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "custom-resource-definition-5789" for this suite.
-•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works  [Conformance]","total":280,"completed":186,"skipped":2951,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should have monotonically increasing restart count [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating pod liveness-be84109b-a30a-43ca-961d-2661068b439f in namespace container-probe-2508
+Feb 12 10:34:53.416: INFO: Started pod liveness-be84109b-a30a-43ca-961d-2661068b439f in namespace container-probe-2508
+STEP: checking the pod's current state and verifying that restartCount is present
+Feb 12 10:34:53.422: INFO: Initial restart count of pod liveness-be84109b-a30a-43ca-961d-2661068b439f is 0
+Feb 12 10:35:09.481: INFO: Restart count of pod container-probe-2508/liveness-be84109b-a30a-43ca-961d-2661068b439f is now 1 (16.059424411s elapsed)
+Feb 12 10:35:31.563: INFO: Restart count of pod container-probe-2508/liveness-be84109b-a30a-43ca-961d-2661068b439f is now 2 (38.140635525s elapsed)
+Feb 12 10:35:51.660: INFO: Restart count of pod container-probe-2508/liveness-be84109b-a30a-43ca-961d-2661068b439f is now 3 (58.238566815s elapsed)
+Feb 12 10:36:11.742: INFO: Restart count of pod container-probe-2508/liveness-be84109b-a30a-43ca-961d-2661068b439f is now 4 (1m18.319625734s elapsed)
+Feb 12 10:37:09.979: INFO: Restart count of pod container-probe-2508/liveness-be84109b-a30a-43ca-961d-2661068b439f is now 5 (2m16.557124157s elapsed)
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:37:10.005: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-2508" for this suite.
+
+• [SLOW TEST:138.796 seconds]
+[k8s.io] Probing container
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should have monotonically increasing restart count [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Secrets 
-  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [k8s.io] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance]","total":280,"completed":166,"skipped":2706,"failed":0}
+SS
+------------------------------
+[sig-storage] Projected downwardAPI 
+  should provide podname only [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:21:02.539: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-9014
+Feb 12 10:37:10.031: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-5438
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-49819cd2-3c1a-4c2f-9897-143652697abd
-STEP: Creating a pod to test consume secrets
-Aug 30 17:21:02.722: INFO: Waiting up to 5m0s for pod "pod-secrets-7b8e2088-f11c-445f-8f31-9f26bf43108a" in namespace "secrets-9014" to be "success or failure"
-Aug 30 17:21:02.727: INFO: Pod "pod-secrets-7b8e2088-f11c-445f-8f31-9f26bf43108a": Phase="Pending", Reason="", readiness=false. Elapsed: 5.049148ms
-Aug 30 17:21:04.733: INFO: Pod "pod-secrets-7b8e2088-f11c-445f-8f31-9f26bf43108a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011013358s
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should provide podname only [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Feb 12 10:37:10.215: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7d764f1c-8fa2-40f3-a4e9-ed019336d355" in namespace "projected-5438" to be "success or failure"
+Feb 12 10:37:10.220: INFO: Pod "downwardapi-volume-7d764f1c-8fa2-40f3-a4e9-ed019336d355": Phase="Pending", Reason="", readiness=false. Elapsed: 5.293034ms
+Feb 12 10:37:12.227: INFO: Pod "downwardapi-volume-7d764f1c-8fa2-40f3-a4e9-ed019336d355": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012353002s
+Feb 12 10:37:14.234: INFO: Pod "downwardapi-volume-7d764f1c-8fa2-40f3-a4e9-ed019336d355": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.019095321s
 STEP: Saw pod success
-Aug 30 17:21:04.733: INFO: Pod "pod-secrets-7b8e2088-f11c-445f-8f31-9f26bf43108a" satisfied condition "success or failure"
-Aug 30 17:21:04.740: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-secrets-7b8e2088-f11c-445f-8f31-9f26bf43108a container secret-volume-test: 
+Feb 12 10:37:14.234: INFO: Pod "downwardapi-volume-7d764f1c-8fa2-40f3-a4e9-ed019336d355" satisfied condition "success or failure"
+Feb 12 10:37:14.240: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-7d764f1c-8fa2-40f3-a4e9-ed019336d355 container client-container: 
 STEP: delete the pod
-Aug 30 17:21:04.809: INFO: Waiting for pod pod-secrets-7b8e2088-f11c-445f-8f31-9f26bf43108a to disappear
-Aug 30 17:21:04.815: INFO: Pod pod-secrets-7b8e2088-f11c-445f-8f31-9f26bf43108a no longer exists
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:04.815: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-9014" for this suite.
-•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":187,"skipped":3002,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Feb 12 10:37:14.313: INFO: Waiting for pod downwardapi-volume-7d764f1c-8fa2-40f3-a4e9-ed019336d355 to disappear
+Feb 12 10:37:14.318: INFO: Pod downwardapi-volume-7d764f1c-8fa2-40f3-a4e9-ed019336d355 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:37:14.318: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-5438" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance]","total":280,"completed":167,"skipped":2708,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl version 
-  should check is all data is printed  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] Namespaces [Serial] 
+  should ensure that all pods are removed when a namespace is deleted [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Namespaces [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:21:04.832: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-759
+Feb 12 10:37:14.339: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename namespaces
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in namespaces-7009
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[It] should check is all data is printed  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:21:05.025: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 version'
-Aug 30 17:21:05.083: INFO: stderr: ""
-Aug 30 17:21:05.083: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"17\", GitVersion:\"v1.17.9\", GitCommit:\"4fb7ed12476d57b8437ada90b4f93b17ffaeed99\", GitTreeState:\"clean\", BuildDate:\"2020-07-15T16:18:16Z\", GoVersion:\"go1.13.9\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nServer Version: version.Info{Major:\"1\", Minor:\"17\", GitVersion:\"v1.17.9\", GitCommit:\"4fb7ed12476d57b8437ada90b4f93b17ffaeed99\", GitTreeState:\"clean\", BuildDate:\"2020-07-15T16:10:45Z\", GoVersion:\"go1.13.9\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:05.083: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-759" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl version should check is all data is printed  [Conformance]","total":280,"completed":188,"skipped":3035,"failed":0}
-SSSSSSSSSSSSS
+[It] should ensure that all pods are removed when a namespace is deleted [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a test namespace
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-3260
+STEP: Waiting for a default service account to be provisioned in namespace
+STEP: Creating a pod in the namespace
+STEP: Waiting for the pod to have running status
+STEP: Deleting the namespace
+STEP: Waiting for the namespace to be removed.
+STEP: Recreating the namespace
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-6411
+STEP: Verifying there are no pods in the namespace
+[AfterEach] [sig-api-machinery] Namespaces [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:37:44.868: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "namespaces-7009" for this suite.
+STEP: Destroying namespace "nsdeletetest-3260" for this suite.
+Feb 12 10:37:44.896: INFO: Namespace nsdeletetest-3260 was already deleted
+STEP: Destroying namespace "nsdeletetest-6411" for this suite.
+
+• [SLOW TEST:30.566 seconds]
+[sig-api-machinery] Namespaces [Serial]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should ensure that all pods are removed when a namespace is deleted [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should be able to deny attaching pod [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance]","total":280,"completed":168,"skipped":2738,"failed":0}
+SSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial] 
+  evicts pods with minTolerationSeconds [Disruptive] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:21:05.100: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-8903
+Feb 12 10:37:44.906: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename taint-multiple-pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in taint-multiple-pods-6945
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Aug 30 17:21:05.475: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Aug 30 17:21:07.491: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404865, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404865, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404865, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734404865, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Aug 30 17:21:10.513: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should be able to deny attaching pod [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering the webhook via the AdmissionRegistration API
-STEP: create a pod
-STEP: 'kubectl attach' the pod, should be denied by the webhook
-Aug 30 17:21:12.720: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 attach --namespace=webhook-8903 to-be-attached-pod -i -c=container1'
-Aug 30 17:21:12.803: INFO: rc: 1
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:12.814: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-8903" for this suite.
-STEP: Destroying namespace "webhook-8903-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+[BeforeEach] [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/taints.go:344
+Feb 12 10:37:45.075: INFO: Waiting up to 1m0s for all nodes to be ready
+Feb 12 10:38:45.111: INFO: Waiting for terminating namespaces to be deleted...
+[It] evicts pods with minTolerationSeconds [Disruptive] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:38:45.119: INFO: Starting informer...
+STEP: Starting pods...
+Feb 12 10:38:45.346: INFO: Pod1 is running on nostalgic-fermat-cdc5d8777-xg74t. Tainting Node
+Feb 12 10:38:49.586: INFO: Pod2 is running on nostalgic-fermat-cdc5d8777-xg74t. Tainting Node
+STEP: Trying to apply a taint on the Node
+STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute
+STEP: Waiting for Pod1 and Pod2 to be deleted
+Feb 12 10:38:56.801: INFO: Noticed Pod "taint-eviction-b1" gets evicted.
+Feb 12 10:39:16.964: INFO: Noticed Pod "taint-eviction-b2" gets evicted.
+STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute
+[AfterEach] [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:39:16.986: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "taint-multiple-pods-6945" for this suite.
 
-• [SLOW TEST:7.802 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should be able to deny attaching pod [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:92.118 seconds]
+[sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
+  evicts pods with minTolerationSeconds [Disruptive] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance]","total":280,"completed":189,"skipped":3048,"failed":0}
-SSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-scheduling] NoExecuteTaintManager Multiple Pods [Serial] evicts pods with minTolerationSeconds [Disruptive] [Conformance]","total":280,"completed":169,"skipped":2758,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl cluster-info 
-  should check if Kubernetes master services is included in cluster-info  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] ReplicationController 
+  should adopt matching pods on creation [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] ReplicationController
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:21:12.902: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-5628
+Feb 12 10:39:17.031: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename replication-controller
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-1270
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[It] should check if Kubernetes master services is included in cluster-info  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: validating cluster-info
-Aug 30 17:21:13.060: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 cluster-info'
-Aug 30 17:21:13.129: INFO: stderr: ""
-Aug 30 17:21:13.129: INFO: stdout: "\x1b[0;32mKubernetes master\x1b[0m is running at \x1b[0;33mhttps://10.240.16.1:443\x1b[0m\n\x1b[0;32mkube-dns\x1b[0m is running at \x1b[0;33mhttps://10.240.16.1:443/api/v1/namespaces/kube-system/services/kube-dns:dns-tcp/proxy\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:13.129: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-5628" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes master services is included in cluster-info  [Conformance]","total":280,"completed":190,"skipped":3068,"failed":0}
-SSSSS
+[It] should adopt matching pods on creation [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Given a Pod with a 'name' label pod-adoption is created
+STEP: When a replication controller with a matching selector is created
+STEP: Then the orphan pod is adopted
+[AfterEach] [sig-apps] ReplicationController
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:39:22.264: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replication-controller-1270" for this suite.
+
+• [SLOW TEST:5.253 seconds]
+[sig-apps] ReplicationController
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should adopt matching pods on creation [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-apps] ReplicationController should adopt matching pods on creation [Conformance]","total":280,"completed":170,"skipped":2803,"failed":0}
+SSSSSSSSSSSSSSSS
 ------------------------------
 [sig-storage] Projected configMap 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:21:13.146: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:39:22.284: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-6994
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9664
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name projected-configmap-test-volume-map-e423e462-ca93-48d7-9ee5-b17bc303a653
+[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name projected-configmap-test-volume-0037f067-7706-4b74-ad5c-f7aa8fe41b18
 STEP: Creating a pod to test consume configMaps
-Aug 30 17:21:13.324: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-c123d2c0-21e2-46a8-a726-c1e2e763b978" in namespace "projected-6994" to be "success or failure"
-Aug 30 17:21:13.330: INFO: Pod "pod-projected-configmaps-c123d2c0-21e2-46a8-a726-c1e2e763b978": Phase="Pending", Reason="", readiness=false. Elapsed: 5.171141ms
-Aug 30 17:21:15.335: INFO: Pod "pod-projected-configmaps-c123d2c0-21e2-46a8-a726-c1e2e763b978": Phase="Running", Reason="", readiness=true. Elapsed: 2.010466664s
-Aug 30 17:21:17.341: INFO: Pod "pod-projected-configmaps-c123d2c0-21e2-46a8-a726-c1e2e763b978": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016386575s
+Feb 12 10:39:22.466: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-6e40f2a5-5583-46ab-9d24-bb4c5acfb2e6" in namespace "projected-9664" to be "success or failure"
+Feb 12 10:39:22.475: INFO: Pod "pod-projected-configmaps-6e40f2a5-5583-46ab-9d24-bb4c5acfb2e6": Phase="Pending", Reason="", readiness=false. Elapsed: 8.723094ms
+Feb 12 10:39:24.490: INFO: Pod "pod-projected-configmaps-6e40f2a5-5583-46ab-9d24-bb4c5acfb2e6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02364165s
+Feb 12 10:39:26.496: INFO: Pod "pod-projected-configmaps-6e40f2a5-5583-46ab-9d24-bb4c5acfb2e6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.030246057s
 STEP: Saw pod success
-Aug 30 17:21:17.341: INFO: Pod "pod-projected-configmaps-c123d2c0-21e2-46a8-a726-c1e2e763b978" satisfied condition "success or failure"
-Aug 30 17:21:17.345: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-projected-configmaps-c123d2c0-21e2-46a8-a726-c1e2e763b978 container projected-configmap-volume-test: 
+Feb 12 10:39:26.496: INFO: Pod "pod-projected-configmaps-6e40f2a5-5583-46ab-9d24-bb4c5acfb2e6" satisfied condition "success or failure"
+Feb 12 10:39:26.505: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-projected-configmaps-6e40f2a5-5583-46ab-9d24-bb4c5acfb2e6 container projected-configmap-volume-test: 
 STEP: delete the pod
-Aug 30 17:21:17.414: INFO: Waiting for pod pod-projected-configmaps-c123d2c0-21e2-46a8-a726-c1e2e763b978 to disappear
-Aug 30 17:21:17.419: INFO: Pod pod-projected-configmaps-c123d2c0-21e2-46a8-a726-c1e2e763b978 no longer exists
+Feb 12 10:39:37.002: INFO: Waiting for pod pod-projected-configmaps-6e40f2a5-5583-46ab-9d24-bb4c5acfb2e6 to disappear
+Feb 12 10:39:37.008: INFO: Pod pod-projected-configmaps-6e40f2a5-5583-46ab-9d24-bb4c5acfb2e6 no longer exists
 [AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:17.419: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-6994" for this suite.
-•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":191,"skipped":3073,"failed":0}
-SSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-network] Services 
-  should be able to change the type from ExternalName to ClusterIP [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:21:17.436: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename services
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-2605
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should be able to change the type from ExternalName to ClusterIP [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a service externalname-service with the type=ExternalName in namespace services-2605
-STEP: changing the ExternalName service to type=ClusterIP
-STEP: creating replication controller externalname-service in namespace services-2605
-I0830 17:21:17.667426      23 runners.go:189] Created replication controller with name: externalname-service, namespace: services-2605, replica count: 2
-Aug 30 17:21:20.717: INFO: Creating new exec pod
-I0830 17:21:20.717750      23 runners.go:189] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-Aug 30 17:21:25.737: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=services-2605 execpodcsgvc -- /bin/sh -x -c nc -zv -t -w 2 externalname-service 80'
-Aug 30 17:21:26.313: INFO: stderr: "+ nc -zv -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n"
-Aug 30 17:21:26.313: INFO: stdout: ""
-Aug 30 17:21:26.314: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=services-2605 execpodcsgvc -- /bin/sh -x -c nc -zv -t -w 2 10.240.26.73 80'
-Aug 30 17:21:26.852: INFO: stderr: "+ nc -zv -t -w 2 10.240.26.73 80\nConnection to 10.240.26.73 80 port [tcp/http] succeeded!\n"
-Aug 30 17:21:26.852: INFO: stdout: ""
-Aug 30 17:21:26.852: INFO: Cleaning up the ExternalName to ClusterIP test service
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:26.878: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-2605" for this suite.
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:39:37.008: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9664" for this suite.
 
-• [SLOW TEST:9.459 seconds]
-[sig-network] Services
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should be able to change the type from ExternalName to ClusterIP [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:14.750 seconds]
+[sig-storage] Projected configMap
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_configmap.go:34
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance]","total":280,"completed":192,"skipped":3094,"failed":0}
-S
+{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":171,"skipped":2819,"failed":0}
+SSSSSSSSSSSSSS
 ------------------------------
-[sig-network] DNS 
-  should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Security Context When creating a pod with privileged 
+  should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Security Context
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:21:26.895: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename dns
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-5910
+Feb 12 10:39:37.038: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename security-context-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in security-context-test-16
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a test headless service
-STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-5910 A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-5910;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-5910 A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-5910;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-5910.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-5910.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-5910.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-5910.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-5910.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-5910.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-5910.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-5910.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-5910.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-5910.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-5910.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-5910.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5910.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 109.21.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.21.109_udp@PTR;check="$$(dig +tcp +noall +answer +search 109.21.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.21.109_tcp@PTR;sleep 1; done
-
-STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-5910 A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-5910;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-5910 A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-5910;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-5910.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-5910.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-5910.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-5910.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-5910.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-5910.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-5910.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-5910.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-5910.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-5910.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-5910.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-5910.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-5910.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 109.21.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.21.109_udp@PTR;check="$$(dig +tcp +noall +answer +search 109.21.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.21.109_tcp@PTR;sleep 1; done
-
-STEP: creating a pod to probe DNS
-STEP: submitting the pod to kubernetes
-STEP: retrieving the pod
-STEP: looking for the results for each expected name from probers
-Aug 30 17:21:29.229: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.264: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.275: INFO: Unable to read wheezy_udp@dns-test-service.dns-5910 from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.320: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5910 from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.330: INFO: Unable to read wheezy_udp@dns-test-service.dns-5910.svc from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.339: INFO: Unable to read wheezy_tcp@dns-test-service.dns-5910.svc from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.349: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-5910.svc from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.359: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-5910.svc from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.885: INFO: Unable to read jessie_udp@dns-test-service from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.897: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.906: INFO: Unable to read jessie_udp@dns-test-service.dns-5910 from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.915: INFO: Unable to read jessie_tcp@dns-test-service.dns-5910 from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.923: INFO: Unable to read jessie_udp@dns-test-service.dns-5910.svc from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.933: INFO: Unable to read jessie_tcp@dns-test-service.dns-5910.svc from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.951: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-5910.svc from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:29.961: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-5910.svc from pod dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7: the server could not find the requested resource (get pods dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7)
-Aug 30 17:21:30.399: INFO: Lookups using dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-5910 wheezy_tcp@dns-test-service.dns-5910 wheezy_udp@dns-test-service.dns-5910.svc wheezy_tcp@dns-test-service.dns-5910.svc wheezy_udp@_http._tcp.dns-test-service.dns-5910.svc wheezy_tcp@_http._tcp.dns-test-service.dns-5910.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-5910 jessie_tcp@dns-test-service.dns-5910 jessie_udp@dns-test-service.dns-5910.svc jessie_tcp@dns-test-service.dns-5910.svc jessie_udp@_http._tcp.dns-test-service.dns-5910.svc jessie_tcp@_http._tcp.dns-test-service.dns-5910.svc]
-
-Aug 30 17:21:37.712: INFO: DNS probes using dns-5910/dns-test-7e453f47-16a6-40f0-86ba-a4deee87dfc7 succeeded
-
-STEP: deleting the pod
-STEP: deleting the test service
-STEP: deleting the test headless service
-[AfterEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:37.786: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-5910" for this suite.
-
-• [SLOW TEST:10.905 seconds]
-[sig-network] DNS
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]","total":280,"completed":193,"skipped":3095,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] [k8s.io] Security Context
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39
+[It] should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:39:37.274: INFO: Waiting up to 5m0s for pod "busybox-privileged-false-9152e8a6-99dd-4584-8b64-788e1d1096a3" in namespace "security-context-test-16" to be "success or failure"
+Feb 12 10:39:37.283: INFO: Pod "busybox-privileged-false-9152e8a6-99dd-4584-8b64-788e1d1096a3": Phase="Pending", Reason="", readiness=false. Elapsed: 8.884868ms
+Feb 12 10:39:39.289: INFO: Pod "busybox-privileged-false-9152e8a6-99dd-4584-8b64-788e1d1096a3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014754387s
+Feb 12 10:39:39.289: INFO: Pod "busybox-privileged-false-9152e8a6-99dd-4584-8b64-788e1d1096a3" satisfied condition "success or failure"
+Feb 12 10:39:39.340: INFO: Got logs for pod "busybox-privileged-false-9152e8a6-99dd-4584-8b64-788e1d1096a3": "ip: RTNETLINK answers: Operation not permitted\n"
+[AfterEach] [k8s.io] Security Context
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:39:39.340: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "security-context-test-16" for this suite.
+•{"msg":"PASSED [k8s.io] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":172,"skipped":2833,"failed":0}
+SSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Projected downwardAPI 
-  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Downward API volume 
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:21:37.802: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-7216
+Feb 12 10:39:39.361: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-5415
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Creating a pod to test downward API volume plugin
-Aug 30 17:21:37.983: INFO: Waiting up to 5m0s for pod "downwardapi-volume-06aa5a72-978f-434c-894d-313964f80ca3" in namespace "projected-7216" to be "success or failure"
-Aug 30 17:21:37.988: INFO: Pod "downwardapi-volume-06aa5a72-978f-434c-894d-313964f80ca3": Phase="Pending", Reason="", readiness=false. Elapsed: 4.534811ms
-Aug 30 17:21:39.994: INFO: Pod "downwardapi-volume-06aa5a72-978f-434c-894d-313964f80ca3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010293749s
-Aug 30 17:21:41.999: INFO: Pod "downwardapi-volume-06aa5a72-978f-434c-894d-313964f80ca3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01588502s
+Feb 12 10:39:39.570: INFO: Waiting up to 5m0s for pod "downwardapi-volume-640664d3-8937-4782-a994-35713f603554" in namespace "downward-api-5415" to be "success or failure"
+Feb 12 10:39:39.576: INFO: Pod "downwardapi-volume-640664d3-8937-4782-a994-35713f603554": Phase="Pending", Reason="", readiness=false. Elapsed: 6.176747ms
+Feb 12 10:39:41.583: INFO: Pod "downwardapi-volume-640664d3-8937-4782-a994-35713f603554": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.01307453s
 STEP: Saw pod success
-Aug 30 17:21:41.999: INFO: Pod "downwardapi-volume-06aa5a72-978f-434c-894d-313964f80ca3" satisfied condition "success or failure"
-Aug 30 17:21:42.004: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-06aa5a72-978f-434c-894d-313964f80ca3 container client-container: 
+Feb 12 10:39:41.583: INFO: Pod "downwardapi-volume-640664d3-8937-4782-a994-35713f603554" satisfied condition "success or failure"
+Feb 12 10:39:41.589: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-640664d3-8937-4782-a994-35713f603554 container client-container: 
 STEP: delete the pod
-Aug 30 17:21:42.072: INFO: Waiting for pod downwardapi-volume-06aa5a72-978f-434c-894d-313964f80ca3 to disappear
-Aug 30 17:21:42.076: INFO: Pod downwardapi-volume-06aa5a72-978f-434c-894d-313964f80ca3 no longer exists
-[AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:42.076: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-7216" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":194,"skipped":3117,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSS
+Feb 12 10:39:41.630: INFO: Waiting for pod downwardapi-volume-640664d3-8937-4782-a994-35713f603554 to disappear
+Feb 12 10:39:41.634: INFO: Pod downwardapi-volume-640664d3-8937-4782-a994-35713f603554 no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:39:41.634: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-5415" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance]","total":280,"completed":173,"skipped":2849,"failed":0}
+SSSSSSSS
 ------------------------------
-[sig-apps] ReplicaSet 
-  should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] ReplicaSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Downward API volume 
+  should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:21:42.095: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename replicaset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replicaset-4164
+Feb 12 10:39:41.655: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-5894
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:21:42.252: INFO: Creating ReplicaSet my-hostname-basic-7f79c213-99b1-4acc-b58c-55a8454da939
-Aug 30 17:21:42.264: INFO: Pod name my-hostname-basic-7f79c213-99b1-4acc-b58c-55a8454da939: Found 0 pods out of 1
-Aug 30 17:21:47.272: INFO: Pod name my-hostname-basic-7f79c213-99b1-4acc-b58c-55a8454da939: Found 1 pods out of 1
-Aug 30 17:21:47.272: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-7f79c213-99b1-4acc-b58c-55a8454da939" is running
-Aug 30 17:21:47.277: INFO: Pod "my-hostname-basic-7f79c213-99b1-4acc-b58c-55a8454da939-vfb42" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-08-30 17:21:42 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-08-30 17:21:44 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-08-30 17:21:44 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2020-08-30 17:21:42 +0000 UTC Reason: Message:}])
-Aug 30 17:21:47.277: INFO: Trying to dial the pod
-Aug 30 17:21:52.385: INFO: Controller my-hostname-basic-7f79c213-99b1-4acc-b58c-55a8454da939: Got expected result from replica 1 [my-hostname-basic-7f79c213-99b1-4acc-b58c-55a8454da939-vfb42]: "my-hostname-basic-7f79c213-99b1-4acc-b58c-55a8454da939-vfb42", 1 of 1 required successes so far
-[AfterEach] [sig-apps] ReplicaSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:52.385: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replicaset-4164" for this suite.
-
-• [SLOW TEST:10.306 seconds]
-[sig-apps] ReplicaSet
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should serve a basic image on each replica with a public image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] ReplicaSet should serve a basic image on each replica with a public image  [Conformance]","total":280,"completed":195,"skipped":3143,"failed":0}
-SSS
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should update annotations on modification [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating the pod
+Feb 12 10:39:44.408: INFO: Successfully updated pod "annotationupdate8609db05-9cce-415c-a13c-13642005b727"
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:39:46.451: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-5894" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance]","total":280,"completed":174,"skipped":2857,"failed":0}
+SSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Projected configMap 
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:21:52.401: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1667
+Feb 12 10:39:46.475: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8360
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0666 on tmpfs
-Aug 30 17:21:52.572: INFO: Waiting up to 5m0s for pod "pod-df88dd45-03a6-406a-ad93-0d33dcf2b5f4" in namespace "emptydir-1667" to be "success or failure"
-Aug 30 17:21:52.578: INFO: Pod "pod-df88dd45-03a6-406a-ad93-0d33dcf2b5f4": Phase="Pending", Reason="", readiness=false. Elapsed: 5.708736ms
-Aug 30 17:21:54.584: INFO: Pod "pod-df88dd45-03a6-406a-ad93-0d33dcf2b5f4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011099403s
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name projected-configmap-test-volume-fcc690b8-14b3-4989-b02f-1861ddead0e5
+STEP: Creating a pod to test consume configMaps
+Feb 12 10:39:46.661: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-f18d0a07-ce23-4824-9647-5fca47aa3ef0" in namespace "projected-8360" to be "success or failure"
+Feb 12 10:39:46.671: INFO: Pod "pod-projected-configmaps-f18d0a07-ce23-4824-9647-5fca47aa3ef0": Phase="Pending", Reason="", readiness=false. Elapsed: 9.288732ms
+Feb 12 10:39:48.678: INFO: Pod "pod-projected-configmaps-f18d0a07-ce23-4824-9647-5fca47aa3ef0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01681383s
+Feb 12 10:39:50.686: INFO: Pod "pod-projected-configmaps-f18d0a07-ce23-4824-9647-5fca47aa3ef0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.024216472s
 STEP: Saw pod success
-Aug 30 17:21:54.584: INFO: Pod "pod-df88dd45-03a6-406a-ad93-0d33dcf2b5f4" satisfied condition "success or failure"
-Aug 30 17:21:54.588: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-df88dd45-03a6-406a-ad93-0d33dcf2b5f4 container test-container: 
+Feb 12 10:39:50.686: INFO: Pod "pod-projected-configmaps-f18d0a07-ce23-4824-9647-5fca47aa3ef0" satisfied condition "success or failure"
+Feb 12 10:39:50.701: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-projected-configmaps-f18d0a07-ce23-4824-9647-5fca47aa3ef0 container projected-configmap-volume-test: 
 STEP: delete the pod
-Aug 30 17:21:54.657: INFO: Waiting for pod pod-df88dd45-03a6-406a-ad93-0d33dcf2b5f4 to disappear
-Aug 30 17:21:54.662: INFO: Pod pod-df88dd45-03a6-406a-ad93-0d33dcf2b5f4 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:54.662: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-1667" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":196,"skipped":3146,"failed":0}
-SSSSSSSSS
+Feb 12 10:39:50.746: INFO: Waiting for pod pod-projected-configmaps-f18d0a07-ce23-4824-9647-5fca47aa3ef0 to disappear
+Feb 12 10:39:50.751: INFO: Pod pod-projected-configmaps-f18d0a07-ce23-4824-9647-5fca47aa3ef0 no longer exists
+[AfterEach] [sig-storage] Projected configMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:39:50.751: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-8360" for this suite.
+•{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":175,"skipped":2864,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should mutate pod and apply defaults after mutation [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Probing container 
+  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:21:54.677: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-7323
+Feb 12 10:39:50.774: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-7454
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Aug 30 17:21:55.212: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Aug 30 17:21:58.239: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should mutate pod and apply defaults after mutation [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering the mutating pod webhook via the AdmissionRegistration API
-STEP: create a pod that should be updated by the webhook
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:21:58.533: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-7323" for this suite.
-STEP: Destroying namespace "webhook-7323-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
-•{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance]","total":280,"completed":197,"skipped":3155,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating pod busybox-3fec0305-ddf2-4ae2-89e4-273b81469a62 in namespace container-probe-7454
+Feb 12 10:39:52.970: INFO: Started pod busybox-3fec0305-ddf2-4ae2-89e4-273b81469a62 in namespace container-probe-7454
+STEP: checking the pod's current state and verifying that restartCount is present
+Feb 12 10:39:52.976: INFO: Initial restart count of pod busybox-3fec0305-ddf2-4ae2-89e4-273b81469a62 is 0
+Feb 12 10:40:41.188: INFO: Restart count of pod container-probe-7454/busybox-3fec0305-ddf2-4ae2-89e4-273b81469a62 is now 1 (48.211797579s elapsed)
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:40:41.208: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-7454" for this suite.
+
+• [SLOW TEST:50.454 seconds]
+[k8s.io] Probing container
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-node] Downward API 
-  should provide host IP as an env var [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] Downward API
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:21:58.618: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-4492
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide host IP as an env var [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward api env vars
-Aug 30 17:21:58.798: INFO: Waiting up to 5m0s for pod "downward-api-b442d4c5-f0ae-4f09-8593-0930ce32cee6" in namespace "downward-api-4492" to be "success or failure"
-Aug 30 17:21:58.810: INFO: Pod "downward-api-b442d4c5-f0ae-4f09-8593-0930ce32cee6": Phase="Pending", Reason="", readiness=false. Elapsed: 11.614199ms
-Aug 30 17:22:00.816: INFO: Pod "downward-api-b442d4c5-f0ae-4f09-8593-0930ce32cee6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018269028s
-Aug 30 17:22:02.822: INFO: Pod "downward-api-b442d4c5-f0ae-4f09-8593-0930ce32cee6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023758077s
-STEP: Saw pod success
-Aug 30 17:22:02.822: INFO: Pod "downward-api-b442d4c5-f0ae-4f09-8593-0930ce32cee6" satisfied condition "success or failure"
-Aug 30 17:22:02.827: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downward-api-b442d4c5-f0ae-4f09-8593-0930ce32cee6 container dapi-container: 
-STEP: delete the pod
-Aug 30 17:22:02.899: INFO: Waiting for pod downward-api-b442d4c5-f0ae-4f09-8593-0930ce32cee6 to disappear
-Aug 30 17:22:02.904: INFO: Pod downward-api-b442d4c5-f0ae-4f09-8593-0930ce32cee6 no longer exists
-[AfterEach] [sig-node] Downward API
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:22:02.904: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-4492" for this suite.
-•{"msg":"PASSED [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance]","total":280,"completed":198,"skipped":3183,"failed":0}
-S
+{"msg":"PASSED [k8s.io] Probing container should be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":280,"completed":176,"skipped":2905,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-storage] EmptyDir volumes 
-  should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:22:02.918: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:40:41.229: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-8223
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1483
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0666 on tmpfs
-Aug 30 17:22:03.091: INFO: Waiting up to 5m0s for pod "pod-a88ef8f2-fe21-4192-8c65-b798abf50f1e" in namespace "emptydir-8223" to be "success or failure"
-Aug 30 17:22:03.103: INFO: Pod "pod-a88ef8f2-fe21-4192-8c65-b798abf50f1e": Phase="Pending", Reason="", readiness=false. Elapsed: 11.892351ms
-Aug 30 17:22:05.109: INFO: Pod "pod-a88ef8f2-fe21-4192-8c65-b798abf50f1e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.017407432s
+[It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir volume type on node default medium
+Feb 12 10:40:41.404: INFO: Waiting up to 5m0s for pod "pod-e0a05207-4ec5-49ae-b9c1-c07a84667bc6" in namespace "emptydir-1483" to be "success or failure"
+Feb 12 10:40:41.412: INFO: Pod "pod-e0a05207-4ec5-49ae-b9c1-c07a84667bc6": Phase="Pending", Reason="", readiness=false. Elapsed: 8.130455ms
+Feb 12 10:40:43.420: INFO: Pod "pod-e0a05207-4ec5-49ae-b9c1-c07a84667bc6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016233358s
 STEP: Saw pod success
-Aug 30 17:22:05.109: INFO: Pod "pod-a88ef8f2-fe21-4192-8c65-b798abf50f1e" satisfied condition "success or failure"
-Aug 30 17:22:05.114: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-a88ef8f2-fe21-4192-8c65-b798abf50f1e container test-container: 
+Feb 12 10:40:43.420: INFO: Pod "pod-e0a05207-4ec5-49ae-b9c1-c07a84667bc6" satisfied condition "success or failure"
+Feb 12 10:40:43.428: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-e0a05207-4ec5-49ae-b9c1-c07a84667bc6 container test-container: 
 STEP: delete the pod
-Aug 30 17:22:05.180: INFO: Waiting for pod pod-a88ef8f2-fe21-4192-8c65-b798abf50f1e to disappear
-Aug 30 17:22:05.184: INFO: Pod pod-a88ef8f2-fe21-4192-8c65-b798abf50f1e no longer exists
+Feb 12 10:40:43.505: INFO: Waiting for pod pod-e0a05207-4ec5-49ae-b9c1-c07a84667bc6 to disappear
+Feb 12 10:40:43.511: INFO: Pod pod-e0a05207-4ec5-49ae-b9c1-c07a84667bc6 no longer exists
 [AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:22:05.184: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-8223" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":199,"skipped":3184,"failed":0}
-SSSSSS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:40:43.511: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-1483" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":177,"skipped":2944,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] Job 
-  should delete a job [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Job
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  removes definition from spec when one version gets changed to not be served [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:22:05.206: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename job
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in job-9495
+Feb 12 10:40:43.534: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename crd-publish-openapi
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-5322
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should delete a job [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a job
-STEP: Ensuring active pods == parallelism
-STEP: delete a job
-STEP: deleting Job.batch foo in namespace job-9495, will wait for the garbage collector to delete the pods
-Aug 30 17:22:09.450: INFO: Deleting Job.batch foo took: 11.430463ms
-Aug 30 17:22:09.950: INFO: Terminating Job.batch foo pods took: 500.322356ms
-STEP: Ensuring job was deleted
-[AfterEach] [sig-apps] Job
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:22:54.355: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "job-9495" for this suite.
+[It] removes definition from spec when one version gets changed to not be served [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: set up a multi version CRD
+Feb 12 10:40:43.702: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: mark a version not serverd
+STEP: check the unserved version gets removed
+STEP: check the other version is not changed
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:40:59.349: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-5322" for this suite.
 
-• [SLOW TEST:49.165 seconds]
-[sig-apps] Job
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should delete a job [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:15.836 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  removes definition from spec when one version gets changed to not be served [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] Job should delete a job [Conformance]","total":280,"completed":200,"skipped":3190,"failed":0}
-[sig-cli] Kubectl client Kubectl describe 
-  should check if kubectl describe prints relevant information for rc and pods  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:22:54.371: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-1432
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[It] should check if kubectl describe prints relevant information for rc and pods  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:22:54.528: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-1432'
-Aug 30 17:22:54.751: INFO: stderr: ""
-Aug 30 17:22:54.751: INFO: stdout: "replicationcontroller/agnhost-master created\n"
-Aug 30 17:22:54.751: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-1432'
-Aug 30 17:22:54.983: INFO: stderr: ""
-Aug 30 17:22:54.983: INFO: stdout: "service/agnhost-master created\n"
-STEP: Waiting for Agnhost master to start.
-Aug 30 17:22:55.988: INFO: Selector matched 1 pods for map[app:agnhost]
-Aug 30 17:22:55.988: INFO: Found 0 / 1
-Aug 30 17:22:56.989: INFO: Selector matched 1 pods for map[app:agnhost]
-Aug 30 17:22:56.989: INFO: Found 1 / 1
-Aug 30 17:22:56.989: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
-Aug 30 17:22:56.995: INFO: Selector matched 1 pods for map[app:agnhost]
-Aug 30 17:22:56.995: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
-Aug 30 17:22:56.995: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 describe pod agnhost-master-78g9p --namespace=kubectl-1432'
-Aug 30 17:22:57.079: INFO: stderr: ""
-Aug 30 17:22:57.079: INFO: stdout: "Name:         agnhost-master-78g9p\nNamespace:    kubectl-1432\nPriority:     0\nNode:         adoring-wozniak-54dcfd79fc-948mf/46.101.135.210\nStart Time:   Sun, 30 Aug 2020 17:22:54 +0000\nLabels:       app=agnhost\n              role=master\nAnnotations:  cni.projectcalico.org/podIP: 172.25.0.209/32\nStatus:       Running\nIP:           172.25.0.209\nIPs:\n  IP:           172.25.0.209\nControlled By:  ReplicationController/agnhost-master\nContainers:\n  agnhost-master:\n    Container ID:   docker://a726f34e47631cb65ddcee9cc9da17c8bd42bf48c83d16a2d2fddb11e369b046\n    Image:          gcr.io/kubernetes-e2e-test-images/agnhost:2.8\n    Image ID:       docker-pullable://gcr.io/kubernetes-e2e-test-images/agnhost@sha256:daf5332100521b1256d0e3c56d697a238eaec3af48897ed9167cbadd426773b5\n    Port:           6379/TCP\n    Host Port:      0/TCP\n    State:          Running\n      Started:      Sun, 30 Aug 2020 17:22:55 +0000\n    Ready:          True\n    Restart Count:  0\n    Environment:    \n    Mounts:\n      /var/run/secrets/kubernetes.io/serviceaccount from default-token-ndhbh (ro)\nConditions:\n  Type              Status\n  Initialized       True \n  Ready             True \n  ContainersReady   True \n  PodScheduled      True \nVolumes:\n  default-token-ndhbh:\n    Type:        Secret (a volume populated by a Secret)\n    SecretName:  default-token-ndhbh\n    Optional:    false\nQoS Class:       BestEffort\nNode-Selectors:  \nTolerations:     node.kubernetes.io/not-ready:NoExecute for 300s\n                 node.kubernetes.io/unreachable:NoExecute for 300s\nEvents:\n  Type    Reason     Age        From                                       Message\n  ----    ------     ----       ----                                       -------\n  Normal  Scheduled    default-scheduler                          Successfully assigned kubectl-1432/agnhost-master-78g9p to adoring-wozniak-54dcfd79fc-948mf\n  Normal  Pulled     2s         kubelet, adoring-wozniak-54dcfd79fc-948mf  Container image \"gcr.io/kubernetes-e2e-test-images/agnhost:2.8\" already present on machine\n  Normal  Created    2s         kubelet, adoring-wozniak-54dcfd79fc-948mf  Created container agnhost-master\n  Normal  Started    1s         kubelet, adoring-wozniak-54dcfd79fc-948mf  Started container agnhost-master\n"
-Aug 30 17:22:57.079: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 describe rc agnhost-master --namespace=kubectl-1432'
-Aug 30 17:22:57.168: INFO: stderr: ""
-Aug 30 17:22:57.168: INFO: stdout: "Name:         agnhost-master\nNamespace:    kubectl-1432\nSelector:     app=agnhost,role=master\nLabels:       app=agnhost\n              role=master\nAnnotations:  \nReplicas:     1 current / 1 desired\nPods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n  Labels:  app=agnhost\n           role=master\n  Containers:\n   agnhost-master:\n    Image:        gcr.io/kubernetes-e2e-test-images/agnhost:2.8\n    Port:         6379/TCP\n    Host Port:    0/TCP\n    Environment:  \n    Mounts:       \n  Volumes:        \nEvents:\n  Type    Reason            Age   From                    Message\n  ----    ------            ----  ----                    -------\n  Normal  SuccessfulCreate  3s    replication-controller  Created pod: agnhost-master-78g9p\n"
-Aug 30 17:22:57.168: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 describe service agnhost-master --namespace=kubectl-1432'
-Aug 30 17:22:57.248: INFO: stderr: ""
-Aug 30 17:22:57.248: INFO: stdout: "Name:              agnhost-master\nNamespace:         kubectl-1432\nLabels:            app=agnhost\n                   role=master\nAnnotations:       \nSelector:          app=agnhost,role=master\nType:              ClusterIP\nIP:                10.240.18.223\nPort:                6379/TCP\nTargetPort:        agnhost-server/TCP\nEndpoints:         172.25.0.209:6379\nSession Affinity:  None\nEvents:            \n"
-Aug 30 17:22:57.257: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 describe node adoring-wozniak-54dcfd79fc-6rshr'
-Aug 30 17:22:57.362: INFO: stderr: ""
-Aug 30 17:22:57.362: INFO: stdout: "Name:               adoring-wozniak-54dcfd79fc-6rshr\nRoles:              \nLabels:             beta.kubernetes.io/arch=amd64\n                    beta.kubernetes.io/os=linux\n                    kubernetes.io/arch=amd64\n                    kubernetes.io/hostname=adoring-wozniak-54dcfd79fc-6rshr\n                    kubernetes.io/os=linux\n                    machine-controller/owned-by=40ce4e72-1841-4057-a4ec-16381070e5bf\n                    system/cluster=x2pxtrxdh8\n                    system/project=kr6n49b5rm\n                    x-kubernetes.io/distribution=ubuntu\nAnnotations:        cluster.k8s.io/machine: kube-system/adoring-wozniak-54dcfd79fc-6rshr\n                    flannel.alpha.coreos.com/backend-data: {\"VtepMAC\":\"b2:d8:4e:ac:b9:a7\"}\n                    flannel.alpha.coreos.com/backend-type: vxlan\n                    flannel.alpha.coreos.com/kube-subnet-manager: true\n                    flannel.alpha.coreos.com/public-ip: 46.101.153.64\n                    node.alpha.kubernetes.io/ttl: 0\n                    projectcalico.org/IPv4IPIPTunnelAddr: 172.25.1.1\n                    volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp:  Sun, 30 Aug 2020 16:28:14 +0000\nTaints:             \nUnschedulable:      false\nLease:\n  HolderIdentity:  adoring-wozniak-54dcfd79fc-6rshr\n  AcquireTime:     \n  RenewTime:       Sun, 30 Aug 2020 17:22:55 +0000\nConditions:\n  Type             Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message\n  ----             ------  -----------------                 ------------------                ------                       -------\n  MemoryPressure   False   Sun, 30 Aug 2020 17:21:06 +0000   Sun, 30 Aug 2020 16:28:14 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available\n  DiskPressure     False   Sun, 30 Aug 2020 17:21:06 +0000   Sun, 30 Aug 2020 16:28:14 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure\n  PIDPressure      False   Sun, 30 Aug 2020 17:21:06 +0000   Sun, 30 Aug 2020 16:28:14 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available\n  Ready            True    Sun, 30 Aug 2020 17:21:06 +0000   Sun, 30 Aug 2020 16:28:34 +0000   KubeletReady                 kubelet is posting ready status. AppArmor enabled\nAddresses:\n  InternalIP:  46.101.153.64\n  Hostname:    adoring-wozniak-54dcfd79fc-6rshr\nCapacity:\n  cpu:                4\n  ephemeral-storage:  50633164Ki\n  hugepages-1Gi:      0\n  hugepages-2Mi:      0\n  memory:             8167684Ki\n  pods:               110\nAllocatable:\n  cpu:                3800m\n  ephemeral-storage:  44516040218\n  hugepages-1Gi:      0\n  hugepages-2Mi:      0\n  memory:             7860484Ki\n  pods:               110\nSystem Info:\n  Machine ID:                 810db956215e4e3795031a26bde87471\n  System UUID:                810DB956-215E-4E37-9503-1A26BDE87471\n  Boot ID:                    6faecf09-95fe-4ae9-9349-7ebfef62a596\n  Kernel Version:             4.15.0-112-generic\n  OS Image:                   Ubuntu 18.04.5 LTS\n  Operating System:           linux\n  Architecture:               amd64\n  Container Runtime Version:  docker://18.9.9\n  Kubelet Version:            v1.17.9\n  Kube-Proxy Version:         v1.17.9\nPodCIDR:                      172.25.1.0/24\nPodCIDRs:                     172.25.1.0/24\nNon-terminated Pods:          (13 in total)\n  Namespace                   Name                                                       CPU Requests  CPU Limits  Memory Requests  Memory Limits  AGE\n  ---------                   ----                                                       ------------  ----------  ---------------  -------------  ---\n  kube-system                 canal-54glj                                                250m (6%)     0 (0%)      0 (0%)           0 (0%)         54m\n  kube-system                 coredns-54457d966b-6cw7g                                   50m (1%)      100m (2%)   32Mi (0%)        64Mi (0%)      4m51s\n  kube-system                 coredns-54457d966b-fbnz4                                   50m (1%)      100m (2%)   32Mi (0%)        64Mi (0%)      56m\n  kube-system                 kube-proxy-wxdxv                                           75m (1%)      250m (6%)   50Mi (0%)        250Mi (3%)     54m\n  kube-system                 logrotate-57bmz                                            75m (1%)      250m (6%)   50Mi (0%)        250Mi (3%)     54m\n  kube-system                 node-local-dns-cvxnf                                       0 (0%)        0 (0%)      0 (0%)           0 (0%)         54m\n  kube-system                 openvpn-client-78d595f58b-vpgdr                            30m (0%)      200m (5%)   30Mi (0%)        82Mi (1%)      4m51s\n  kube-system                 user-ssh-keys-agent-xwrzj                                  0 (0%)        0 (0%)      0 (0%)           0 (0%)         54m\n  kubernetes-dashboard        dashboard-metrics-scraper-59bfc65dc9-252k4                 50m (1%)      100m (2%)   32Mi (0%)        64Mi (0%)      4m51s\n  kubernetes-dashboard        dashboard-metrics-scraper-59bfc65dc9-t82kj                 50m (1%)      100m (2%)   32Mi (0%)        64Mi (0%)      4m51s\n  sonobuoy                    sonobuoy                                                   0 (0%)        0 (0%)      0 (0%)           0 (0%)         53m\n  sonobuoy                    sonobuoy-e2e-job-cf49606f646f4c8a                          0 (0%)        0 (0%)      0 (0%)           0 (0%)         53m\n  sonobuoy                    sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-nd5nc    0 (0%)        0 (0%)      0 (0%)           0 (0%)         53m\nAllocated resources:\n  (Total limits may be over 100 percent, i.e., overcommitted.)\n  Resource           Requests    Limits\n  --------           --------    ------\n  cpu                630m (16%)  1100m (28%)\n  memory             258Mi (3%)  838Mi (10%)\n  ephemeral-storage  0 (0%)      0 (0%)\nEvents:\n  Type    Reason                 Age                From                                          Message\n  ----    ------                 ----               ----                                          -------\n  Normal  NodeHasNoDiskPressure  54m (x8 over 54m)  kubelet, adoring-wozniak-54dcfd79fc-6rshr     Node adoring-wozniak-54dcfd79fc-6rshr status is now: NodeHasNoDiskPressure\n  Normal  NodeHasSufficientPID   54m (x8 over 54m)  kubelet, adoring-wozniak-54dcfd79fc-6rshr     Node adoring-wozniak-54dcfd79fc-6rshr status is now: NodeHasSufficientPID\n  Normal  Starting               54m                kube-proxy, adoring-wozniak-54dcfd79fc-6rshr  Starting kube-proxy.\n"
-Aug 30 17:22:57.362: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 describe namespace kubectl-1432'
-Aug 30 17:22:57.444: INFO: stderr: ""
-Aug 30 17:22:57.444: INFO: stdout: "Name:         kubectl-1432\nLabels:       e2e-framework=kubectl\n              e2e-run=eda2017a-4ba1-4abb-bd48-e853704860b3\nAnnotations:  \nStatus:       Active\n\nNo resource quota.\n\nNo LimitRange resource.\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:22:57.444: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-1432" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods  [Conformance]","total":280,"completed":201,"skipped":3190,"failed":0}
-SSSS
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance]","total":280,"completed":178,"skipped":2966,"failed":0}
+SSS
 ------------------------------
 [sig-cli] Kubectl client Update Demo 
   should create and stop a replication controller  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:22:57.461: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:40:59.370: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-9559
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-4940
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
 [BeforeEach] Update Demo
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:324
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:325
 [It] should create and stop a replication controller  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: creating a replication controller
-Aug 30 17:22:57.634: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-9559'
-Aug 30 17:22:57.797: INFO: stderr: ""
-Aug 30 17:22:57.797: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
+Feb 12 10:40:59.526: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-4940'
+Feb 12 10:40:59.977: INFO: stderr: ""
+Feb 12 10:40:59.977: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
 STEP: waiting for all containers in name=update-demo pods to come up.
-Aug 30 17:22:57.797: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9559'
-Aug 30 17:22:57.885: INFO: stderr: ""
-Aug 30 17:22:57.885: INFO: stdout: "update-demo-nautilus-6wf9z update-demo-nautilus-hwwwv "
-Aug 30 17:22:57.885: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-6wf9z -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9559'
-Aug 30 17:22:57.968: INFO: stderr: ""
-Aug 30 17:22:57.968: INFO: stdout: ""
-Aug 30 17:22:57.968: INFO: update-demo-nautilus-6wf9z is created but not running
-Aug 30 17:23:02.969: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-9559'
-Aug 30 17:23:03.037: INFO: stderr: ""
-Aug 30 17:23:03.037: INFO: stdout: "update-demo-nautilus-6wf9z update-demo-nautilus-hwwwv "
-Aug 30 17:23:03.037: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-6wf9z -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9559'
-Aug 30 17:23:03.107: INFO: stderr: ""
-Aug 30 17:23:03.107: INFO: stdout: "true"
-Aug 30 17:23:03.107: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-6wf9z -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9559'
-Aug 30 17:23:03.174: INFO: stderr: ""
-Aug 30 17:23:03.174: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Aug 30 17:23:03.174: INFO: validating pod update-demo-nautilus-6wf9z
-Aug 30 17:23:03.306: INFO: got data: {
+Feb 12 10:40:59.977: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4940'
+Feb 12 10:41:00.089: INFO: stderr: ""
+Feb 12 10:41:00.089: INFO: stdout: "update-demo-nautilus-58f6d update-demo-nautilus-7td9j "
+Feb 12 10:41:00.090: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-58f6d -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4940'
+Feb 12 10:41:00.190: INFO: stderr: ""
+Feb 12 10:41:00.190: INFO: stdout: ""
+Feb 12 10:41:00.190: INFO: update-demo-nautilus-58f6d is created but not running
+Feb 12 10:41:05.190: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-4940'
+Feb 12 10:41:05.285: INFO: stderr: ""
+Feb 12 10:41:05.285: INFO: stdout: "update-demo-nautilus-58f6d update-demo-nautilus-7td9j "
+Feb 12 10:41:05.286: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-58f6d -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4940'
+Feb 12 10:41:05.382: INFO: stderr: ""
+Feb 12 10:41:05.382: INFO: stdout: "true"
+Feb 12 10:41:05.382: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-58f6d -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4940'
+Feb 12 10:41:05.470: INFO: stderr: ""
+Feb 12 10:41:05.471: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Feb 12 10:41:05.471: INFO: validating pod update-demo-nautilus-58f6d
+Feb 12 10:41:05.509: INFO: got data: {
   "image": "nautilus.jpg"
 }
 
-Aug 30 17:23:03.306: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Aug 30 17:23:03.306: INFO: update-demo-nautilus-6wf9z is verified up and running
-Aug 30 17:23:03.306: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-hwwwv -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-9559'
-Aug 30 17:23:03.371: INFO: stderr: ""
-Aug 30 17:23:03.371: INFO: stdout: "true"
-Aug 30 17:23:03.371: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-hwwwv -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-9559'
-Aug 30 17:23:03.432: INFO: stderr: ""
-Aug 30 17:23:03.432: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Aug 30 17:23:03.432: INFO: validating pod update-demo-nautilus-hwwwv
-Aug 30 17:23:03.564: INFO: got data: {
+Feb 12 10:41:05.509: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Feb 12 10:41:05.509: INFO: update-demo-nautilus-58f6d is verified up and running
+Feb 12 10:41:05.510: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-7td9j -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-4940'
+Feb 12 10:41:05.588: INFO: stderr: ""
+Feb 12 10:41:05.588: INFO: stdout: "true"
+Feb 12 10:41:05.588: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods update-demo-nautilus-7td9j -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-4940'
+Feb 12 10:41:05.664: INFO: stderr: ""
+Feb 12 10:41:05.664: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
+Feb 12 10:41:05.664: INFO: validating pod update-demo-nautilus-7td9j
+Feb 12 10:41:05.776: INFO: got data: {
   "image": "nautilus.jpg"
 }
 
-Aug 30 17:23:03.565: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Aug 30 17:23:03.565: INFO: update-demo-nautilus-hwwwv is verified up and running
+Feb 12 10:41:05.776: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
+Feb 12 10:41:05.776: INFO: update-demo-nautilus-7td9j is verified up and running
 STEP: using delete to clean up resources
-Aug 30 17:23:03.565: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete --grace-period=0 --force -f - --namespace=kubectl-9559'
-Aug 30 17:23:03.634: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Aug 30 17:23:03.634: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n"
-Aug 30 17:23:03.634: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-9559'
-Aug 30 17:23:03.701: INFO: stderr: "No resources found in kubectl-9559 namespace.\n"
-Aug 30 17:23:03.701: INFO: stdout: ""
-Aug 30 17:23:03.701: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -l name=update-demo --namespace=kubectl-9559 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
-Aug 30 17:23:03.765: INFO: stderr: ""
-Aug 30 17:23:03.765: INFO: stdout: "update-demo-nautilus-6wf9z\nupdate-demo-nautilus-hwwwv\n"
-Aug 30 17:23:04.266: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-9559'
-Aug 30 17:23:04.352: INFO: stderr: "No resources found in kubectl-9559 namespace.\n"
-Aug 30 17:23:04.352: INFO: stdout: ""
-Aug 30 17:23:04.352: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -l name=update-demo --namespace=kubectl-9559 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
-Aug 30 17:23:04.440: INFO: stderr: ""
-Aug 30 17:23:04.440: INFO: stdout: ""
+Feb 12 10:41:05.776: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete --grace-period=0 --force -f - --namespace=kubectl-4940'
+Feb 12 10:41:05.861: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Feb 12 10:41:05.861: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n"
+Feb 12 10:41:05.861: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-4940'
+Feb 12 10:41:05.942: INFO: stderr: "No resources found in kubectl-4940 namespace.\n"
+Feb 12 10:41:05.942: INFO: stdout: ""
+Feb 12 10:41:05.942: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -l name=update-demo --namespace=kubectl-4940 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Feb 12 10:41:06.017: INFO: stderr: ""
+Feb 12 10:41:06.017: INFO: stdout: "update-demo-nautilus-58f6d\nupdate-demo-nautilus-7td9j\n"
+Feb 12 10:41:06.518: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-4940'
+Feb 12 10:41:06.617: INFO: stderr: "No resources found in kubectl-4940 namespace.\n"
+Feb 12 10:41:06.617: INFO: stdout: ""
+Feb 12 10:41:06.617: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -l name=update-demo --namespace=kubectl-4940 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Feb 12 10:41:06.691: INFO: stderr: ""
+Feb 12 10:41:06.691: INFO: stdout: ""
 [AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:23:04.440: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-9559" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:41:06.691: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-4940" for this suite.
 
-• [SLOW TEST:6.995 seconds]
+• [SLOW TEST:7.350 seconds]
 [sig-cli] Kubectl client
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
   Update Demo
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:322
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:323
     should create and stop a replication controller  [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Update Demo should create and stop a replication controller  [Conformance]","total":280,"completed":202,"skipped":3194,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-cli] Kubectl client Update Demo should create and stop a replication controller  [Conformance]","total":280,"completed":179,"skipped":2969,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
-  works for CRD preserving unknown fields at the schema root [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] 
+  should be able to convert a non homogeneous list of CRs [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:23:04.458: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename crd-publish-openapi
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-6074
+Feb 12 10:41:06.721: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename crd-webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-webhook-9928
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] works for CRD preserving unknown fields at the schema root [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:23:04.623: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: client-side validation (kubectl create and apply) allows request with any unknown properties
-Aug 30 17:23:07.479: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-6074 create -f -'
-Aug 30 17:23:07.892: INFO: stderr: ""
-Aug 30 17:23:07.892: INFO: stdout: "e2e-test-crd-publish-openapi-8313-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n"
-Aug 30 17:23:07.892: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-6074 delete e2e-test-crd-publish-openapi-8313-crds test-cr'
-Aug 30 17:23:08.022: INFO: stderr: ""
-Aug 30 17:23:08.022: INFO: stdout: "e2e-test-crd-publish-openapi-8313-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n"
-Aug 30 17:23:08.022: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-6074 apply -f -'
-Aug 30 17:23:08.287: INFO: stderr: ""
-Aug 30 17:23:08.287: INFO: stdout: "e2e-test-crd-publish-openapi-8313-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n"
-Aug 30 17:23:08.287: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=crd-publish-openapi-6074 delete e2e-test-crd-publish-openapi-8313-crds test-cr'
-Aug 30 17:23:08.390: INFO: stderr: ""
-Aug 30 17:23:08.390: INFO: stdout: "e2e-test-crd-publish-openapi-8313-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n"
-STEP: kubectl explain works to explain CR
-Aug 30 17:23:08.390: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 explain e2e-test-crd-publish-openapi-8313-crds'
-Aug 30 17:23:08.627: INFO: stderr: ""
-Aug 30 17:23:08.627: INFO: stdout: "KIND:     E2e-test-crd-publish-openapi-8313-crd\nVERSION:  crd-publish-openapi-test-unknown-at-root.example.com/v1\n\nDESCRIPTION:\n     \n"
-[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:23:11.979: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-publish-openapi-6074" for this suite.
+[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:125
+STEP: Setting up server cert
+STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication
+STEP: Deploying the custom resource conversion webhook pod
+STEP: Wait for the deployment to be ready
+Feb 12 10:41:07.256: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set
+Feb 12 10:41:09.279: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723266, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723266, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723266, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723266, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-78dcf5dd84\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Feb 12 10:41:12.309: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1
+[It] should be able to convert a non homogeneous list of CRs [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:41:12.315: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Creating a v1 custom resource
+STEP: Create a v2 custom resource
+STEP: List CRs in v1
+STEP: List CRs in v2
+[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:41:14.033: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-webhook-9928" for this suite.
+[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_conversion_webhook.go:136
 
-• [SLOW TEST:7.536 seconds]
-[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  works for CRD preserving unknown fields at the schema root [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:7.568 seconds]
+[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should be able to convert a non homogeneous list of CRs [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance]","total":280,"completed":203,"skipped":3217,"failed":0}
-SSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance]","total":280,"completed":180,"skipped":3003,"failed":0}
+SSSSSSSSSS
 ------------------------------
-[sig-apps] Daemon set [Serial] 
-  should retry creating failed daemon pods [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] Garbage collector 
+  should delete RS created by deployment when not orphaning [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:23:11.994: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename daemonsets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-6474
+Feb 12 10:41:14.289: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-8913
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
-[It] should retry creating failed daemon pods [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a simple DaemonSet "daemon-set"
-STEP: Check that daemon pods launch on every node of the cluster.
-Aug 30 17:23:12.215: INFO: Number of nodes with available pods: 0
-Aug 30 17:23:12.215: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 17:23:13.227: INFO: Number of nodes with available pods: 0
-Aug 30 17:23:13.227: INFO: Node adoring-wozniak-54dcfd79fc-6rshr is running more than one daemon pod
-Aug 30 17:23:14.228: INFO: Number of nodes with available pods: 2
-Aug 30 17:23:14.228: INFO: Number of running nodes: 2, number of available pods: 2
-STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived.
-Aug 30 17:23:14.260: INFO: Number of nodes with available pods: 1
-Aug 30 17:23:14.260: INFO: Node adoring-wozniak-54dcfd79fc-948mf is running more than one daemon pod
-Aug 30 17:23:15.272: INFO: Number of nodes with available pods: 1
-Aug 30 17:23:15.272: INFO: Node adoring-wozniak-54dcfd79fc-948mf is running more than one daemon pod
-Aug 30 17:23:16.273: INFO: Number of nodes with available pods: 2
-Aug 30 17:23:16.273: INFO: Number of running nodes: 2, number of available pods: 2
-STEP: Wait for the failed daemon pod to be completely deleted.
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
-STEP: Deleting DaemonSet "daemon-set"
-STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6474, will wait for the garbage collector to delete the pods
-Aug 30 17:23:16.348: INFO: Deleting DaemonSet.extensions daemon-set took: 11.940516ms
-Aug 30 17:23:16.449: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.207982ms
-Aug 30 17:23:20.355: INFO: Number of nodes with available pods: 0
-Aug 30 17:23:20.355: INFO: Number of running nodes: 0, number of available pods: 0
-Aug 30 17:23:20.359: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-6474/daemonsets","resourceVersion":"24006"},"items":null}
-
-Aug 30 17:23:20.364: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-6474/pods","resourceVersion":"24006"},"items":null}
+[It] should delete RS created by deployment when not orphaning [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: create the deployment
+STEP: Wait for the Deployment to create new ReplicaSet
+STEP: delete the deployment
+STEP: wait for all rs to be garbage collected
+STEP: expected 0 pods, got 2 pods
+STEP: Gathering metrics
+W0212 10:41:15.340459      21 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Feb 12 10:41:15.340: INFO: For apiserver_request_total:
+For apiserver_request_latency_seconds:
+For apiserver_init_events_total:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
 
-[AfterEach] [sig-apps] Daemon set [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:23:20.379: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "daemonsets-6474" for this suite.
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:41:15.340: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-8913" for this suite.
+•{"msg":"PASSED [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance]","total":280,"completed":181,"skipped":3013,"failed":0}
 
-• [SLOW TEST:8.401 seconds]
-[sig-apps] Daemon set [Serial]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should retry creating failed daemon pods [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance]","total":280,"completed":204,"skipped":3234,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-storage] ConfigMap 
-  updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:23:20.399: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-3907
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name configmap-test-upd-585f7b15-60d2-44aa-a880-b52a7d890a5b
-STEP: Creating the pod
-STEP: Updating configmap configmap-test-upd-585f7b15-60d2-44aa-a880-b52a7d890a5b
-STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:23:24.740: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-3907" for this suite.
-•{"msg":"PASSED [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":205,"skipped":3257,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with secret pod [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Subpath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Projected downwardAPI 
+  should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:23:24.758: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename subpath
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-5886
+Feb 12 10:41:15.377: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-5633
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
-STEP: Setting up data
-[It] should support subpaths with secret pod [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod pod-subpath-test-secret-8h6x
-STEP: Creating a pod to test atomic-volume-subpath
-Aug 30 17:23:24.942: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-8h6x" in namespace "subpath-5886" to be "success or failure"
-Aug 30 17:23:24.946: INFO: Pod "pod-subpath-test-secret-8h6x": Phase="Pending", Reason="", readiness=false. Elapsed: 4.233639ms
-Aug 30 17:23:26.963: INFO: Pod "pod-subpath-test-secret-8h6x": Phase="Running", Reason="", readiness=true. Elapsed: 2.020940213s
-Aug 30 17:23:28.968: INFO: Pod "pod-subpath-test-secret-8h6x": Phase="Running", Reason="", readiness=true. Elapsed: 4.026443797s
-Aug 30 17:23:30.975: INFO: Pod "pod-subpath-test-secret-8h6x": Phase="Running", Reason="", readiness=true. Elapsed: 6.032983762s
-Aug 30 17:23:32.981: INFO: Pod "pod-subpath-test-secret-8h6x": Phase="Running", Reason="", readiness=true. Elapsed: 8.039441782s
-Aug 30 17:23:34.987: INFO: Pod "pod-subpath-test-secret-8h6x": Phase="Running", Reason="", readiness=true. Elapsed: 10.045363864s
-Aug 30 17:23:36.993: INFO: Pod "pod-subpath-test-secret-8h6x": Phase="Running", Reason="", readiness=true. Elapsed: 12.050679787s
-Aug 30 17:23:38.999: INFO: Pod "pod-subpath-test-secret-8h6x": Phase="Running", Reason="", readiness=true. Elapsed: 14.056718488s
-Aug 30 17:23:41.004: INFO: Pod "pod-subpath-test-secret-8h6x": Phase="Running", Reason="", readiness=true. Elapsed: 16.062280821s
-Aug 30 17:23:43.010: INFO: Pod "pod-subpath-test-secret-8h6x": Phase="Running", Reason="", readiness=true. Elapsed: 18.068534495s
-Aug 30 17:23:45.017: INFO: Pod "pod-subpath-test-secret-8h6x": Phase="Running", Reason="", readiness=true. Elapsed: 20.074740255s
-Aug 30 17:23:47.024: INFO: Pod "pod-subpath-test-secret-8h6x": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.081983529s
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should provide container's memory limit [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Feb 12 10:41:15.690: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7f298a3c-92a1-418a-8972-f4ba6a0d8236" in namespace "projected-5633" to be "success or failure"
+Feb 12 10:41:15.708: INFO: Pod "downwardapi-volume-7f298a3c-92a1-418a-8972-f4ba6a0d8236": Phase="Pending", Reason="", readiness=false. Elapsed: 17.903762ms
+Feb 12 10:41:17.715: INFO: Pod "downwardapi-volume-7f298a3c-92a1-418a-8972-f4ba6a0d8236": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024546147s
 STEP: Saw pod success
-Aug 30 17:23:47.024: INFO: Pod "pod-subpath-test-secret-8h6x" satisfied condition "success or failure"
-Aug 30 17:23:47.029: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-subpath-test-secret-8h6x container test-container-subpath-secret-8h6x: 
+Feb 12 10:41:17.715: INFO: Pod "downwardapi-volume-7f298a3c-92a1-418a-8972-f4ba6a0d8236" satisfied condition "success or failure"
+Feb 12 10:41:17.722: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-7f298a3c-92a1-418a-8972-f4ba6a0d8236 container client-container: 
 STEP: delete the pod
-Aug 30 17:23:47.106: INFO: Waiting for pod pod-subpath-test-secret-8h6x to disappear
-Aug 30 17:23:47.111: INFO: Pod pod-subpath-test-secret-8h6x no longer exists
-STEP: Deleting pod pod-subpath-test-secret-8h6x
-Aug 30 17:23:47.111: INFO: Deleting pod "pod-subpath-test-secret-8h6x" in namespace "subpath-5886"
-[AfterEach] [sig-storage] Subpath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:23:47.117: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "subpath-5886" for this suite.
-
-• [SLOW TEST:22.376 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
-  Atomic writer volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
-    should support subpaths with secret pod [LinuxOnly] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [LinuxOnly] [Conformance]","total":280,"completed":206,"skipped":3288,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should mutate custom resource with different stored version [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:23:47.135: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-4214
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Aug 30 17:23:47.641: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Aug 30 17:23:50.671: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should mutate custom resource with different stored version [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:23:50.677: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Registering the mutating webhook for custom resource e2e-test-webhook-7635-crds.webhook.example.com via the AdmissionRegistration API
-STEP: Creating a custom resource while v1 is storage version
-STEP: Patching Custom Resource Definition to set v2 as storage
-STEP: Patching the custom resource while v2 is storage version
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:23:52.216: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-4214" for this suite.
-STEP: Destroying namespace "webhook-4214-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
-
-• [SLOW TEST:5.169 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should mutate custom resource with different stored version [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance]","total":280,"completed":207,"skipped":3318,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-network] Services 
-  should be able to change the type from ClusterIP to ExternalName [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:23:52.313: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename services
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-4896
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should be able to change the type from ClusterIP to ExternalName [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a service clusterip-service with the type=ClusterIP in namespace services-4896
-STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service
-STEP: creating service externalsvc in namespace services-4896
-STEP: creating replication controller externalsvc in namespace services-4896
-I0830 17:23:52.575499      23 runners.go:189] Created replication controller with name: externalsvc, namespace: services-4896, replica count: 2
-I0830 17:23:55.625977      23 runners.go:189] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-STEP: changing the ClusterIP service to type=ExternalName
-Aug 30 17:23:55.657: INFO: Creating new exec pod
-Aug 30 17:23:57.676: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=services-4896 execpod4945t -- /bin/sh -x -c nslookup clusterip-service'
-Aug 30 17:23:58.216: INFO: stderr: "+ nslookup clusterip-service\n"
-Aug 30 17:23:58.216: INFO: stdout: "Server:\t\t10.240.16.10\nAddress:\t10.240.16.10#53\n\nclusterip-service.services-4896.svc.cluster.local\tcanonical name = externalsvc.services-4896.svc.cluster.local.\nName:\texternalsvc.services-4896.svc.cluster.local\nAddress: 10.240.27.141\n\n"
-STEP: deleting ReplicationController externalsvc in namespace services-4896, will wait for the garbage collector to delete the pods
-Aug 30 17:23:58.282: INFO: Deleting ReplicationController externalsvc took: 11.30533ms
-Aug 30 17:23:58.382: INFO: Terminating ReplicationController externalsvc pods took: 100.171299ms
-Aug 30 17:24:10.009: INFO: Cleaning up the ClusterIP to ExternalName test service
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:24:10.031: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-4896" for this suite.
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
-
-• [SLOW TEST:17.736 seconds]
-[sig-network] Services
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should be able to change the type from ClusterIP to ExternalName [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance]","total":280,"completed":208,"skipped":3386,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Feb 12 10:41:17.809: INFO: Waiting for pod downwardapi-volume-7f298a3c-92a1-418a-8972-f4ba6a0d8236 to disappear
+Feb 12 10:41:17.816: INFO: Pod downwardapi-volume-7f298a3c-92a1-418a-8972-f4ba6a0d8236 no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:41:17.816: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-5633" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance]","total":280,"completed":182,"skipped":3013,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-scheduling] NoExecuteTaintManager Single Pod [Serial] 
-  removing taint cancels eviction [Disruptive] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-scheduling] NoExecuteTaintManager Single Pod [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch 
+  watch on custom resource definition objects [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:24:10.049: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename taint-single-pod
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in taint-single-pod-9872
+Feb 12 10:41:17.840: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename crd-watch
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-watch-7370
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] NoExecuteTaintManager Single Pod [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/taints.go:163
-Aug 30 17:24:10.231: INFO: Waiting up to 1m0s for all nodes to be ready
-Aug 30 17:25:10.254: INFO: Waiting for terminating namespaces to be deleted...
-[It] removing taint cancels eviction [Disruptive] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:25:10.259: INFO: Starting informer...
-STEP: Starting pod...
-Aug 30 17:25:10.480: INFO: Pod is running on adoring-wozniak-54dcfd79fc-948mf. Tainting Node
-STEP: Trying to apply a taint on the Node
-STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute
-STEP: Waiting short time to make sure Pod is queued for deletion
-Aug 30 17:25:10.500: INFO: Pod wasn't evicted. Proceeding
-Aug 30 17:25:10.500: INFO: Removing taint from Node
-STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute
-STEP: Waiting some time to make sure that toleration time passed.
-Aug 30 17:26:25.526: INFO: Pod wasn't evicted. Test successful
-[AfterEach] [sig-scheduling] NoExecuteTaintManager Single Pod [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:26:25.527: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "taint-single-pod-9872" for this suite.
+[It] watch on custom resource definition objects [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:41:18.025: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Creating first CR 
+Feb 12 10:41:18.653: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2021-02-12T10:41:18Z generation:1 name:name1 resourceVersion:28002 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:1098c71f-2722-4b71-866a-bfb3edcee1a4] num:map[num1:9223372036854775807 num2:1000000]]}
+STEP: Creating second CR
+Feb 12 10:41:28.665: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2021-02-12T10:41:28Z generation:1 name:name2 resourceVersion:28089 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:6dbdf7e4-653d-420f-b5f1-840654a639de] num:map[num1:9223372036854775807 num2:1000000]]}
+STEP: Modifying first CR
+Feb 12 10:41:38.674: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2021-02-12T10:41:18Z generation:2 name:name1 resourceVersion:28135 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:1098c71f-2722-4b71-866a-bfb3edcee1a4] num:map[num1:9223372036854775807 num2:1000000]]}
+STEP: Modifying second CR
+Feb 12 10:41:48.684: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2021-02-12T10:41:28Z generation:2 name:name2 resourceVersion:28179 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:6dbdf7e4-653d-420f-b5f1-840654a639de] num:map[num1:9223372036854775807 num2:1000000]]}
+STEP: Deleting first CR
+Feb 12 10:41:58.700: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2021-02-12T10:41:18Z generation:2 name:name1 resourceVersion:28225 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:1098c71f-2722-4b71-866a-bfb3edcee1a4] num:map[num1:9223372036854775807 num2:1000000]]}
+STEP: Deleting second CR
+Feb 12 10:42:08.717: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2021-02-12T10:41:28Z generation:2 name:name2 resourceVersion:28269 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:6dbdf7e4-653d-420f-b5f1-840654a639de] num:map[num1:9223372036854775807 num2:1000000]]}
+[AfterEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:42:19.236: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-watch-7370" for this suite.
 
-• [SLOW TEST:135.494 seconds]
-[sig-scheduling] NoExecuteTaintManager Single Pod [Serial]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
-  removing taint cancels eviction [Disruptive] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:61.415 seconds]
+[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  CustomResourceDefinition Watch
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_watch.go:41
+    watch on custom resource definition objects [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-scheduling] NoExecuteTaintManager Single Pod [Serial] removing taint cancels eviction [Disruptive] [Conformance]","total":280,"completed":209,"skipped":3414,"failed":0}
-SSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance]","total":280,"completed":183,"skipped":3065,"failed":0}
+SS
 ------------------------------
 [sig-api-machinery] Garbage collector 
-  should delete pods created by rc when not orphaning [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:26:25.545: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:42:19.256: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-4368
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-8370
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should delete pods created by rc when not orphaning [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the rc
-STEP: delete the rc
-STEP: wait for all pods to be garbage collected
+[It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: create the rc1
+STEP: create the rc2
+STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well
+STEP: delete the rc simpletest-rc-to-be-deleted
+STEP: wait for the rc to be deleted
 STEP: Gathering metrics
-W0830 17:26:35.748159      23 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-Aug 30 17:26:35.748: INFO: For apiserver_request_total:
+W0212 10:42:29.537350      21 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Feb 12 10:42:29.537: INFO: For apiserver_request_total:
 For apiserver_request_latency_seconds:
 For apiserver_init_events_total:
 For garbage_collector_attempt_to_delete_queue_latency:
@@ -8944,3376 +8506,3884 @@ For errors_total:
 For evicted_pods_total:
 
 [AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:26:35.748: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-4368" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:42:29.537: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-8370" for this suite.
 
-• [SLOW TEST:10.218 seconds]
+• [SLOW TEST:10.299 seconds]
 [sig-api-machinery] Garbage collector
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should delete pods created by rc when not orphaning [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]","total":280,"completed":184,"skipped":3067,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  pod should support shared volumes between containers [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:42:29.556: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-6680
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] pod should support shared volumes between containers [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating Pod
+STEP: Waiting for the pod running
+STEP: Geting the pod
+STEP: Reading file content from the nginx-container
+Feb 12 10:42:31.783: INFO: ExecWithOptions {Command:[/bin/sh -c cat /usr/share/volumeshare/shareddata.txt] Namespace:emptydir-6680 PodName:pod-sharedvolume-0e04b713-ee89-42d1-9e07-40a75666c5c2 ContainerName:busybox-main-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:42:31.783: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:42:32.307: INFO: Exec stderr: ""
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:42:32.307: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-6680" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance]","total":280,"completed":185,"skipped":3130,"failed":0}
+S
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance]","total":280,"completed":210,"skipped":3430,"failed":0}
-SSSSSSSSSSSSSSSSSSSSS
+[sig-node] Downward API 
+  should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-node] Downward API
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:42:32.325: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7798
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward api env vars
+Feb 12 10:42:32.498: INFO: Waiting up to 5m0s for pod "downward-api-96032159-ff81-47db-b0db-9ac0439dec77" in namespace "downward-api-7798" to be "success or failure"
+Feb 12 10:42:32.505: INFO: Pod "downward-api-96032159-ff81-47db-b0db-9ac0439dec77": Phase="Pending", Reason="", readiness=false. Elapsed: 6.764699ms
+Feb 12 10:42:34.512: INFO: Pod "downward-api-96032159-ff81-47db-b0db-9ac0439dec77": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013426675s
+Feb 12 10:42:36.519: INFO: Pod "downward-api-96032159-ff81-47db-b0db-9ac0439dec77": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020502448s
+STEP: Saw pod success
+Feb 12 10:42:36.519: INFO: Pod "downward-api-96032159-ff81-47db-b0db-9ac0439dec77" satisfied condition "success or failure"
+Feb 12 10:42:36.527: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downward-api-96032159-ff81-47db-b0db-9ac0439dec77 container dapi-container: 
+STEP: delete the pod
+Feb 12 10:42:36.610: INFO: Waiting for pod downward-api-96032159-ff81-47db-b0db-9ac0439dec77 to disappear
+Feb 12 10:42:36.615: INFO: Pod downward-api-96032159-ff81-47db-b0db-9ac0439dec77 no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:42:36.616: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-7798" for this suite.
+•{"msg":"PASSED [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]","total":280,"completed":186,"skipped":3131,"failed":0}
+SSSSSSSSSS
 ------------------------------
-[k8s.io] [sig-node] Events 
-  should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] [sig-node] Events
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] DNS 
+  should provide DNS for the cluster  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:26:35.763: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename events
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in events-5257
+Feb 12 10:42:36.635: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename dns
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-6183
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
+[It] should provide DNS for the cluster  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6183.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;sleep 1; done
+
+STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-6183.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;sleep 1; done
+
+STEP: creating a pod to probe DNS
 STEP: submitting the pod to kubernetes
-STEP: verifying the pod is in kubernetes
 STEP: retrieving the pod
-Aug 30 17:26:37.960: INFO: &Pod{ObjectMeta:{send-events-227d055c-ea1a-41eb-8cd8-39d9673a5cb5  events-5257 /api/v1/namespaces/events-5257/pods/send-events-227d055c-ea1a-41eb-8cd8-39d9673a5cb5 52e73ac7-b400-42e4-adb3-e5cbfc62ef96 25119 0 2020-08-30 17:26:35 +0000 UTC   map[name:foo time:929334961] map[cni.projectcalico.org/podIP:172.25.0.221/32] [] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-mpw6z,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-mpw6z,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:p,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[serve-hostname],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:,HostPort:0,ContainerPort:80,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-mpw6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*30,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:adoring-wozniak-54dcfd79fc-948mf,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:26:35 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:26:37 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:26:37 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2020-08-30 17:26:35 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:46.101.135.210,PodIP:172.25.0.221,StartTime:2020-08-30 17:26:35 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:p,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2020-08-30 17:26:37 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/agnhost@sha256:daf5332100521b1256d0e3c56d697a238eaec3af48897ed9167cbadd426773b5,ContainerID:docker://f25ee0d7d6cd223046af7f19dc786125de6785681b0597143abcb7f5b439ad4c,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.0.221,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+STEP: looking for the results for each expected name from probers
+Feb 12 10:42:41.572: INFO: DNS probes using dns-6183/dns-test-022ea288-398b-4033-b940-afd90c63630b succeeded
 
-STEP: checking for scheduler event about the pod
-Aug 30 17:26:39.969: INFO: Saw scheduler event for our pod.
-STEP: checking for kubelet event about the pod
-Aug 30 17:26:41.976: INFO: Saw kubelet event for our pod.
 STEP: deleting the pod
-[AfterEach] [k8s.io] [sig-node] Events
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:26:41.984: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "events-5257" for this suite.
+[AfterEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:42:41.592: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-6183" for this suite.
+•{"msg":"PASSED [sig-network] DNS should provide DNS for the cluster  [Conformance]","total":280,"completed":187,"skipped":3141,"failed":0}
 
-• [SLOW TEST:6.236 seconds]
-[k8s.io] [sig-node] Events
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] [sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running  [Conformance]","total":280,"completed":211,"skipped":3451,"failed":0}
-[sig-cli] Kubectl client Kubectl patch 
-  should add annotations for pods in rc  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[sig-storage] EmptyDir wrapper volumes 
+  should not conflict [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir wrapper volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:42:41.614: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir-wrapper
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-wrapper-7852
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should not conflict [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Cleaning up the secret
+STEP: Cleaning up the configmap
+STEP: Cleaning up the pod
+[AfterEach] [sig-storage] EmptyDir wrapper volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:42:43.881: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-wrapper-7852" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance]","total":280,"completed":188,"skipped":3141,"failed":0}
+SS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:42:43.899: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-6218
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir 0644 on tmpfs
+Feb 12 10:42:44.076: INFO: Waiting up to 5m0s for pod "pod-f4fa50c5-2212-413f-8b09-8c211c14d955" in namespace "emptydir-6218" to be "success or failure"
+Feb 12 10:42:44.083: INFO: Pod "pod-f4fa50c5-2212-413f-8b09-8c211c14d955": Phase="Pending", Reason="", readiness=false. Elapsed: 6.3261ms
+Feb 12 10:42:46.089: INFO: Pod "pod-f4fa50c5-2212-413f-8b09-8c211c14d955": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012746163s
+Feb 12 10:42:48.095: INFO: Pod "pod-f4fa50c5-2212-413f-8b09-8c211c14d955": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018785206s
+STEP: Saw pod success
+Feb 12 10:42:48.095: INFO: Pod "pod-f4fa50c5-2212-413f-8b09-8c211c14d955" satisfied condition "success or failure"
+Feb 12 10:42:48.100: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-f4fa50c5-2212-413f-8b09-8c211c14d955 container test-container: 
+STEP: delete the pod
+Feb 12 10:42:48.174: INFO: Waiting for pod pod-f4fa50c5-2212-413f-8b09-8c211c14d955 to disappear
+Feb 12 10:42:48.178: INFO: Pod pod-f4fa50c5-2212-413f-8b09-8c211c14d955 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:42:48.179: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-6218" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":189,"skipped":3143,"failed":0}
+SSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:42:48.199: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-2425
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir 0666 on node default medium
+Feb 12 10:42:48.366: INFO: Waiting up to 5m0s for pod "pod-9ee9b207-b1bc-4cf6-83f4-246b0fce9901" in namespace "emptydir-2425" to be "success or failure"
+Feb 12 10:42:48.372: INFO: Pod "pod-9ee9b207-b1bc-4cf6-83f4-246b0fce9901": Phase="Pending", Reason="", readiness=false. Elapsed: 6.286901ms
+Feb 12 10:42:50.377: INFO: Pod "pod-9ee9b207-b1bc-4cf6-83f4-246b0fce9901": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011083524s
+STEP: Saw pod success
+Feb 12 10:42:50.377: INFO: Pod "pod-9ee9b207-b1bc-4cf6-83f4-246b0fce9901" satisfied condition "success or failure"
+Feb 12 10:42:50.384: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-9ee9b207-b1bc-4cf6-83f4-246b0fce9901 container test-container: 
+STEP: delete the pod
+Feb 12 10:42:50.428: INFO: Waiting for pod pod-9ee9b207-b1bc-4cf6-83f4-246b0fce9901 to disappear
+Feb 12 10:42:50.432: INFO: Pod pod-9ee9b207-b1bc-4cf6-83f4-246b0fce9901 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:42:50.432: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-2425" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":190,"skipped":3149,"failed":0}
+SSSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client Kubectl describe 
+  should check if kubectl describe prints relevant information for rc and pods  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:26:41.999: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:42:50.454: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-5183
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-6744
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[It] should add annotations for pods in rc  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating Agnhost RC
-Aug 30 17:26:42.157: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-5183'
-Aug 30 17:26:42.396: INFO: stderr: ""
-Aug 30 17:26:42.396: INFO: stdout: "replicationcontroller/agnhost-master created\n"
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[It] should check if kubectl describe prints relevant information for rc and pods  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:42:50.632: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-6744'
+Feb 12 10:42:50.940: INFO: stderr: ""
+Feb 12 10:42:50.940: INFO: stdout: "replicationcontroller/agnhost-master created\n"
+Feb 12 10:42:50.940: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-6744'
+Feb 12 10:42:51.121: INFO: stderr: ""
+Feb 12 10:42:51.121: INFO: stdout: "service/agnhost-master created\n"
 STEP: Waiting for Agnhost master to start.
-Aug 30 17:26:43.402: INFO: Selector matched 1 pods for map[app:agnhost]
-Aug 30 17:26:43.402: INFO: Found 0 / 1
-Aug 30 17:26:44.403: INFO: Selector matched 1 pods for map[app:agnhost]
-Aug 30 17:26:44.403: INFO: Found 1 / 1
-Aug 30 17:26:44.403: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
-STEP: patching all pods
-Aug 30 17:26:44.408: INFO: Selector matched 1 pods for map[app:agnhost]
-Aug 30 17:26:44.408: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
-Aug 30 17:26:44.408: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 patch pod agnhost-master-lm769 --namespace=kubectl-5183 -p {"metadata":{"annotations":{"x":"y"}}}'
-Aug 30 17:26:44.485: INFO: stderr: ""
-Aug 30 17:26:44.485: INFO: stdout: "pod/agnhost-master-lm769 patched\n"
-STEP: checking annotations
-Aug 30 17:26:44.490: INFO: Selector matched 1 pods for map[app:agnhost]
-Aug 30 17:26:44.490: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Feb 12 10:42:52.128: INFO: Selector matched 1 pods for map[app:agnhost]
+Feb 12 10:42:52.128: INFO: Found 0 / 1
+Feb 12 10:42:53.130: INFO: Selector matched 1 pods for map[app:agnhost]
+Feb 12 10:42:53.130: INFO: Found 1 / 1
+Feb 12 10:42:53.130: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+Feb 12 10:42:53.140: INFO: Selector matched 1 pods for map[app:agnhost]
+Feb 12 10:42:53.140: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Feb 12 10:42:53.140: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 describe pod agnhost-master-xnvns --namespace=kubectl-6744'
+Feb 12 10:42:53.260: INFO: stderr: ""
+Feb 12 10:42:53.260: INFO: stdout: "Name:         agnhost-master-xnvns\nNamespace:    kubectl-6744\nPriority:     0\nNode:         nostalgic-fermat-cdc5d8777-xg74t/138.68.97.197\nStart Time:   Fri, 12 Feb 2021 10:42:50 +0000\nLabels:       app=agnhost\n              role=master\nAnnotations:  cni.projectcalico.org/podIP: 172.25.1.173/32\nStatus:       Running\nIP:           172.25.1.173\nIPs:\n  IP:           172.25.1.173\nControlled By:  ReplicationController/agnhost-master\nContainers:\n  agnhost-master:\n    Container ID:   docker://8f62f05d94d8780f8d1a683d7cd9dffe8a7d74bb4d95dc8da1a417020b6b3b2c\n    Image:          gcr.io/kubernetes-e2e-test-images/agnhost:2.8\n    Image ID:       docker-pullable://gcr.io/kubernetes-e2e-test-images/agnhost@sha256:daf5332100521b1256d0e3c56d697a238eaec3af48897ed9167cbadd426773b5\n    Port:           6379/TCP\n    Host Port:      0/TCP\n    State:          Running\n      Started:      Fri, 12 Feb 2021 10:42:51 +0000\n    Ready:          True\n    Restart Count:  0\n    Environment:    \n    Mounts:\n      /var/run/secrets/kubernetes.io/serviceaccount from default-token-ltqgv (ro)\nConditions:\n  Type              Status\n  Initialized       True \n  Ready             True \n  ContainersReady   True \n  PodScheduled      True \nVolumes:\n  default-token-ltqgv:\n    Type:        Secret (a volume populated by a Secret)\n    SecretName:  default-token-ltqgv\n    Optional:    false\nQoS Class:       BestEffort\nNode-Selectors:  \nTolerations:     node.kubernetes.io/not-ready:NoExecute for 300s\n                 node.kubernetes.io/unreachable:NoExecute for 300s\nEvents:\n  Type    Reason     Age   From                                       Message\n  ----    ------     ----  ----                                       -------\n  Normal  Scheduled  2s    default-scheduler                          Successfully assigned kubectl-6744/agnhost-master-xnvns to nostalgic-fermat-cdc5d8777-xg74t\n  Normal  Pulled     2s    kubelet, nostalgic-fermat-cdc5d8777-xg74t  Container image \"gcr.io/kubernetes-e2e-test-images/agnhost:2.8\" already present on machine\n  Normal  Created    2s    kubelet, nostalgic-fermat-cdc5d8777-xg74t  Created container agnhost-master\n  Normal  Started    2s    kubelet, nostalgic-fermat-cdc5d8777-xg74t  Started container agnhost-master\n"
+Feb 12 10:42:53.260: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 describe rc agnhost-master --namespace=kubectl-6744'
+Feb 12 10:42:53.394: INFO: stderr: ""
+Feb 12 10:42:53.394: INFO: stdout: "Name:         agnhost-master\nNamespace:    kubectl-6744\nSelector:     app=agnhost,role=master\nLabels:       app=agnhost\n              role=master\nAnnotations:  \nReplicas:     1 current / 1 desired\nPods Status:  1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n  Labels:  app=agnhost\n           role=master\n  Containers:\n   agnhost-master:\n    Image:        gcr.io/kubernetes-e2e-test-images/agnhost:2.8\n    Port:         6379/TCP\n    Host Port:    0/TCP\n    Environment:  \n    Mounts:       \n  Volumes:        \nEvents:\n  Type    Reason            Age   From                    Message\n  ----    ------            ----  ----                    -------\n  Normal  SuccessfulCreate  3s    replication-controller  Created pod: agnhost-master-xnvns\n"
+Feb 12 10:42:53.394: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 describe service agnhost-master --namespace=kubectl-6744'
+Feb 12 10:42:53.494: INFO: stderr: ""
+Feb 12 10:42:53.494: INFO: stdout: "Name:              agnhost-master\nNamespace:         kubectl-6744\nLabels:            app=agnhost\n                   role=master\nAnnotations:       \nSelector:          app=agnhost,role=master\nType:              ClusterIP\nIP:                10.240.27.91\nPort:                6379/TCP\nTargetPort:        agnhost-server/TCP\nEndpoints:         172.25.1.173:6379\nSession Affinity:  None\nEvents:            \n"
+Feb 12 10:42:53.511: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 describe node nostalgic-fermat-cdc5d8777-dbhjj'
+Feb 12 10:42:53.937: INFO: stderr: ""
+Feb 12 10:42:53.937: INFO: stdout: "Name:               nostalgic-fermat-cdc5d8777-dbhjj\nRoles:              \nLabels:             beta.kubernetes.io/arch=amd64\n                    beta.kubernetes.io/os=linux\n                    kubernetes.io/arch=amd64\n                    kubernetes.io/hostname=nostalgic-fermat-cdc5d8777-dbhjj\n                    kubernetes.io/os=linux\n                    machine-controller/owned-by=0a279a5d-be94-4b18-90a7-760afd893325\n                    system/cluster=xbfxh5n7q7\n                    system/project=kr6n49b5rm\n                    x-kubernetes.io/distribution=ubuntu\nAnnotations:        cluster.k8s.io/machine: kube-system/nostalgic-fermat-cdc5d8777-dbhjj\n                    flannel.alpha.coreos.com/backend-data: {\"VtepMAC\":\"56:cb:a3:ab:ee:ba\"}\n                    flannel.alpha.coreos.com/backend-type: vxlan\n                    flannel.alpha.coreos.com/kube-subnet-manager: true\n                    flannel.alpha.coreos.com/public-ip: 104.248.134.247\n                    node.alpha.kubernetes.io/ttl: 0\n                    projectcalico.org/IPv4IPIPTunnelAddr: 172.25.2.1\n                    volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp:  Fri, 12 Feb 2021 09:48:16 +0000\nTaints:             \nUnschedulable:      false\nLease:\n  HolderIdentity:  nostalgic-fermat-cdc5d8777-dbhjj\n  AcquireTime:     \n  RenewTime:       Fri, 12 Feb 2021 10:42:45 +0000\nConditions:\n  Type             Status  LastHeartbeatTime                 LastTransitionTime                Reason                       Message\n  ----             ------  -----------------                 ------------------                ------                       -------\n  MemoryPressure   False   Fri, 12 Feb 2021 10:39:09 +0000   Fri, 12 Feb 2021 09:48:16 +0000   KubeletHasSufficientMemory   kubelet has sufficient memory available\n  DiskPressure     False   Fri, 12 Feb 2021 10:39:09 +0000   Fri, 12 Feb 2021 09:48:16 +0000   KubeletHasNoDiskPressure     kubelet has no disk pressure\n  PIDPressure      False   Fri, 12 Feb 2021 10:39:09 +0000   Fri, 12 Feb 2021 09:48:16 +0000   KubeletHasSufficientPID      kubelet has sufficient PID available\n  Ready            True    Fri, 12 Feb 2021 10:39:09 +0000   Fri, 12 Feb 2021 09:48:36 +0000   KubeletReady                 kubelet is posting ready status. AppArmor enabled\nAddresses:\n  InternalIP:  104.248.134.247\n  Hostname:    nostalgic-fermat-cdc5d8777-dbhjj\nCapacity:\n  cpu:                2\n  ephemeral-storage:  25226960Ki\n  hugepages-1Gi:      0\n  hugepages-2Mi:      0\n  memory:             4039272Ki\n  pods:               110\nAllocatable:\n  cpu:                1600m\n  ephemeral-storage:  21101682650\n  hugepages-1Gi:      0\n  hugepages-2Mi:      0\n  memory:             3527272Ki\n  pods:               110\nSystem Info:\n  Machine ID:                 c41e459fe67c4295a17b346174dfcaa4\n  System UUID:                C41E459F-E67C-4295-A17B-346174DFCAA4\n  Boot ID:                    1c695bc6-b981-4900-825d-bd38a3972a80\n  Kernel Version:             4.15.0-121-generic\n  OS Image:                   Ubuntu 18.04.5 LTS\n  Operating System:           linux\n  Architecture:               amd64\n  Container Runtime Version:  docker://19.3.12\n  Kubelet Version:            v1.17.16\n  Kube-Proxy Version:         v1.17.16\nPodCIDR:                      172.25.2.0/24\nPodCIDRs:                     172.25.2.0/24\nNon-terminated Pods:          (9 in total)\n  Namespace                   Name                                                       CPU Requests  CPU Limits  Memory Requests  Memory Limits  AGE\n  ---------                   ----                                                       ------------  ----------  ---------------  -------------  ---\n  kube-system                 canal-k48gh                                                250m (15%)    0 (0%)      0 (0%)           0 (0%)         54m\n  kube-system                 coredns-6449d494cc-vrdxh                                   50m (3%)      100m (6%)   32Mi (0%)        64Mi (1%)      4m4s\n  kube-system                 kube-proxy-c88c9                                           75m (4%)      250m (15%)  50Mi (1%)        250Mi (7%)     54m\n  kube-system                 logrotate-4f9dq                                            75m (4%)      250m (15%)  50Mi (1%)        250Mi (7%)     54m\n  kube-system                 node-local-dns-k2khb                                       0 (0%)        0 (0%)      0 (0%)           0 (0%)         54m\n  kube-system                 openvpn-client-78d595f58b-2pzkd                            30m (1%)      200m (12%)  30Mi (0%)        82Mi (2%)      4m4s\n  kube-system                 user-ssh-keys-agent-8lrpd                                  0 (0%)        0 (0%)      0 (0%)           0 (0%)         54m\n  sonobuoy                    sonobuoy-e2e-job-690c418daa0f4403                          0 (0%)        0 (0%)      0 (0%)           0 (0%)         54m\n  sonobuoy                    sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-vkql5    0 (0%)        0 (0%)      0 (0%)           0 (0%)         54m\nAllocated resources:\n  (Total limits may be over 100 percent, i.e., overcommitted.)\n  Resource           Requests    Limits\n  --------           --------    ------\n  cpu                480m (30%)  800m (50%)\n  memory             162Mi (4%)  646Mi (18%)\n  ephemeral-storage  0 (0%)      0 (0%)\nEvents:\n  Type    Reason                   Age                From                                          Message\n  ----    ------                   ----               ----                                          -------\n  Normal  NodeHasSufficientMemory  54m (x8 over 54m)  kubelet, nostalgic-fermat-cdc5d8777-dbhjj     Node nostalgic-fermat-cdc5d8777-dbhjj status is now: NodeHasSufficientMemory\n  Normal  NodeHasSufficientPID     54m (x8 over 54m)  kubelet, nostalgic-fermat-cdc5d8777-dbhjj     Node nostalgic-fermat-cdc5d8777-dbhjj status is now: NodeHasSufficientPID\n  Normal  Starting                 54m                kube-proxy, nostalgic-fermat-cdc5d8777-dbhjj  Starting kube-proxy.\n"
+Feb 12 10:42:53.937: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 describe namespace kubectl-6744'
+Feb 12 10:42:54.073: INFO: stderr: ""
+Feb 12 10:42:54.073: INFO: stdout: "Name:         kubectl-6744\nLabels:       e2e-framework=kubectl\n              e2e-run=290da14e-be04-4851-88fa-116a06471991\nAnnotations:  \nStatus:       Active\n\nNo resource quota.\n\nNo LimitRange resource.\n"
 [AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:26:44.490: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-5183" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc  [Conformance]","total":280,"completed":212,"skipped":3451,"failed":0}
-SSSSSS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:42:54.073: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-6744" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods  [Conformance]","total":280,"completed":191,"skipped":3163,"failed":0}
+SSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-scheduling] SchedulerPredicates [Serial] 
-  validates resource limits of pods that are allowed to run  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with downward pod [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Subpath
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:26:44.504: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename sched-pred
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-4333
+Feb 12 10:42:54.097: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-9200
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
-Aug 30 17:26:44.666: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
-Aug 30 17:26:44.683: INFO: Waiting for terminating namespaces to be deleted...
-Aug 30 17:26:44.687: INFO: 
-Logging pods the kubelet thinks is on node adoring-wozniak-54dcfd79fc-6rshr before test
-Aug 30 17:26:44.753: INFO: kube-proxy-wxdxv from kube-system started at 2020-08-30 16:28:14 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.753: INFO: 	Container kube-proxy ready: true, restart count 0
-Aug 30 17:26:44.753: INFO: node-local-dns-cvxnf from kube-system started at 2020-08-30 16:28:34 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.753: INFO: 	Container node-cache ready: true, restart count 0
-Aug 30 17:26:44.753: INFO: logrotate-57bmz from kube-system started at 2020-08-30 16:28:34 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.753: INFO: 	Container logrotate ready: true, restart count 0
-Aug 30 17:26:44.753: INFO: sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-nd5nc from sonobuoy started at 2020-08-30 16:29:24 +0000 UTC (2 container statuses recorded)
-Aug 30 17:26:44.753: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 17:26:44.753: INFO: 	Container systemd-logs ready: true, restart count 0
-Aug 30 17:26:44.753: INFO: coredns-54457d966b-fbnz4 from kube-system started at 2020-08-30 16:28:39 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.753: INFO: 	Container coredns ready: true, restart count 0
-Aug 30 17:26:44.753: INFO: sonobuoy-e2e-job-cf49606f646f4c8a from sonobuoy started at 2020-08-30 16:29:23 +0000 UTC (2 container statuses recorded)
-Aug 30 17:26:44.753: INFO: 	Container e2e ready: true, restart count 0
-Aug 30 17:26:44.753: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 17:26:44.753: INFO: dashboard-metrics-scraper-59bfc65dc9-252k4 from kubernetes-dashboard started at 2020-08-30 17:18:06 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.754: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
-Aug 30 17:26:44.754: INFO: canal-54glj from kube-system started at 2020-08-30 16:28:14 +0000 UTC (2 container statuses recorded)
-Aug 30 17:26:44.754: INFO: 	Container calico-node ready: true, restart count 0
-Aug 30 17:26:44.754: INFO: 	Container kube-flannel ready: true, restart count 0
-Aug 30 17:26:44.754: INFO: user-ssh-keys-agent-xwrzj from kube-system started at 2020-08-30 16:28:14 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.754: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
-Aug 30 17:26:44.754: INFO: coredns-54457d966b-6cw7g from kube-system started at 2020-08-30 17:18:06 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.754: INFO: 	Container coredns ready: true, restart count 0
-Aug 30 17:26:44.754: INFO: sonobuoy from sonobuoy started at 2020-08-30 16:29:18 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.754: INFO: 	Container kube-sonobuoy ready: true, restart count 0
-Aug 30 17:26:44.754: INFO: openvpn-client-78d595f58b-vpgdr from kube-system started at 2020-08-30 17:18:06 +0000 UTC (2 container statuses recorded)
-Aug 30 17:26:44.754: INFO: 	Container dnat-controller ready: true, restart count 0
-Aug 30 17:26:44.754: INFO: 	Container openvpn-client ready: true, restart count 0
-Aug 30 17:26:44.754: INFO: dashboard-metrics-scraper-59bfc65dc9-t82kj from kubernetes-dashboard started at 2020-08-30 17:18:06 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.754: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
-Aug 30 17:26:44.754: INFO: 
-Logging pods the kubelet thinks is on node adoring-wozniak-54dcfd79fc-948mf before test
-Aug 30 17:26:44.837: INFO: canal-lg5hn from kube-system started at 2020-08-30 16:28:09 +0000 UTC (2 container statuses recorded)
-Aug 30 17:26:44.837: INFO: 	Container calico-node ready: true, restart count 0
-Aug 30 17:26:44.837: INFO: 	Container kube-flannel ready: true, restart count 0
-Aug 30 17:26:44.837: INFO: agnhost-master-lm769 from kubectl-5183 started at 2020-08-30 17:26:42 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.837: INFO: 	Container agnhost-master ready: true, restart count 0
-Aug 30 17:26:44.837: INFO: sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-cdm8v from sonobuoy started at 2020-08-30 16:29:23 +0000 UTC (2 container statuses recorded)
-Aug 30 17:26:44.837: INFO: 	Container sonobuoy-worker ready: true, restart count 0
-Aug 30 17:26:44.837: INFO: 	Container systemd-logs ready: true, restart count 0
-Aug 30 17:26:44.837: INFO: node-local-dns-s2mgx from kube-system started at 2020-08-30 17:25:10 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.837: INFO: 	Container node-cache ready: true, restart count 0
-Aug 30 17:26:44.837: INFO: logrotate-bnfcr from kube-system started at 2020-08-30 17:25:24 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.837: INFO: 	Container logrotate ready: true, restart count 0
-Aug 30 17:26:44.837: INFO: kube-proxy-v88gx from kube-system started at 2020-08-30 16:28:09 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.837: INFO: 	Container kube-proxy ready: true, restart count 0
-Aug 30 17:26:44.837: INFO: user-ssh-keys-agent-vkbs9 from kube-system started at 2020-08-30 16:28:09 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.837: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
-Aug 30 17:26:44.837: INFO: send-events-227d055c-ea1a-41eb-8cd8-39d9673a5cb5 from events-5257 started at 2020-08-30 17:26:35 +0000 UTC (1 container statuses recorded)
-Aug 30 17:26:44.837: INFO: 	Container p ready: true, restart count 0
-[It] validates resource limits of pods that are allowed to run  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: verifying the node has the label node adoring-wozniak-54dcfd79fc-6rshr
-STEP: verifying the node has the label node adoring-wozniak-54dcfd79fc-948mf
-Aug 30 17:26:44.885: INFO: Pod send-events-227d055c-ea1a-41eb-8cd8-39d9673a5cb5 requesting resource cpu=0m on Node adoring-wozniak-54dcfd79fc-948mf
-Aug 30 17:26:44.885: INFO: Pod canal-54glj requesting resource cpu=250m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.885: INFO: Pod canal-lg5hn requesting resource cpu=250m on Node adoring-wozniak-54dcfd79fc-948mf
-Aug 30 17:26:44.885: INFO: Pod coredns-54457d966b-6cw7g requesting resource cpu=50m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.885: INFO: Pod coredns-54457d966b-fbnz4 requesting resource cpu=50m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.885: INFO: Pod kube-proxy-v88gx requesting resource cpu=75m on Node adoring-wozniak-54dcfd79fc-948mf
-Aug 30 17:26:44.885: INFO: Pod kube-proxy-wxdxv requesting resource cpu=75m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.885: INFO: Pod logrotate-57bmz requesting resource cpu=75m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.885: INFO: Pod logrotate-bnfcr requesting resource cpu=75m on Node adoring-wozniak-54dcfd79fc-948mf
-Aug 30 17:26:44.885: INFO: Pod node-local-dns-cvxnf requesting resource cpu=0m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.885: INFO: Pod node-local-dns-s2mgx requesting resource cpu=0m on Node adoring-wozniak-54dcfd79fc-948mf
-Aug 30 17:26:44.885: INFO: Pod openvpn-client-78d595f58b-vpgdr requesting resource cpu=30m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.885: INFO: Pod user-ssh-keys-agent-vkbs9 requesting resource cpu=0m on Node adoring-wozniak-54dcfd79fc-948mf
-Aug 30 17:26:44.885: INFO: Pod user-ssh-keys-agent-xwrzj requesting resource cpu=0m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.885: INFO: Pod agnhost-master-lm769 requesting resource cpu=0m on Node adoring-wozniak-54dcfd79fc-948mf
-Aug 30 17:26:44.885: INFO: Pod dashboard-metrics-scraper-59bfc65dc9-252k4 requesting resource cpu=50m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.885: INFO: Pod dashboard-metrics-scraper-59bfc65dc9-t82kj requesting resource cpu=50m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.885: INFO: Pod sonobuoy requesting resource cpu=0m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.885: INFO: Pod sonobuoy-e2e-job-cf49606f646f4c8a requesting resource cpu=0m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.885: INFO: Pod sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-cdm8v requesting resource cpu=0m on Node adoring-wozniak-54dcfd79fc-948mf
-Aug 30 17:26:44.885: INFO: Pod sonobuoy-systemd-logs-daemon-set-fc27dc07d16945d5-nd5nc requesting resource cpu=0m on Node adoring-wozniak-54dcfd79fc-6rshr
-STEP: Starting Pods to consume most of the cluster CPU.
-Aug 30 17:26:44.885: INFO: Creating a pod which consumes cpu=2219m on Node adoring-wozniak-54dcfd79fc-6rshr
-Aug 30 17:26:44.895: INFO: Creating a pod which consumes cpu=2380m on Node adoring-wozniak-54dcfd79fc-948mf
-STEP: Creating another pod that requires unavailable amount of CPU.
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-05156bb3-b84b-4a11-a3c8-f689a3a69a08.16301bc5fd892c6c], Reason = [Scheduled], Message = [Successfully assigned sched-pred-4333/filler-pod-05156bb3-b84b-4a11-a3c8-f689a3a69a08 to adoring-wozniak-54dcfd79fc-6rshr]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-05156bb3-b84b-4a11-a3c8-f689a3a69a08.16301bc63b56ba57], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-05156bb3-b84b-4a11-a3c8-f689a3a69a08.16301bc63faad98d], Reason = [Created], Message = [Created container filler-pod-05156bb3-b84b-4a11-a3c8-f689a3a69a08]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-05156bb3-b84b-4a11-a3c8-f689a3a69a08.16301bc64a2e9fc1], Reason = [Started], Message = [Started container filler-pod-05156bb3-b84b-4a11-a3c8-f689a3a69a08]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-0d7f8a65-b4c2-4299-ad5b-a838b5c282fa.16301bc5fdf4940d], Reason = [Scheduled], Message = [Successfully assigned sched-pred-4333/filler-pod-0d7f8a65-b4c2-4299-ad5b-a838b5c282fa to adoring-wozniak-54dcfd79fc-948mf]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-0d7f8a65-b4c2-4299-ad5b-a838b5c282fa.16301bc636154c81], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-0d7f8a65-b4c2-4299-ad5b-a838b5c282fa.16301bc639e04a7f], Reason = [Created], Message = [Created container filler-pod-0d7f8a65-b4c2-4299-ad5b-a838b5c282fa]
-STEP: Considering event: 
-Type = [Normal], Name = [filler-pod-0d7f8a65-b4c2-4299-ad5b-a838b5c282fa.16301bc6434fa5bb], Reason = [Started], Message = [Started container filler-pod-0d7f8a65-b4c2-4299-ad5b-a838b5c282fa]
-STEP: Considering event: 
-Type = [Warning], Name = [additional-pod.16301bc6772d5d6f], Reason = [FailedScheduling], Message = [0/2 nodes are available: 2 Insufficient cpu.]
-STEP: Considering event: 
-Type = [Warning], Name = [additional-pod.16301bc6780f5291], Reason = [FailedScheduling], Message = [0/2 nodes are available: 2 Insufficient cpu.]
-STEP: removing the label node off the node adoring-wozniak-54dcfd79fc-6rshr
-STEP: verifying the node doesn't have the label node
-STEP: removing the label node off the node adoring-wozniak-54dcfd79fc-948mf
-STEP: verifying the node doesn't have the label node
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:26:48.001: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "sched-pred-4333" for this suite.
-[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
-•{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run  [Conformance]","total":280,"completed":213,"skipped":3457,"failed":0}
-SSSSSSSSSSS
+[BeforeEach] Atomic writer volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
+STEP: Setting up data
+[It] should support subpaths with downward pod [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating pod pod-subpath-test-downwardapi-kqck
+STEP: Creating a pod to test atomic-volume-subpath
+Feb 12 10:42:54.310: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-kqck" in namespace "subpath-9200" to be "success or failure"
+Feb 12 10:42:54.316: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Pending", Reason="", readiness=false. Elapsed: 6.175177ms
+Feb 12 10:42:56.323: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Running", Reason="", readiness=true. Elapsed: 2.012467318s
+Feb 12 10:42:58.329: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Running", Reason="", readiness=true. Elapsed: 4.018633516s
+Feb 12 10:43:00.336: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Running", Reason="", readiness=true. Elapsed: 6.025749715s
+Feb 12 10:43:02.342: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Running", Reason="", readiness=true. Elapsed: 8.031715219s
+Feb 12 10:43:04.348: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Running", Reason="", readiness=true. Elapsed: 10.038153921s
+Feb 12 10:43:06.355: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Running", Reason="", readiness=true. Elapsed: 12.044436826s
+Feb 12 10:43:08.361: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Running", Reason="", readiness=true. Elapsed: 14.050427424s
+Feb 12 10:43:10.367: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Running", Reason="", readiness=true. Elapsed: 16.056987665s
+Feb 12 10:43:12.374: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Running", Reason="", readiness=true. Elapsed: 18.063696841s
+Feb 12 10:43:14.383: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Running", Reason="", readiness=true. Elapsed: 20.072419024s
+Feb 12 10:43:16.390: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Running", Reason="", readiness=true. Elapsed: 22.079647825s
+Feb 12 10:43:18.398: INFO: Pod "pod-subpath-test-downwardapi-kqck": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.088247842s
+STEP: Saw pod success
+Feb 12 10:43:18.399: INFO: Pod "pod-subpath-test-downwardapi-kqck" satisfied condition "success or failure"
+Feb 12 10:43:18.405: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-subpath-test-downwardapi-kqck container test-container-subpath-downwardapi-kqck: 
+STEP: delete the pod
+Feb 12 10:43:18.488: INFO: Waiting for pod pod-subpath-test-downwardapi-kqck to disappear
+Feb 12 10:43:18.495: INFO: Pod pod-subpath-test-downwardapi-kqck no longer exists
+STEP: Deleting pod pod-subpath-test-downwardapi-kqck
+Feb 12 10:43:18.495: INFO: Deleting pod "pod-subpath-test-downwardapi-kqck" in namespace "subpath-9200"
+[AfterEach] [sig-storage] Subpath
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:43:18.502: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-9200" for this suite.
+
+• [SLOW TEST:24.426 seconds]
+[sig-storage] Subpath
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
+  Atomic writer volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
+    should support subpaths with downward pod [LinuxOnly] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  should perform rolling updates and roll backs of template modifications [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [LinuxOnly] [Conformance]","total":280,"completed":192,"skipped":3183,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:26:48.017: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename statefulset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-8626
+Feb 12 10:43:18.524: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-6274
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
-STEP: Creating service test in namespace statefulset-8626
-[It] should perform rolling updates and roll backs of template modifications [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a new StatefulSet
-Aug 30 17:26:48.208: INFO: Found 0 stateful pods, waiting for 3
-Aug 30 17:26:58.215: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
-Aug 30 17:26:58.215: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
-Aug 30 17:26:58.215: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
-Aug 30 17:26:58.231: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8626 ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Aug 30 17:26:58.767: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Aug 30 17:26:58.767: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Aug 30 17:26:58.767: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-STEP: Updating StatefulSet template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine
-Aug 30 17:27:08.809: INFO: Updating stateful set ss2
-STEP: Creating a new revision
-STEP: Updating Pods in reverse ordinal order
-Aug 30 17:27:18.843: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8626 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:27:19.362: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Aug 30 17:27:19.362: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Aug 30 17:27:19.362: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-STEP: Rolling back to a previous revision
-Aug 30 17:27:39.409: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8626 ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Aug 30 17:27:39.980: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Aug 30 17:27:39.980: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Aug 30 17:27:39.980: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Aug 30 17:27:50.022: INFO: Updating stateful set ss2
-STEP: Rolling back update in reverse ordinal order
-Aug 30 17:28:00.048: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8626 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:28:00.577: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Aug 30 17:28:00.577: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Aug 30 17:28:00.577: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Aug 30 17:28:00.606: INFO: Waiting for StatefulSet statefulset-8626/ss2 to complete update
-Aug 30 17:28:00.606: INFO: Waiting for Pod statefulset-8626/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
-Aug 30 17:28:00.606: INFO: Waiting for Pod statefulset-8626/ss2-1 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
-Aug 30 17:28:00.606: INFO: Waiting for Pod statefulset-8626/ss2-2 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
-Aug 30 17:28:10.618: INFO: Waiting for StatefulSet statefulset-8626/ss2 to complete update
-Aug 30 17:28:10.618: INFO: Waiting for Pod statefulset-8626/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
-Aug 30 17:28:20.617: INFO: Waiting for StatefulSet statefulset-8626/ss2 to complete update
-Aug 30 17:28:20.617: INFO: Waiting for Pod statefulset-8626/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
-Aug 30 17:28:30.618: INFO: Deleting all statefulset in ns statefulset-8626
-Aug 30 17:28:30.622: INFO: Scaling statefulset ss2 to 0
-Aug 30 17:28:40.644: INFO: Waiting for statefulset status.replicas updated to 0
-Aug 30 17:28:40.650: INFO: Deleting statefulset ss2
-[AfterEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:28:40.676: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-8626" for this suite.
-
-• [SLOW TEST:112.680 seconds]
-[sig-apps] StatefulSet
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-    should perform rolling updates and roll backs of template modifications [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance]","total":280,"completed":214,"skipped":3468,"failed":0}
-SSSSSSSSSSSSSSSSSSSS
+[It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir 0777 on node default medium
+Feb 12 10:43:18.713: INFO: Waiting up to 5m0s for pod "pod-03176d10-a3f1-40f5-ae74-41d69430f8e5" in namespace "emptydir-6274" to be "success or failure"
+Feb 12 10:43:18.721: INFO: Pod "pod-03176d10-a3f1-40f5-ae74-41d69430f8e5": Phase="Pending", Reason="", readiness=false. Elapsed: 7.725627ms
+Feb 12 10:43:20.728: INFO: Pod "pod-03176d10-a3f1-40f5-ae74-41d69430f8e5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014443537s
+Feb 12 10:43:22.733: INFO: Pod "pod-03176d10-a3f1-40f5-ae74-41d69430f8e5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.019842749s
+STEP: Saw pod success
+Feb 12 10:43:22.733: INFO: Pod "pod-03176d10-a3f1-40f5-ae74-41d69430f8e5" satisfied condition "success or failure"
+Feb 12 10:43:22.739: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-03176d10-a3f1-40f5-ae74-41d69430f8e5 container test-container: 
+STEP: delete the pod
+Feb 12 10:43:22.774: INFO: Waiting for pod pod-03176d10-a3f1-40f5-ae74-41d69430f8e5 to disappear
+Feb 12 10:43:22.780: INFO: Pod pod-03176d10-a3f1-40f5-ae74-41d69430f8e5 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:43:22.780: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-6274" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":193,"skipped":3233,"failed":0}
+SSSSSSSSSSSS
 ------------------------------
-[k8s.io] InitContainer [NodeConformance] 
-  should invoke init containers on a RestartAlways pod [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute prestop http hook properly [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Container Lifecycle Hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:28:40.697: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename init-container
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-144
+Feb 12 10:43:22.802: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-lifecycle-hook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-1299
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
-[It] should invoke init containers on a RestartAlways pod [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating the pod
-Aug 30 17:28:40.858: INFO: PodSpec: initContainers in spec.initContainers
-[AfterEach] [k8s.io] InitContainer [NodeConformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:28:44.834: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "init-container-144" for this suite.
-•{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance]","total":280,"completed":215,"skipped":3488,"failed":0}
-SSSS
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute prestop http hook properly [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: create the pod with lifecycle hook
+STEP: delete the pod with lifecycle hook
+Feb 12 10:43:27.023: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Feb 12 10:43:27.030: INFO: Pod pod-with-prestop-http-hook still exists
+Feb 12 10:43:29.031: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Feb 12 10:43:29.038: INFO: Pod pod-with-prestop-http-hook still exists
+Feb 12 10:43:31.031: INFO: Waiting for pod pod-with-prestop-http-hook to disappear
+Feb 12 10:43:31.037: INFO: Pod pod-with-prestop-http-hook no longer exists
+STEP: check prestop hook
+[AfterEach] [k8s.io] Container Lifecycle Hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:43:31.093: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-1299" for this suite.
+
+• [SLOW TEST:8.311 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  when create a pod with lifecycle hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
+    should execute prestop http hook properly [NodeConformance] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Downward API volume 
+{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance]","total":280,"completed":194,"skipped":3245,"failed":0}
+SSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Projected downwardAPI 
   should provide container's cpu limit [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:28:44.849: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7207
+Feb 12 10:43:31.117: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-2132
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
 [It] should provide container's cpu limit [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Creating a pod to test downward API volume plugin
-Aug 30 17:28:45.030: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fd46d762-cdcf-4874-a76f-bd5314b04e8d" in namespace "downward-api-7207" to be "success or failure"
-Aug 30 17:28:45.037: INFO: Pod "downwardapi-volume-fd46d762-cdcf-4874-a76f-bd5314b04e8d": Phase="Pending", Reason="", readiness=false. Elapsed: 6.110258ms
-Aug 30 17:28:47.042: INFO: Pod "downwardapi-volume-fd46d762-cdcf-4874-a76f-bd5314b04e8d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011982126s
+Feb 12 10:43:31.290: INFO: Waiting up to 5m0s for pod "downwardapi-volume-741f71b3-8514-4304-bdbe-17a4764e142d" in namespace "projected-2132" to be "success or failure"
+Feb 12 10:43:31.297: INFO: Pod "downwardapi-volume-741f71b3-8514-4304-bdbe-17a4764e142d": Phase="Pending", Reason="", readiness=false. Elapsed: 6.781151ms
+Feb 12 10:43:33.303: INFO: Pod "downwardapi-volume-741f71b3-8514-4304-bdbe-17a4764e142d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012722175s
 STEP: Saw pod success
-Aug 30 17:28:47.042: INFO: Pod "downwardapi-volume-fd46d762-cdcf-4874-a76f-bd5314b04e8d" satisfied condition "success or failure"
-Aug 30 17:28:47.048: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-fd46d762-cdcf-4874-a76f-bd5314b04e8d container client-container: 
+Feb 12 10:43:33.303: INFO: Pod "downwardapi-volume-741f71b3-8514-4304-bdbe-17a4764e142d" satisfied condition "success or failure"
+Feb 12 10:43:33.309: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-741f71b3-8514-4304-bdbe-17a4764e142d container client-container: 
 STEP: delete the pod
-Aug 30 17:28:47.120: INFO: Waiting for pod downwardapi-volume-fd46d762-cdcf-4874-a76f-bd5314b04e8d to disappear
-Aug 30 17:28:47.124: INFO: Pod downwardapi-volume-fd46d762-cdcf-4874-a76f-bd5314b04e8d no longer exists
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:28:47.124: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-7207" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance]","total":280,"completed":216,"skipped":3492,"failed":0}
-SSSSSSSSSSSSSSSSSSSSS
+Feb 12 10:43:33.358: INFO: Waiting for pod downwardapi-volume-741f71b3-8514-4304-bdbe-17a4764e142d to disappear
+Feb 12 10:43:33.364: INFO: Pod downwardapi-volume-741f71b3-8514-4304-bdbe-17a4764e142d no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:43:33.364: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-2132" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance]","total":280,"completed":195,"skipped":3266,"failed":0}
+
 ------------------------------
-[sig-cli] Kubectl client Kubectl run --rm job 
-  should create a job from an image, then delete the job  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Pods 
+  should contain environment variables for services [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:28:47.142: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-8842
+Feb 12 10:43:33.385: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-5084
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[It] should create a job from an image, then delete the job  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: executing a command with run --rm and attach with stdin
-Aug 30 17:28:47.312: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 --namespace=kubectl-8842 run e2e-test-rm-busybox-job --image=docker.io/library/busybox:1.29 --rm=true --generator=job/v1 --restart=OnFailure --attach=true --stdin -- sh -c cat && echo 'stdin closed''
-Aug 30 17:28:50.006: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\nIf you don't see a command prompt, try pressing enter.\n"
-Aug 30 17:28:50.006: INFO: stdout: "abcd1234stdin closed\njob.batch \"e2e-test-rm-busybox-job\" deleted\n"
-STEP: verifying the job e2e-test-rm-busybox-job was deleted
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:28:52.017: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-8842" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run --rm job should create a job from an image, then delete the job  [Conformance]","total":280,"completed":217,"skipped":3513,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
+[It] should contain environment variables for services [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:43:35.612: INFO: Waiting up to 5m0s for pod "client-envvars-6fd2336b-48f4-4477-90f7-ec5ca877586e" in namespace "pods-5084" to be "success or failure"
+Feb 12 10:43:35.620: INFO: Pod "client-envvars-6fd2336b-48f4-4477-90f7-ec5ca877586e": Phase="Pending", Reason="", readiness=false. Elapsed: 7.028495ms
+Feb 12 10:43:37.627: INFO: Pod "client-envvars-6fd2336b-48f4-4477-90f7-ec5ca877586e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014284202s
+STEP: Saw pod success
+Feb 12 10:43:37.627: INFO: Pod "client-envvars-6fd2336b-48f4-4477-90f7-ec5ca877586e" satisfied condition "success or failure"
+Feb 12 10:43:37.633: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod client-envvars-6fd2336b-48f4-4477-90f7-ec5ca877586e container env3cont: 
+STEP: delete the pod
+Feb 12 10:43:37.718: INFO: Waiting for pod client-envvars-6fd2336b-48f4-4477-90f7-ec5ca877586e to disappear
+Feb 12 10:43:37.729: INFO: Pod client-envvars-6fd2336b-48f4-4477-90f7-ec5ca877586e no longer exists
+[AfterEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:43:37.729: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-5084" for this suite.
+•{"msg":"PASSED [k8s.io] Pods should contain environment variables for services [NodeConformance] [Conformance]","total":280,"completed":196,"skipped":3266,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] ResourceQuota 
-  should create a ResourceQuota and capture the life of a replica set. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] Proxy version v1 
+  should proxy logs on node using proxy subresource  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] version v1
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:28:52.034: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename resourcequota
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-9570
+Feb 12 10:43:37.751: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename proxy
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-7865
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should create a ResourceQuota and capture the life of a replica set. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Counting existing ResourceQuota
-STEP: Creating a ResourceQuota
-STEP: Ensuring resource quota status is calculated
-STEP: Creating a ReplicaSet
-STEP: Ensuring resource quota status captures replicaset creation
-STEP: Deleting a ReplicaSet
-STEP: Ensuring resource quota status released usage
-[AfterEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:29:03.260: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-9570" for this suite.
-
-• [SLOW TEST:11.242 seconds]
-[sig-api-machinery] ResourceQuota
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should create a ResourceQuota and capture the life of a replica set. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance]","total":280,"completed":218,"skipped":3542,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[It] should proxy logs on node using proxy subresource  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:43:37.982: INFO: (0) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 54.152356ms)
+Feb 12 10:43:38.032: INFO: (1) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 49.976593ms)
+Feb 12 10:43:38.044: INFO: (2) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 12.085108ms)
+Feb 12 10:43:38.055: INFO: (3) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 10.831221ms)
+Feb 12 10:43:38.066: INFO: (4) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 10.272633ms)
+Feb 12 10:43:38.080: INFO: (5) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 14.051682ms)
+Feb 12 10:43:38.091: INFO: (6) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 10.951377ms)
+Feb 12 10:43:38.100: INFO: (7) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 8.908447ms)
+Feb 12 10:43:38.112: INFO: (8) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 12.181112ms)
+Feb 12 10:43:38.123: INFO: (9) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 10.804403ms)
+Feb 12 10:43:38.132: INFO: (10) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 9.371465ms)
+Feb 12 10:43:38.147: INFO: (11) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 14.381699ms)
+Feb 12 10:43:38.156: INFO: (12) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 9.284931ms)
+Feb 12 10:43:38.165: INFO: (13) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 8.505592ms)
+Feb 12 10:43:38.176: INFO: (14) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 11.396012ms)
+Feb 12 10:43:38.187: INFO: (15) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 10.462039ms)
+Feb 12 10:43:38.197: INFO: (16) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 9.656029ms)
+Feb 12 10:43:38.205: INFO: (17) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 8.910882ms)
+Feb 12 10:43:38.222: INFO: (18) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 16.10837ms)
+Feb 12 10:43:38.231: INFO: (19) /api/v1/nodes/nostalgic-fermat-cdc5d8777-r6snm/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 9.2081ms)
+[AfterEach] version v1
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:43:38.231: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "proxy-7865" for this suite.
+•{"msg":"PASSED [sig-network] Proxy version v1 should proxy logs on node using proxy subresource  [Conformance]","total":280,"completed":197,"skipped":3326,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir wrapper volumes 
-  should not cause race condition when used for configmaps [Serial] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir wrapper volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] ReplicaSet 
+  should adopt matching pods on creation and release no longer matching pods [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] ReplicaSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:29:03.276: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir-wrapper
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-wrapper-4906
+Feb 12 10:43:38.252: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename replicaset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replicaset-310
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not cause race condition when used for configmaps [Serial] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating 50 configmaps
-STEP: Creating RC which spawns configmap-volume pods
-Aug 30 17:29:03.835: INFO: Pod name wrapped-volume-race-c4fa8e8e-ab69-4cb6-bb58-48920bf0a2d6: Found 0 pods out of 5
-Aug 30 17:29:08.844: INFO: Pod name wrapped-volume-race-c4fa8e8e-ab69-4cb6-bb58-48920bf0a2d6: Found 5 pods out of 5
-STEP: Ensuring each pod is running
-STEP: deleting ReplicationController wrapped-volume-race-c4fa8e8e-ab69-4cb6-bb58-48920bf0a2d6 in namespace emptydir-wrapper-4906, will wait for the garbage collector to delete the pods
-Aug 30 17:29:18.948: INFO: Deleting ReplicationController wrapped-volume-race-c4fa8e8e-ab69-4cb6-bb58-48920bf0a2d6 took: 13.863546ms
-Aug 30 17:29:19.448: INFO: Terminating ReplicationController wrapped-volume-race-c4fa8e8e-ab69-4cb6-bb58-48920bf0a2d6 pods took: 500.205623ms
-STEP: Creating RC which spawns configmap-volume pods
-Aug 30 17:29:29.975: INFO: Pod name wrapped-volume-race-d492aa4c-7589-4850-bb3e-4bc4191c098f: Found 0 pods out of 5
-Aug 30 17:29:34.993: INFO: Pod name wrapped-volume-race-d492aa4c-7589-4850-bb3e-4bc4191c098f: Found 5 pods out of 5
-STEP: Ensuring each pod is running
-STEP: deleting ReplicationController wrapped-volume-race-d492aa4c-7589-4850-bb3e-4bc4191c098f in namespace emptydir-wrapper-4906, will wait for the garbage collector to delete the pods
-Aug 30 17:29:45.096: INFO: Deleting ReplicationController wrapped-volume-race-d492aa4c-7589-4850-bb3e-4bc4191c098f took: 13.417828ms
-Aug 30 17:29:45.596: INFO: Terminating ReplicationController wrapped-volume-race-d492aa4c-7589-4850-bb3e-4bc4191c098f pods took: 500.163726ms
-STEP: Creating RC which spawns configmap-volume pods
-Aug 30 17:30:04.321: INFO: Pod name wrapped-volume-race-ab147d24-b88e-45a3-974b-8b5f31125cd3: Found 0 pods out of 5
-Aug 30 17:30:09.333: INFO: Pod name wrapped-volume-race-ab147d24-b88e-45a3-974b-8b5f31125cd3: Found 5 pods out of 5
-STEP: Ensuring each pod is running
-STEP: deleting ReplicationController wrapped-volume-race-ab147d24-b88e-45a3-974b-8b5f31125cd3 in namespace emptydir-wrapper-4906, will wait for the garbage collector to delete the pods
-Aug 30 17:30:19.439: INFO: Deleting ReplicationController wrapped-volume-race-ab147d24-b88e-45a3-974b-8b5f31125cd3 took: 15.173264ms
-Aug 30 17:30:19.939: INFO: Terminating ReplicationController wrapped-volume-race-ab147d24-b88e-45a3-974b-8b5f31125cd3 pods took: 500.337345ms
-STEP: Cleaning up the configMaps
-[AfterEach] [sig-storage] EmptyDir wrapper volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:30:35.794: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-wrapper-4906" for this suite.
+[It] should adopt matching pods on creation and release no longer matching pods [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Given a Pod with a 'name' label pod-adoption-release is created
+STEP: When a replicaset with a matching selector is created
+STEP: Then the orphan pod is adopted
+STEP: When the matched label of one of its pods change
+Feb 12 10:43:43.477: INFO: Pod name pod-adoption-release: Found 1 pods out of 1
+STEP: Then the pod is released
+[AfterEach] [sig-apps] ReplicaSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:43:44.514: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replicaset-310" for this suite.
 
-• [SLOW TEST:92.534 seconds]
-[sig-storage] EmptyDir wrapper volumes
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
-  should not cause race condition when used for configmaps [Serial] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:6.288 seconds]
+[sig-apps] ReplicaSet
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should adopt matching pods on creation and release no longer matching pods [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance]","total":280,"completed":219,"skipped":3575,"failed":0}
-S
+{"msg":"PASSED [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance]","total":280,"completed":198,"skipped":3353,"failed":0}
+SSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  listing validating webhooks should work [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] Networking Granular Checks: Pods 
+  should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Networking
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:30:35.811: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-9547
+Feb 12 10:43:44.544: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename pod-network-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-6175
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Aug 30 17:30:36.355: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Aug 30 17:30:38.370: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405436, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405436, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405436, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405436, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Aug 30 17:30:41.396: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] listing validating webhooks should work [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Listing all of the created validation webhooks
-STEP: Creating a configMap that does not comply to the validation webhook rules
-STEP: Deleting the collection of validation webhooks
-STEP: Creating a configMap that does not comply to the validation webhook rules
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:30:41.771: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-9547" for this suite.
-STEP: Destroying namespace "webhook-9547-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+[It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Performing setup for networking test in namespace pod-network-test-6175
+STEP: creating a selector
+STEP: Creating the service pods in kubernetes
+Feb 12 10:43:44.715: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+STEP: Creating test pods
+Feb 12 10:44:06.860: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.25.2.66:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-6175 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:44:06.860: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:44:07.338: INFO: Found all expected endpoints: [netserver-0]
+Feb 12 10:44:07.343: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.25.0.41:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-6175 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:44:07.343: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:44:07.921: INFO: Found all expected endpoints: [netserver-1]
+Feb 12 10:44:07.932: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://172.25.1.183:8080/hostName | grep -v '^\s*$'] Namespace:pod-network-test-6175 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:44:07.932: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:44:08.536: INFO: Found all expected endpoints: [netserver-2]
+[AfterEach] [sig-network] Networking
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:44:08.537: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pod-network-test-6175" for this suite.
 
-• [SLOW TEST:6.057 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  listing validating webhooks should work [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:24.021 seconds]
+[sig-network] Networking
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
+  Granular Checks: Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
+    should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance]","total":280,"completed":220,"skipped":3576,"failed":0}
-SSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":199,"skipped":3364,"failed":0}
+SSSSSS
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should deny crd creation [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] ReplicationController 
+  should release no longer matching pods [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] ReplicationController
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:30:41.868: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-9343
+Feb 12 10:44:08.567: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename replication-controller
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-8586
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Aug 30 17:30:42.238: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Aug 30 17:30:44.262: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405442, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405442, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405442, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405442, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Aug 30 17:30:47.288: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should deny crd creation [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Registering the crd webhook via the AdmissionRegistration API
-STEP: Creating a custom resource definition that should be denied by the webhook
-Aug 30 17:30:47.476: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:30:47.579: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-9343" for this suite.
-STEP: Destroying namespace "webhook-9343-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+[It] should release no longer matching pods [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Given a ReplicationController is created
+STEP: When the matched label of one of its pods change
+Feb 12 10:44:08.740: INFO: Pod name pod-release: Found 0 pods out of 1
+Feb 12 10:44:13.748: INFO: Pod name pod-release: Found 1 pods out of 1
+STEP: Then the pod is released
+[AfterEach] [sig-apps] ReplicationController
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:44:14.774: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replication-controller-8586" for this suite.
 
-• [SLOW TEST:5.817 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should deny crd creation [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:6.225 seconds]
+[sig-apps] ReplicationController
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should release no longer matching pods [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance]","total":280,"completed":221,"skipped":3594,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-apps] ReplicationController should release no longer matching pods [Conformance]","total":280,"completed":200,"skipped":3370,"failed":0}
+SSS
 ------------------------------
 [sig-storage] Projected secret 
-  should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:30:47.689: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:44:14.792: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9024
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1695
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating projection with secret that has name projected-secret-test-map-846658d2-62c0-4dbd-9ba8-2838542a3d15
+[It] should be consumable from pods in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating projection with secret that has name projected-secret-test-9546e182-04a2-4abb-a796-888baa2da3d4
 STEP: Creating a pod to test consume secrets
-Aug 30 17:30:47.891: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-104ef37b-e543-4946-a4c9-51ef09d6adca" in namespace "projected-9024" to be "success or failure"
-Aug 30 17:30:47.896: INFO: Pod "pod-projected-secrets-104ef37b-e543-4946-a4c9-51ef09d6adca": Phase="Pending", Reason="", readiness=false. Elapsed: 5.410914ms
-Aug 30 17:30:49.902: INFO: Pod "pod-projected-secrets-104ef37b-e543-4946-a4c9-51ef09d6adca": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011388434s
+Feb 12 10:44:14.977: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-3e575019-def8-4b49-9892-ae59cba7abfe" in namespace "projected-1695" to be "success or failure"
+Feb 12 10:44:14.984: INFO: Pod "pod-projected-secrets-3e575019-def8-4b49-9892-ae59cba7abfe": Phase="Pending", Reason="", readiness=false. Elapsed: 6.888916ms
+Feb 12 10:44:16.990: INFO: Pod "pod-projected-secrets-3e575019-def8-4b49-9892-ae59cba7abfe": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012644371s
 STEP: Saw pod success
-Aug 30 17:30:49.902: INFO: Pod "pod-projected-secrets-104ef37b-e543-4946-a4c9-51ef09d6adca" satisfied condition "success or failure"
-Aug 30 17:30:49.907: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-projected-secrets-104ef37b-e543-4946-a4c9-51ef09d6adca container projected-secret-volume-test: 
+Feb 12 10:44:16.990: INFO: Pod "pod-projected-secrets-3e575019-def8-4b49-9892-ae59cba7abfe" satisfied condition "success or failure"
+Feb 12 10:44:16.996: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-projected-secrets-3e575019-def8-4b49-9892-ae59cba7abfe container projected-secret-volume-test: 
 STEP: delete the pod
-Aug 30 17:30:50.016: INFO: Waiting for pod pod-projected-secrets-104ef37b-e543-4946-a4c9-51ef09d6adca to disappear
-Aug 30 17:30:50.021: INFO: Pod pod-projected-secrets-104ef37b-e543-4946-a4c9-51ef09d6adca no longer exists
+Feb 12 10:44:17.073: INFO: Waiting for pod pod-projected-secrets-3e575019-def8-4b49-9892-ae59cba7abfe to disappear
+Feb 12 10:44:17.078: INFO: Pod pod-projected-secrets-3e575019-def8-4b49-9892-ae59cba7abfe no longer exists
 [AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:30:50.021: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-9024" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","total":280,"completed":222,"skipped":3619,"failed":0}
-SSS
-------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should mutate custom resource with pruning [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:30:50.040: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-7776
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Aug 30 17:30:50.900: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Aug 30 17:30:52.916: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405450, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405450, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405450, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405450, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Aug 30 17:30:55.936: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should mutate custom resource with pruning [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:30:55.942: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Registering the mutating webhook for custom resource e2e-test-webhook-3188-crds.webhook.example.com via the AdmissionRegistration API
-STEP: Creating a custom resource that should be mutated by the webhook
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:30:57.324: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-7776" for this suite.
-STEP: Destroying namespace "webhook-7776-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
-
-• [SLOW TEST:7.364 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should mutate custom resource with pruning [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance]","total":280,"completed":223,"skipped":3622,"failed":0}
-SSSSS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:44:17.078: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-1695" for this suite.
+•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance]","total":280,"completed":201,"skipped":3373,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Secrets 
-  should fail to create secret due to empty secret key [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Secrets 
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:30:57.405: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:44:17.100: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-7672
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should fail to create secret due to empty secret key [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating projection with secret that has name secret-emptykey-test-593f6803-a3c9-431f-9329-95f18086aaa4
-[AfterEach] [sig-api-machinery] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:30:57.571: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-7672" for this suite.
-•{"msg":"PASSED [sig-api-machinery] Secrets should fail to create secret due to empty secret key [Conformance]","total":280,"completed":224,"skipped":3627,"failed":0}
-SSSSSSSSS
-------------------------------
-[sig-storage] Projected secret 
-  should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:30:57.589: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-1153
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-1823
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating projection with secret that has name projected-secret-test-map-708fe94d-32b6-4c04-923a-9acc8ae9f1e8
+[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating secret with name secret-test-9a7ef819-b0a1-4aea-ba5a-759249fa4678
 STEP: Creating a pod to test consume secrets
-Aug 30 17:30:57.767: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-a6dfb05d-279d-45a8-b198-3757ba596b39" in namespace "projected-1153" to be "success or failure"
-Aug 30 17:30:57.772: INFO: Pod "pod-projected-secrets-a6dfb05d-279d-45a8-b198-3757ba596b39": Phase="Pending", Reason="", readiness=false. Elapsed: 4.472381ms
-Aug 30 17:30:59.778: INFO: Pod "pod-projected-secrets-a6dfb05d-279d-45a8-b198-3757ba596b39": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010426382s
-STEP: Saw pod success
-Aug 30 17:30:59.778: INFO: Pod "pod-projected-secrets-a6dfb05d-279d-45a8-b198-3757ba596b39" satisfied condition "success or failure"
-Aug 30 17:30:59.784: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-projected-secrets-a6dfb05d-279d-45a8-b198-3757ba596b39 container projected-secret-volume-test: 
-STEP: delete the pod
-Aug 30 17:30:59.852: INFO: Waiting for pod pod-projected-secrets-a6dfb05d-279d-45a8-b198-3757ba596b39 to disappear
-Aug 30 17:30:59.860: INFO: Pod pod-projected-secrets-a6dfb05d-279d-45a8-b198-3757ba596b39 no longer exists
-[AfterEach] [sig-storage] Projected secret
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:30:59.860: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-1153" for this suite.
-•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":225,"skipped":3636,"failed":0}
-SSSSSSSSSSSSSSSSSS
-------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:30:59.875: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1506
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0666 on node default medium
-Aug 30 17:31:00.050: INFO: Waiting up to 5m0s for pod "pod-6efff8f3-ba97-4487-8958-81320d2e4e21" in namespace "emptydir-1506" to be "success or failure"
-Aug 30 17:31:00.059: INFO: Pod "pod-6efff8f3-ba97-4487-8958-81320d2e4e21": Phase="Pending", Reason="", readiness=false. Elapsed: 8.813133ms
-Aug 30 17:31:02.074: INFO: Pod "pod-6efff8f3-ba97-4487-8958-81320d2e4e21": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.024239199s
+Feb 12 10:44:17.305: INFO: Waiting up to 5m0s for pod "pod-secrets-b6558930-543f-4d14-a24b-a391af605109" in namespace "secrets-1823" to be "success or failure"
+Feb 12 10:44:17.312: INFO: Pod "pod-secrets-b6558930-543f-4d14-a24b-a391af605109": Phase="Pending", Reason="", readiness=false. Elapsed: 6.624ms
+Feb 12 10:44:19.320: INFO: Pod "pod-secrets-b6558930-543f-4d14-a24b-a391af605109": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.01453698s
 STEP: Saw pod success
-Aug 30 17:31:02.074: INFO: Pod "pod-6efff8f3-ba97-4487-8958-81320d2e4e21" satisfied condition "success or failure"
-Aug 30 17:31:02.082: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-6efff8f3-ba97-4487-8958-81320d2e4e21 container test-container: 
+Feb 12 10:44:19.320: INFO: Pod "pod-secrets-b6558930-543f-4d14-a24b-a391af605109" satisfied condition "success or failure"
+Feb 12 10:44:19.326: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-secrets-b6558930-543f-4d14-a24b-a391af605109 container secret-volume-test: 
 STEP: delete the pod
-Aug 30 17:31:02.166: INFO: Waiting for pod pod-6efff8f3-ba97-4487-8958-81320d2e4e21 to disappear
-Aug 30 17:31:02.176: INFO: Pod pod-6efff8f3-ba97-4487-8958-81320d2e4e21 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:31:02.176: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-1506" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":226,"skipped":3654,"failed":0}
-SSSSSSSSSSSSSSSS
+Feb 12 10:44:19.371: INFO: Waiting for pod pod-secrets-b6558930-543f-4d14-a24b-a391af605109 to disappear
+Feb 12 10:44:19.377: INFO: Pod pod-secrets-b6558930-543f-4d14-a24b-a391af605109 no longer exists
+[AfterEach] [sig-storage] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:44:19.377: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-1823" for this suite.
+•{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":202,"skipped":3410,"failed":0}
+SSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-auth] ServiceAccounts 
-  should mount an API token into pods  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-auth] ServiceAccounts
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook 
+  should execute poststart exec hook properly [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Container Lifecycle Hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:31:02.196: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename svcaccounts
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svcaccounts-2493
+Feb 12 10:44:19.402: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-lifecycle-hook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-lifecycle-hook-2095
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should mount an API token into pods  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: getting the auto-created API token
-STEP: reading a file in the container
-Aug 30 17:31:06.943: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-2493 pod-service-account-1ff6b08c-d972-4a7d-a92c-6cef1506c747 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token'
-STEP: reading a file in the container
-Aug 30 17:31:07.493: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-2493 pod-service-account-1ff6b08c-d972-4a7d-a92c-6cef1506c747 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt'
-STEP: reading a file in the container
-Aug 30 17:31:08.064: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-2493 pod-service-account-1ff6b08c-d972-4a7d-a92c-6cef1506c747 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace'
-[AfterEach] [sig-auth] ServiceAccounts
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:31:08.661: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "svcaccounts-2493" for this suite.
+[BeforeEach] when create a pod with lifecycle hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:64
+STEP: create the container to handle the HTTPGet hook request.
+[It] should execute poststart exec hook properly [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: create the pod with lifecycle hook
+STEP: check poststart hook
+STEP: delete the pod with lifecycle hook
+Feb 12 10:44:27.682: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Feb 12 10:44:27.689: INFO: Pod pod-with-poststart-exec-hook still exists
+Feb 12 10:44:29.689: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Feb 12 10:44:29.697: INFO: Pod pod-with-poststart-exec-hook still exists
+Feb 12 10:44:31.689: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Feb 12 10:44:31.696: INFO: Pod pod-with-poststart-exec-hook still exists
+Feb 12 10:44:33.689: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear
+Feb 12 10:44:33.695: INFO: Pod pod-with-poststart-exec-hook no longer exists
+[AfterEach] [k8s.io] Container Lifecycle Hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:44:33.695: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-lifecycle-hook-2095" for this suite.
 
-• [SLOW TEST:6.481 seconds]
-[sig-auth] ServiceAccounts
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/auth/framework.go:23
-  should mount an API token into pods  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-auth] ServiceAccounts should mount an API token into pods  [Conformance]","total":280,"completed":227,"skipped":3670,"failed":0}
-SSSSSSSSSSSSS
+• [SLOW TEST:14.323 seconds]
+[k8s.io] Container Lifecycle Hook
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  when create a pod with lifecycle hook
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/lifecycle_hook.go:42
+    should execute poststart exec hook properly [NodeConformance] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-api-machinery] ResourceQuota 
-  should be able to update and delete ResourceQuota. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:31:08.677: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename resourcequota
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-4757
-STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to update and delete ResourceQuota. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a ResourceQuota
-STEP: Getting a ResourceQuota
-STEP: Updating a ResourceQuota
-STEP: Verifying a ResourceQuota was modified
-STEP: Deleting a ResourceQuota
-STEP: Verifying the deleted ResourceQuota
-[AfterEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:31:08.881: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-4757" for this suite.
-•{"msg":"PASSED [sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance]","total":280,"completed":228,"skipped":3683,"failed":0}
-SSSSSSSS
+{"msg":"PASSED [k8s.io] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance]","total":280,"completed":203,"skipped":3431,"failed":0}
+SSSSS
 ------------------------------
-[sig-apps] ReplicationController 
-  should surface a failure condition on a common issue like exceeded quota [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-cli] Kubectl client Kubectl patch 
+  should add annotations for pods in rc  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:31:08.897: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename replication-controller
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-4947
+Feb 12 10:44:33.730: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7992
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should surface a failure condition on a common issue like exceeded quota [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:31:09.064: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace
-STEP: Creating rc "condition-test" that asks for more than the allowed pod quota
-STEP: Checking rc "condition-test" has the desired failure condition set
-STEP: Scaling down rc "condition-test" to satisfy pod quota
-Aug 30 17:31:11.119: INFO: Updating replication controller "condition-test"
-STEP: Checking rc "condition-test" has no failure condition set
-[AfterEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:31:11.123: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replication-controller-4947" for this suite.
-•{"msg":"PASSED [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance]","total":280,"completed":229,"skipped":3691,"failed":0}
-SSSSSS
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[It] should add annotations for pods in rc  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating Agnhost RC
+Feb 12 10:44:33.888: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-7992'
+Feb 12 10:44:34.080: INFO: stderr: ""
+Feb 12 10:44:34.080: INFO: stdout: "replicationcontroller/agnhost-master created\n"
+STEP: Waiting for Agnhost master to start.
+Feb 12 10:44:35.086: INFO: Selector matched 1 pods for map[app:agnhost]
+Feb 12 10:44:35.086: INFO: Found 0 / 1
+Feb 12 10:44:36.086: INFO: Selector matched 1 pods for map[app:agnhost]
+Feb 12 10:44:36.086: INFO: Found 0 / 1
+Feb 12 10:44:37.085: INFO: Selector matched 1 pods for map[app:agnhost]
+Feb 12 10:44:37.085: INFO: Found 1 / 1
+Feb 12 10:44:37.085: INFO: WaitFor completed with timeout 5m0s.  Pods found = 1 out of 1
+STEP: patching all pods
+Feb 12 10:44:37.092: INFO: Selector matched 1 pods for map[app:agnhost]
+Feb 12 10:44:37.092: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+Feb 12 10:44:37.092: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 patch pod agnhost-master-dqrgv --namespace=kubectl-7992 -p {"metadata":{"annotations":{"x":"y"}}}'
+Feb 12 10:44:37.187: INFO: stderr: ""
+Feb 12 10:44:37.187: INFO: stdout: "pod/agnhost-master-dqrgv patched\n"
+STEP: checking annotations
+Feb 12 10:44:37.193: INFO: Selector matched 1 pods for map[app:agnhost]
+Feb 12 10:44:37.193: INFO: ForEach: Found 1 pods from the filter.  Now looping through them.
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:44:37.194: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-7992" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc  [Conformance]","total":280,"completed":204,"skipped":3436,"failed":0}
+SSSSSSSSSSSSS
 ------------------------------
-[sig-storage] ConfigMap 
-  should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Projected secret 
+  should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:31:11.141: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-2587
+Feb 12 10:44:37.214: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9579
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name configmap-test-volume-map-dde0efd5-0c0f-4946-be10-fec9ec20f9bf
-STEP: Creating a pod to test consume configMaps
-Aug 30 17:31:11.318: INFO: Waiting up to 5m0s for pod "pod-configmaps-f901b73a-9ec3-405f-9106-b7fb78fe8e04" in namespace "configmap-2587" to be "success or failure"
-Aug 30 17:31:11.322: INFO: Pod "pod-configmaps-f901b73a-9ec3-405f-9106-b7fb78fe8e04": Phase="Pending", Reason="", readiness=false. Elapsed: 4.331573ms
-Aug 30 17:31:13.329: INFO: Pod "pod-configmaps-f901b73a-9ec3-405f-9106-b7fb78fe8e04": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010574837s
+[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating projection with secret that has name projected-secret-test-7c1d4f7b-6e96-41ca-a520-e8313042fa4d
+STEP: Creating a pod to test consume secrets
+Feb 12 10:44:37.481: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-fc159b69-67ea-4c3d-8a80-0b0ec575829e" in namespace "projected-9579" to be "success or failure"
+Feb 12 10:44:37.490: INFO: Pod "pod-projected-secrets-fc159b69-67ea-4c3d-8a80-0b0ec575829e": Phase="Pending", Reason="", readiness=false. Elapsed: 9.346556ms
+Feb 12 10:44:39.496: INFO: Pod "pod-projected-secrets-fc159b69-67ea-4c3d-8a80-0b0ec575829e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015297471s
 STEP: Saw pod success
-Aug 30 17:31:13.329: INFO: Pod "pod-configmaps-f901b73a-9ec3-405f-9106-b7fb78fe8e04" satisfied condition "success or failure"
-Aug 30 17:31:13.334: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-configmaps-f901b73a-9ec3-405f-9106-b7fb78fe8e04 container configmap-volume-test: 
+Feb 12 10:44:39.496: INFO: Pod "pod-projected-secrets-fc159b69-67ea-4c3d-8a80-0b0ec575829e" satisfied condition "success or failure"
+Feb 12 10:44:39.502: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-projected-secrets-fc159b69-67ea-4c3d-8a80-0b0ec575829e container projected-secret-volume-test: 
 STEP: delete the pod
-Aug 30 17:31:13.402: INFO: Waiting for pod pod-configmaps-f901b73a-9ec3-405f-9106-b7fb78fe8e04 to disappear
-Aug 30 17:31:13.407: INFO: Pod pod-configmaps-f901b73a-9ec3-405f-9106-b7fb78fe8e04 no longer exists
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:31:13.408: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-2587" for this suite.
-•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":230,"skipped":3697,"failed":0}
+Feb 12 10:44:39.571: INFO: Waiting for pod pod-projected-secrets-fc159b69-67ea-4c3d-8a80-0b0ec575829e to disappear
+Feb 12 10:44:39.575: INFO: Pod pod-projected-secrets-fc159b69-67ea-4c3d-8a80-0b0ec575829e no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:44:39.575: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9579" for this suite.
+•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":205,"skipped":3449,"failed":0}
 SS
 ------------------------------
-[sig-api-machinery] Garbage collector 
-  should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-cli] Kubectl client Kubectl label 
+  should update the label on a resource  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:31:13.426: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-8156
+Feb 12 10:44:39.593: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-4570
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the rc
-STEP: delete the rc
-STEP: wait for the rc to be deleted
-STEP: Gathering metrics
-W0830 17:31:19.635251      23 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-Aug 30 17:31:19.635: INFO: For apiserver_request_total:
-For apiserver_request_latency_seconds:
-For apiserver_init_events_total:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
-
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:31:19.635: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-8156" for this suite.
-
-• [SLOW TEST:6.223 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]","total":280,"completed":231,"skipped":3699,"failed":0}
-SSSSSSSSSSSSS
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[BeforeEach] Kubectl label
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1276
+STEP: creating the pod
+Feb 12 10:44:39.757: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 create -f - --namespace=kubectl-4570'
+Feb 12 10:44:39.982: INFO: stderr: ""
+Feb 12 10:44:39.982: INFO: stdout: "pod/pause created\n"
+Feb 12 10:44:39.982: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause]
+Feb 12 10:44:39.982: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-4570" to be "running and ready"
+Feb 12 10:44:39.988: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 6.434164ms
+Feb 12 10:44:41.995: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 2.013078879s
+Feb 12 10:44:41.995: INFO: Pod "pause" satisfied condition "running and ready"
+Feb 12 10:44:41.995: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause]
+[It] should update the label on a resource  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: adding the label testing-label with value testing-label-value to a pod
+Feb 12 10:44:41.995: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 label pods pause testing-label=testing-label-value --namespace=kubectl-4570'
+Feb 12 10:44:42.080: INFO: stderr: ""
+Feb 12 10:44:42.080: INFO: stdout: "pod/pause labeled\n"
+STEP: verifying the pod has the label testing-label with the value testing-label-value
+Feb 12 10:44:42.081: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pod pause -L testing-label --namespace=kubectl-4570'
+Feb 12 10:44:42.153: INFO: stderr: ""
+Feb 12 10:44:42.153: INFO: stdout: "NAME    READY   STATUS    RESTARTS   AGE   TESTING-LABEL\npause   1/1     Running   0          2s    testing-label-value\n"
+STEP: removing the label testing-label of a pod
+Feb 12 10:44:42.153: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 label pods pause testing-label- --namespace=kubectl-4570'
+Feb 12 10:44:42.259: INFO: stderr: ""
+Feb 12 10:44:42.259: INFO: stdout: "pod/pause labeled\n"
+STEP: verifying the pod doesn't have the label testing-label
+Feb 12 10:44:42.259: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pod pause -L testing-label --namespace=kubectl-4570'
+Feb 12 10:44:42.331: INFO: stderr: ""
+Feb 12 10:44:42.331: INFO: stdout: "NAME    READY   STATUS    RESTARTS   AGE   TESTING-LABEL\npause   1/1     Running   0          2s    \n"
+[AfterEach] Kubectl label
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1283
+STEP: using delete to clean up resources
+Feb 12 10:44:42.331: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete --grace-period=0 --force -f - --namespace=kubectl-4570'
+Feb 12 10:44:42.419: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
+Feb 12 10:44:42.419: INFO: stdout: "pod \"pause\" force deleted\n"
+Feb 12 10:44:42.419: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get rc,svc -l name=pause --no-headers --namespace=kubectl-4570'
+Feb 12 10:44:42.508: INFO: stderr: "No resources found in kubectl-4570 namespace.\n"
+Feb 12 10:44:42.508: INFO: stdout: ""
+Feb 12 10:44:42.508: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 get pods -l name=pause --namespace=kubectl-4570 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
+Feb 12 10:44:42.578: INFO: stderr: ""
+Feb 12 10:44:42.578: INFO: stdout: ""
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:44:42.578: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-4570" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl label should update the label on a resource  [Conformance]","total":280,"completed":206,"skipped":3451,"failed":0}
+SSSSSSSSS
 ------------------------------
 [sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with configmap pod [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should support subpaths with projected pod [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] Subpath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:31:19.649: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:44:42.595: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename subpath
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-5691
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-9425
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
 STEP: Setting up data
-[It] should support subpaths with configmap pod [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod pod-subpath-test-configmap-gxhd
+[It] should support subpaths with projected pod [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating pod pod-subpath-test-projected-96zq
 STEP: Creating a pod to test atomic-volume-subpath
-Aug 30 17:31:19.836: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-gxhd" in namespace "subpath-5691" to be "success or failure"
-Aug 30 17:31:19.845: INFO: Pod "pod-subpath-test-configmap-gxhd": Phase="Pending", Reason="", readiness=false. Elapsed: 8.927303ms
-Aug 30 17:31:21.855: INFO: Pod "pod-subpath-test-configmap-gxhd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018780932s
-Aug 30 17:31:23.861: INFO: Pod "pod-subpath-test-configmap-gxhd": Phase="Pending", Reason="", readiness=false. Elapsed: 4.024870698s
-Aug 30 17:31:25.867: INFO: Pod "pod-subpath-test-configmap-gxhd": Phase="Running", Reason="", readiness=true. Elapsed: 6.030985585s
-Aug 30 17:31:27.873: INFO: Pod "pod-subpath-test-configmap-gxhd": Phase="Running", Reason="", readiness=true. Elapsed: 8.036884194s
-Aug 30 17:31:29.881: INFO: Pod "pod-subpath-test-configmap-gxhd": Phase="Running", Reason="", readiness=true. Elapsed: 10.044861084s
-Aug 30 17:31:31.887: INFO: Pod "pod-subpath-test-configmap-gxhd": Phase="Running", Reason="", readiness=true. Elapsed: 12.050756921s
-Aug 30 17:31:33.893: INFO: Pod "pod-subpath-test-configmap-gxhd": Phase="Running", Reason="", readiness=true. Elapsed: 14.056471882s
-Aug 30 17:31:35.899: INFO: Pod "pod-subpath-test-configmap-gxhd": Phase="Running", Reason="", readiness=true. Elapsed: 16.062489451s
-Aug 30 17:31:37.904: INFO: Pod "pod-subpath-test-configmap-gxhd": Phase="Running", Reason="", readiness=true. Elapsed: 18.068172415s
-Aug 30 17:31:39.910: INFO: Pod "pod-subpath-test-configmap-gxhd": Phase="Running", Reason="", readiness=true. Elapsed: 20.073839055s
-Aug 30 17:31:41.916: INFO: Pod "pod-subpath-test-configmap-gxhd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.079966687s
+Feb 12 10:44:42.788: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-96zq" in namespace "subpath-9425" to be "success or failure"
+Feb 12 10:44:42.793: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Pending", Reason="", readiness=false. Elapsed: 5.269624ms
+Feb 12 10:44:44.806: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01870472s
+Feb 12 10:44:46.813: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Running", Reason="", readiness=true. Elapsed: 4.025234026s
+Feb 12 10:44:48.821: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Running", Reason="", readiness=true. Elapsed: 6.033081309s
+Feb 12 10:44:50.827: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Running", Reason="", readiness=true. Elapsed: 8.038829896s
+Feb 12 10:44:52.834: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Running", Reason="", readiness=true. Elapsed: 10.046698642s
+Feb 12 10:44:54.841: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Running", Reason="", readiness=true. Elapsed: 12.053453644s
+Feb 12 10:44:56.847: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Running", Reason="", readiness=true. Elapsed: 14.059230565s
+Feb 12 10:44:58.853: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Running", Reason="", readiness=true. Elapsed: 16.065507577s
+Feb 12 10:45:00.865: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Running", Reason="", readiness=true. Elapsed: 18.077151946s
+Feb 12 10:45:02.871: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Running", Reason="", readiness=true. Elapsed: 20.083457471s
+Feb 12 10:45:04.878: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Running", Reason="", readiness=true. Elapsed: 22.090015051s
+Feb 12 10:45:06.885: INFO: Pod "pod-subpath-test-projected-96zq": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.097074837s
 STEP: Saw pod success
-Aug 30 17:31:41.916: INFO: Pod "pod-subpath-test-configmap-gxhd" satisfied condition "success or failure"
-Aug 30 17:31:41.921: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-subpath-test-configmap-gxhd container test-container-subpath-configmap-gxhd: 
+Feb 12 10:45:06.885: INFO: Pod "pod-subpath-test-projected-96zq" satisfied condition "success or failure"
+Feb 12 10:45:06.890: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-subpath-test-projected-96zq container test-container-subpath-projected-96zq: 
 STEP: delete the pod
-Aug 30 17:31:41.988: INFO: Waiting for pod pod-subpath-test-configmap-gxhd to disappear
-Aug 30 17:31:41.992: INFO: Pod pod-subpath-test-configmap-gxhd no longer exists
-STEP: Deleting pod pod-subpath-test-configmap-gxhd
-Aug 30 17:31:41.992: INFO: Deleting pod "pod-subpath-test-configmap-gxhd" in namespace "subpath-5691"
+Feb 12 10:45:06.927: INFO: Waiting for pod pod-subpath-test-projected-96zq to disappear
+Feb 12 10:45:06.932: INFO: Pod pod-subpath-test-projected-96zq no longer exists
+STEP: Deleting pod pod-subpath-test-projected-96zq
+Feb 12 10:45:06.932: INFO: Deleting pod "pod-subpath-test-projected-96zq" in namespace "subpath-9425"
 [AfterEach] [sig-storage] Subpath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:31:41.997: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "subpath-5691" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:45:06.937: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-9425" for this suite.
 
-• [SLOW TEST:22.362 seconds]
+• [SLOW TEST:24.361 seconds]
 [sig-storage] Subpath
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
   Atomic writer volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
-    should support subpaths with configmap pod [LinuxOnly] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
+    should support subpaths with projected pod [LinuxOnly] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [LinuxOnly] [Conformance]","total":280,"completed":232,"skipped":3712,"failed":0}
-SSSSSSSS
+{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance]","total":280,"completed":207,"skipped":3460,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] Subpath Atomic writer volumes 
-  should support subpaths with projected pod [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Subpath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] DNS 
+  should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:31:42.011: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename subpath
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-2675
+Feb 12 10:45:06.956: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename dns
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-7125
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] Atomic writer volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
-STEP: Setting up data
-[It] should support subpaths with projected pod [LinuxOnly] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating pod pod-subpath-test-projected-q92q
-STEP: Creating a pod to test atomic-volume-subpath
-Aug 30 17:31:42.203: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-q92q" in namespace "subpath-2675" to be "success or failure"
-Aug 30 17:31:42.207: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Pending", Reason="", readiness=false. Elapsed: 4.418265ms
-Aug 30 17:31:44.213: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Running", Reason="", readiness=true. Elapsed: 2.010213245s
-Aug 30 17:31:46.219: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Running", Reason="", readiness=true. Elapsed: 4.016132822s
-Aug 30 17:31:48.224: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Running", Reason="", readiness=true. Elapsed: 6.021197058s
-Aug 30 17:31:50.230: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Running", Reason="", readiness=true. Elapsed: 8.026919432s
-Aug 30 17:31:52.236: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Running", Reason="", readiness=true. Elapsed: 10.032997155s
-Aug 30 17:31:54.242: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Running", Reason="", readiness=true. Elapsed: 12.03924511s
-Aug 30 17:31:56.247: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Running", Reason="", readiness=true. Elapsed: 14.044575306s
-Aug 30 17:31:58.253: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Running", Reason="", readiness=true. Elapsed: 16.050515043s
-Aug 30 17:32:00.259: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Running", Reason="", readiness=true. Elapsed: 18.055941643s
-Aug 30 17:32:02.264: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Running", Reason="", readiness=true. Elapsed: 20.061725201s
-Aug 30 17:32:04.271: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Running", Reason="", readiness=true. Elapsed: 22.068031518s
-Aug 30 17:32:06.277: INFO: Pod "pod-subpath-test-projected-q92q": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.074033167s
-STEP: Saw pod success
-Aug 30 17:32:06.277: INFO: Pod "pod-subpath-test-projected-q92q" satisfied condition "success or failure"
-Aug 30 17:32:06.282: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-subpath-test-projected-q92q container test-container-subpath-projected-q92q: 
-STEP: delete the pod
-Aug 30 17:32:06.362: INFO: Waiting for pod pod-subpath-test-projected-q92q to disappear
-Aug 30 17:32:06.369: INFO: Pod pod-subpath-test-projected-q92q no longer exists
-STEP: Deleting pod pod-subpath-test-projected-q92q
-Aug 30 17:32:06.369: INFO: Deleting pod "pod-subpath-test-projected-q92q" in namespace "subpath-2675"
-[AfterEach] [sig-storage] Subpath
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:32:06.374: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "subpath-2675" for this suite.
+[It] should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a test headless service
+STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-7125 A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-7125;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-7125 A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-7125;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-7125.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-7125.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-7125.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-7125.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-7125.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-7125.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-7125.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-7125.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-7125.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-7125.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-7125.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-7125.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-7125.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 248.23.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.23.248_udp@PTR;check="$$(dig +tcp +noall +answer +search 248.23.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.23.248_tcp@PTR;sleep 1; done
 
-• [SLOW TEST:24.378 seconds]
-[sig-storage] Subpath
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
-  Atomic writer volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
-    should support subpaths with projected pod [LinuxOnly] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-7125 A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-7125;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-7125 A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-7125;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-7125.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-7125.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-7125.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-7125.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-7125.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-7125.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-7125.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-7125.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-7125.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-7125.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-7125.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-7125.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".dns-7125.pod.cluster.local"}');check="$$(dig +notcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_udp@PodARecord;check="$$(dig +tcp +noall +answer +search $${podARec} A)" && test -n "$$check" && echo OK > /results/jessie_tcp@PodARecord;check="$$(dig +notcp +noall +answer +search 248.23.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.23.248_udp@PTR;check="$$(dig +tcp +noall +answer +search 248.23.240.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.240.23.248_tcp@PTR;sleep 1; done
+
+STEP: creating a pod to probe DNS
+STEP: submitting the pod to kubernetes
+STEP: retrieving the pod
+STEP: looking for the results for each expected name from probers
+Feb 12 10:45:11.281: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.324: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.336: INFO: Unable to read wheezy_udp@dns-test-service.dns-7125 from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.345: INFO: Unable to read wheezy_tcp@dns-test-service.dns-7125 from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.362: INFO: Unable to read wheezy_udp@dns-test-service.dns-7125.svc from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.374: INFO: Unable to read wheezy_tcp@dns-test-service.dns-7125.svc from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.385: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-7125.svc from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.397: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-7125.svc from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.928: INFO: Unable to read jessie_udp@dns-test-service from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.938: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.947: INFO: Unable to read jessie_udp@dns-test-service.dns-7125 from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.956: INFO: Unable to read jessie_tcp@dns-test-service.dns-7125 from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.967: INFO: Unable to read jessie_udp@dns-test-service.dns-7125.svc from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.975: INFO: Unable to read jessie_tcp@dns-test-service.dns-7125.svc from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.985: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-7125.svc from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:11.994: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-7125.svc from pod dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5: the server could not find the requested resource (get pods dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5)
+Feb 12 10:45:12.478: INFO: Lookups using dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-7125 wheezy_tcp@dns-test-service.dns-7125 wheezy_udp@dns-test-service.dns-7125.svc wheezy_tcp@dns-test-service.dns-7125.svc wheezy_udp@_http._tcp.dns-test-service.dns-7125.svc wheezy_tcp@_http._tcp.dns-test-service.dns-7125.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-7125 jessie_tcp@dns-test-service.dns-7125 jessie_udp@dns-test-service.dns-7125.svc jessie_tcp@dns-test-service.dns-7125.svc jessie_udp@_http._tcp.dns-test-service.dns-7125.svc jessie_tcp@_http._tcp.dns-test-service.dns-7125.svc]
+
+Feb 12 10:45:34.778: INFO: DNS probes using dns-7125/dns-test-c59121c4-1cef-4f83-a10f-7234458ef4f5 succeeded
+
+STEP: deleting the pod
+STEP: deleting the test service
+STEP: deleting the test headless service
+[AfterEach] [sig-network] DNS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:45:34.870: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-7125" for this suite.
+
+• [SLOW TEST:27.929 seconds]
+[sig-network] DNS
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [LinuxOnly] [Conformance]","total":280,"completed":233,"skipped":3720,"failed":0}
-SSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]","total":280,"completed":208,"skipped":3519,"failed":0}
+[sig-storage] Projected combined 
+  should project all components that make up the projection API [Projection][NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected combined
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:45:34.886: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8188
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name configmap-projected-all-test-volume-4efeacd9-78bd-43ac-a67d-7dc0f163eb51
+STEP: Creating secret with name secret-projected-all-test-volume-214df027-d4e0-4b5d-94c1-a813d0a60ca6
+STEP: Creating a pod to test Check all projections for projected volume plugin
+Feb 12 10:45:35.094: INFO: Waiting up to 5m0s for pod "projected-volume-b00a278d-698c-4a8a-bbb9-08328451a9f0" in namespace "projected-8188" to be "success or failure"
+Feb 12 10:45:35.100: INFO: Pod "projected-volume-b00a278d-698c-4a8a-bbb9-08328451a9f0": Phase="Pending", Reason="", readiness=false. Elapsed: 5.620859ms
+Feb 12 10:45:37.111: INFO: Pod "projected-volume-b00a278d-698c-4a8a-bbb9-08328451a9f0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.017355162s
+STEP: Saw pod success
+Feb 12 10:45:37.112: INFO: Pod "projected-volume-b00a278d-698c-4a8a-bbb9-08328451a9f0" satisfied condition "success or failure"
+Feb 12 10:45:37.120: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod projected-volume-b00a278d-698c-4a8a-bbb9-08328451a9f0 container projected-all-volume-test: 
+STEP: delete the pod
+Feb 12 10:45:37.248: INFO: Waiting for pod projected-volume-b00a278d-698c-4a8a-bbb9-08328451a9f0 to disappear
+Feb 12 10:45:37.254: INFO: Pod projected-volume-b00a278d-698c-4a8a-bbb9-08328451a9f0 no longer exists
+[AfterEach] [sig-storage] Projected combined
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:45:37.254: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-8188" for this suite.
+•{"msg":"PASSED [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance]","total":280,"completed":209,"skipped":3519,"failed":0}
+SSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl api-versions 
-  should check if v1 is in available api versions  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Pods 
+  should support remote command execution over websockets [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:32:06.390: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7245
+Feb 12 10:45:37.271: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-8040
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[It] should check if v1 is in available api versions  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: validating api versions
-Aug 30 17:32:06.551: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 api-versions'
-Aug 30 17:32:06.615: INFO: stderr: ""
-Aug 30 17:32:06.615: INFO: stdout: "admissionregistration.k8s.io/v1\nadmissionregistration.k8s.io/v1beta1\napiextensions.k8s.io/v1\napiextensions.k8s.io/v1beta1\napiregistration.k8s.io/v1\napiregistration.k8s.io/v1beta1\napps/v1\nauthentication.k8s.io/v1\nauthentication.k8s.io/v1beta1\nauthorization.k8s.io/v1\nauthorization.k8s.io/v1beta1\nautoscaling/v1\nautoscaling/v2beta1\nautoscaling/v2beta2\nbatch/v1\nbatch/v1beta1\ncertificates.k8s.io/v1beta1\ncluster.k8s.io/v1alpha1\ncoordination.k8s.io/v1\ncoordination.k8s.io/v1beta1\ncrd.projectcalico.org/v1\ndiscovery.k8s.io/v1beta1\nevents.k8s.io/v1beta1\nextensions/v1beta1\nmetrics.k8s.io/v1beta1\nnetworking.k8s.io/v1\nnetworking.k8s.io/v1beta1\nnode.k8s.io/v1beta1\npolicy/v1beta1\nrbac.authorization.k8s.io/v1\nrbac.authorization.k8s.io/v1beta1\nscheduling.k8s.io/v1\nscheduling.k8s.io/v1beta1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:32:06.615: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-7245" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions  [Conformance]","total":280,"completed":234,"skipped":3737,"failed":0}
-SSSSSSSSSSSSSS
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
+[It] should support remote command execution over websockets [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:45:37.455: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+[AfterEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:45:39.706: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-8040" for this suite.
+•{"msg":"PASSED [k8s.io] Pods should support remote command execution over websockets [NodeConformance] [Conformance]","total":280,"completed":210,"skipped":3523,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] ResourceQuota 
-  should verify ResourceQuota with best effort scope. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] Networking Granular Checks: Pods 
+  should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Networking
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:32:06.637: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename resourcequota
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-1210
+Feb 12 10:45:39.724: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename pod-network-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pod-network-test-4901
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should verify ResourceQuota with best effort scope. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a ResourceQuota with best effort scope
-STEP: Ensuring ResourceQuota status is calculated
-STEP: Creating a ResourceQuota with not best effort scope
-STEP: Ensuring ResourceQuota status is calculated
-STEP: Creating a best-effort pod
-STEP: Ensuring resource quota with best effort scope captures the pod usage
-STEP: Ensuring resource quota with not best effort ignored the pod usage
-STEP: Deleting the pod
-STEP: Ensuring resource quota status released the pod usage
-STEP: Creating a not best-effort pod
-STEP: Ensuring resource quota with not best effort scope captures the pod usage
-STEP: Ensuring resource quota with best effort scope ignored the pod usage
-STEP: Deleting the pod
-STEP: Ensuring resource quota status released the pod usage
-[AfterEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:32:22.948: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-1210" for this suite.
+[It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Performing setup for networking test in namespace pod-network-test-4901
+STEP: creating a selector
+STEP: Creating the service pods in kubernetes
+Feb 12 10:45:39.908: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable
+STEP: Creating test pods
+Feb 12 10:46:06.057: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.25.2.67 8081 | grep -v '^\s*$'] Namespace:pod-network-test-4901 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:46:06.058: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:46:07.568: INFO: Found all expected endpoints: [netserver-0]
+Feb 12 10:46:07.574: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.25.0.42 8081 | grep -v '^\s*$'] Namespace:pod-network-test-4901 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:46:07.574: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:46:09.190: INFO: Found all expected endpoints: [netserver-1]
+Feb 12 10:46:09.196: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 172.25.1.198 8081 | grep -v '^\s*$'] Namespace:pod-network-test-4901 PodName:host-test-container-pod ContainerName:agnhost Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false}
+Feb 12 10:46:09.197: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:46:10.714: INFO: Found all expected endpoints: [netserver-2]
+[AfterEach] [sig-network] Networking
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:46:10.715: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pod-network-test-4901" for this suite.
 
-• [SLOW TEST:16.327 seconds]
-[sig-api-machinery] ResourceQuota
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should verify ResourceQuota with best effort scope. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:31.014 seconds]
+[sig-network] Networking
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:26
+  Granular Checks: Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/networking.go:29
+    should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance]","total":280,"completed":235,"skipped":3751,"failed":0}
-SS
+{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":211,"skipped":3562,"failed":0}
+SSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Watchers 
-  should observe add, update, and delete watch notifications on configmaps [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  patching/updating a mutating webhook should work [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:32:22.964: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename watch
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-1671
+Feb 12 10:46:10.739: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-6013
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should observe add, update, and delete watch notifications on configmaps [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a watch on configmaps with label A
-STEP: creating a watch on configmaps with label B
-STEP: creating a watch on configmaps with label A or B
-STEP: creating a configmap with label A and ensuring the correct watchers observe the notification
-Aug 30 17:32:23.152: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-1671 /api/v1/namespaces/watch-1671/configmaps/e2e-watch-test-configmap-a cc8f8e00-9ef4-4c0b-bc07-e6c2b722ecc9 28801 0 2020-08-30 17:32:23 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Aug 30 17:32:23.152: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-1671 /api/v1/namespaces/watch-1671/configmaps/e2e-watch-test-configmap-a cc8f8e00-9ef4-4c0b-bc07-e6c2b722ecc9 28801 0 2020-08-30 17:32:23 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-STEP: modifying configmap A and ensuring the correct watchers observe the notification
-Aug 30 17:32:33.168: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-1671 /api/v1/namespaces/watch-1671/configmaps/e2e-watch-test-configmap-a cc8f8e00-9ef4-4c0b-bc07-e6c2b722ecc9 28858 0 2020-08-30 17:32:23 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-Aug 30 17:32:33.169: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-1671 /api/v1/namespaces/watch-1671/configmaps/e2e-watch-test-configmap-a cc8f8e00-9ef4-4c0b-bc07-e6c2b722ecc9 28858 0 2020-08-30 17:32:23 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
-STEP: modifying configmap A again and ensuring the correct watchers observe the notification
-Aug 30 17:32:43.182: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-1671 /api/v1/namespaces/watch-1671/configmaps/e2e-watch-test-configmap-a cc8f8e00-9ef4-4c0b-bc07-e6c2b722ecc9 28893 0 2020-08-30 17:32:23 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Aug 30 17:32:43.182: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-1671 /api/v1/namespaces/watch-1671/configmaps/e2e-watch-test-configmap-a cc8f8e00-9ef4-4c0b-bc07-e6c2b722ecc9 28893 0 2020-08-30 17:32:23 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-STEP: deleting configmap A and ensuring the correct watchers observe the notification
-Aug 30 17:32:53.194: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-1671 /api/v1/namespaces/watch-1671/configmaps/e2e-watch-test-configmap-a cc8f8e00-9ef4-4c0b-bc07-e6c2b722ecc9 28928 0 2020-08-30 17:32:23 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-Aug 30 17:32:53.194: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-1671 /api/v1/namespaces/watch-1671/configmaps/e2e-watch-test-configmap-a cc8f8e00-9ef4-4c0b-bc07-e6c2b722ecc9 28928 0 2020-08-30 17:32:23 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
-STEP: creating a configmap with label B and ensuring the correct watchers observe the notification
-Aug 30 17:33:03.207: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b  watch-1671 /api/v1/namespaces/watch-1671/configmaps/e2e-watch-test-configmap-b 79745b80-4982-4d44-bdef-0915385b460e 28963 0 2020-08-30 17:33:03 +0000 UTC   map[watch-this-configmap:multiple-watchers-B] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Aug 30 17:33:03.207: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b  watch-1671 /api/v1/namespaces/watch-1671/configmaps/e2e-watch-test-configmap-b 79745b80-4982-4d44-bdef-0915385b460e 28963 0 2020-08-30 17:33:03 +0000 UTC   map[watch-this-configmap:multiple-watchers-B] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-STEP: deleting configmap B and ensuring the correct watchers observe the notification
-Aug 30 17:33:13.220: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b  watch-1671 /api/v1/namespaces/watch-1671/configmaps/e2e-watch-test-configmap-b 79745b80-4982-4d44-bdef-0915385b460e 28996 0 2020-08-30 17:33:03 +0000 UTC   map[watch-this-configmap:multiple-watchers-B] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-Aug 30 17:33:13.220: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b  watch-1671 /api/v1/namespaces/watch-1671/configmaps/e2e-watch-test-configmap-b 79745b80-4982-4d44-bdef-0915385b460e 28996 0 2020-08-30 17:33:03 +0000 UTC   map[watch-this-configmap:multiple-watchers-B] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
-[AfterEach] [sig-api-machinery] Watchers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:33:23.221: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "watch-1671" for this suite.
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Feb 12 10:46:11.118: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Feb 12 10:46:14.157: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] patching/updating a mutating webhook should work [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a mutating webhook configuration
+STEP: Updating a mutating webhook configuration's rules to not include the create operation
+STEP: Creating a configMap that should not be mutated
+STEP: Patching a mutating webhook configuration's rules to include the create operation
+STEP: Creating a configMap that should be mutated
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:46:14.581: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-6013" for this suite.
+STEP: Destroying namespace "webhook-6013-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+•{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance]","total":280,"completed":212,"skipped":3578,"failed":0}
 
-• [SLOW TEST:60.273 seconds]
-[sig-api-machinery] Watchers
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should observe add, update, and delete watch notifications on configmaps [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance]","total":280,"completed":236,"skipped":3753,"failed":0}
-SSSSSSSSSSS
 ------------------------------
-[k8s.io] Container Runtime blackbox test when starting a container that exits 
-  should run with the expected status [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should include webhook resources in discovery documents [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:33:23.238: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-runtime
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-119
+Feb 12 10:46:14.693: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-6711
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should run with the expected status [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount'
-STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase'
-STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition
-STEP: Container 'terminate-cmd-rpa': should get the expected 'State'
-STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance]
-STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount'
-STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase'
-STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition
-STEP: Container 'terminate-cmd-rpof': should get the expected 'State'
-STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance]
-STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount'
-STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase'
-STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition
-STEP: Container 'terminate-cmd-rpn': should get the expected 'State'
-STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance]
-[AfterEach] [k8s.io] Container Runtime
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:33:47.739: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-runtime-119" for this suite.
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Feb 12 10:46:15.404: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:46:17.423: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723574, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723574, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723574, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723574, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Feb 12 10:46:20.452: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should include webhook resources in discovery documents [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: fetching the /apis discovery document
+STEP: finding the admissionregistration.k8s.io API group in the /apis discovery document
+STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis discovery document
+STEP: fetching the /apis/admissionregistration.k8s.io discovery document
+STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis/admissionregistration.k8s.io discovery document
+STEP: fetching the /apis/admissionregistration.k8s.io/v1 discovery document
+STEP: finding mutatingwebhookconfigurations and validatingwebhookconfigurations resources in the /apis/admissionregistration.k8s.io/v1 discovery document
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:46:20.468: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-6711" for this suite.
+STEP: Destroying namespace "webhook-6711-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:24.517 seconds]
-[k8s.io] Container Runtime
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  blackbox test
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:38
-    when starting a container that exits
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/runtime.go:39
-      should run with the expected status [NodeConformance] [Conformance]
-      /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:5.881 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should include webhook resources in discovery documents [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance]","total":280,"completed":213,"skipped":3578,"failed":0}
+SSSSSSSSSSSSSSSSSSS
 ------------------------------
-{"msg":"PASSED [k8s.io] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance]","total":280,"completed":237,"skipped":3764,"failed":0}
+[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] 
+  custom resource defaulting for requests and from storage works  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:46:20.575: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename custom-resource-definition
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in custom-resource-definition-9837
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] custom resource defaulting for requests and from storage works  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:46:20.748: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:46:21.508: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "custom-resource-definition-9837" for this suite.
+•{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works  [Conformance]","total":280,"completed":214,"skipped":3597,"failed":0}
 SSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Variable Expansion 
-  should allow composing env vars into new env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Variable Expansion
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Projected downwardAPI 
+  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:33:47.756: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename var-expansion
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-8096
+Feb 12 10:46:21.562: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-3411
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should allow composing env vars into new env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test env composition
-Aug 30 17:33:47.934: INFO: Waiting up to 5m0s for pod "var-expansion-2acf63f0-f207-4ef8-a083-32c0e43353fe" in namespace "var-expansion-8096" to be "success or failure"
-Aug 30 17:33:47.939: INFO: Pod "var-expansion-2acf63f0-f207-4ef8-a083-32c0e43353fe": Phase="Pending", Reason="", readiness=false. Elapsed: 4.91574ms
-Aug 30 17:33:49.945: INFO: Pod "var-expansion-2acf63f0-f207-4ef8-a083-32c0e43353fe": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.009985604s
+[BeforeEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Feb 12 10:46:21.868: INFO: Waiting up to 5m0s for pod "downwardapi-volume-3dfc155c-22de-46b5-9207-560551146f0e" in namespace "projected-3411" to be "success or failure"
+Feb 12 10:46:21.874: INFO: Pod "downwardapi-volume-3dfc155c-22de-46b5-9207-560551146f0e": Phase="Pending", Reason="", readiness=false. Elapsed: 6.56828ms
+Feb 12 10:46:23.885: INFO: Pod "downwardapi-volume-3dfc155c-22de-46b5-9207-560551146f0e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.016702296s
 STEP: Saw pod success
-Aug 30 17:33:49.945: INFO: Pod "var-expansion-2acf63f0-f207-4ef8-a083-32c0e43353fe" satisfied condition "success or failure"
-Aug 30 17:33:49.950: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod var-expansion-2acf63f0-f207-4ef8-a083-32c0e43353fe container dapi-container: 
+Feb 12 10:46:23.885: INFO: Pod "downwardapi-volume-3dfc155c-22de-46b5-9207-560551146f0e" satisfied condition "success or failure"
+Feb 12 10:46:23.899: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-3dfc155c-22de-46b5-9207-560551146f0e container client-container: 
 STEP: delete the pod
-Aug 30 17:33:50.021: INFO: Waiting for pod var-expansion-2acf63f0-f207-4ef8-a083-32c0e43353fe to disappear
-Aug 30 17:33:50.025: INFO: Pod var-expansion-2acf63f0-f207-4ef8-a083-32c0e43353fe no longer exists
-[AfterEach] [k8s.io] Variable Expansion
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:33:50.026: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "var-expansion-8096" for this suite.
-•{"msg":"PASSED [k8s.io] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance]","total":280,"completed":238,"skipped":3789,"failed":0}
+Feb 12 10:46:23.971: INFO: Waiting for pod downwardapi-volume-3dfc155c-22de-46b5-9207-560551146f0e to disappear
+Feb 12 10:46:23.976: INFO: Pod downwardapi-volume-3dfc155c-22de-46b5-9207-560551146f0e no longer exists
+[AfterEach] [sig-storage] Projected downwardAPI
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:46:23.976: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-3411" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":215,"skipped":3622,"failed":0}
 
 ------------------------------
-[k8s.io] Probing container 
-  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] 
+  works for multiple CRDs of same group but different versions [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:33:50.041: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-9340
+Feb 12 10:46:23.997: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename crd-publish-openapi
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-publish-openapi-7448
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:34:10.234: INFO: Container started at 2020-08-30 17:33:51 +0000 UTC, pod became ready at 2020-08-30 17:34:08 +0000 UTC
-[AfterEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:34:10.234: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-9340" for this suite.
+[It] works for multiple CRDs of same group but different versions [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: CRs in the same group but different versions (one multiversion CRD) show up in OpenAPI documentation
+Feb 12 10:46:24.187: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: CRs in the same group but different versions (two CRDs) show up in OpenAPI documentation
+Feb 12 10:46:36.158: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+Feb 12 10:46:39.062: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:46:50.615: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "crd-publish-openapi-7448" for this suite.
 
-• [SLOW TEST:20.210 seconds]
-[k8s.io] Probing container
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:26.637 seconds]
+[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  works for multiple CRDs of same group but different versions [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]","total":280,"completed":239,"skipped":3789,"failed":0}
-S
+{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance]","total":280,"completed":216,"skipped":3622,"failed":0}
+SSSSSSSSSSSSS
+------------------------------
+[sig-network] Services 
+  should serve multiport endpoints from pods  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:46:50.638: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename services
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-5390
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
+[It] should serve multiport endpoints from pods  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating service multi-endpoint-test in namespace services-5390
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-5390 to expose endpoints map[]
+Feb 12 10:46:50.821: INFO: successfully validated that service multi-endpoint-test in namespace services-5390 exposes endpoints map[] (7.826828ms elapsed)
+STEP: Creating pod pod1 in namespace services-5390
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-5390 to expose endpoints map[pod1:[100]]
+Feb 12 10:46:53.885: INFO: successfully validated that service multi-endpoint-test in namespace services-5390 exposes endpoints map[pod1:[100]] (3.051286717s elapsed)
+STEP: Creating pod pod2 in namespace services-5390
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-5390 to expose endpoints map[pod1:[100] pod2:[101]]
+Feb 12 10:46:55.942: INFO: successfully validated that service multi-endpoint-test in namespace services-5390 exposes endpoints map[pod1:[100] pod2:[101]] (2.04928535s elapsed)
+STEP: Deleting pod pod1 in namespace services-5390
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-5390 to expose endpoints map[pod2:[101]]
+Feb 12 10:46:56.979: INFO: successfully validated that service multi-endpoint-test in namespace services-5390 exposes endpoints map[pod2:[101]] (1.025945063s elapsed)
+STEP: Deleting pod pod2 in namespace services-5390
+STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-5390 to expose endpoints map[]
+Feb 12 10:46:58.008: INFO: successfully validated that service multi-endpoint-test in namespace services-5390 exposes endpoints map[] (1.014208591s elapsed)
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:46:58.034: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-5390" for this suite.
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
+
+• [SLOW TEST:7.418 seconds]
+[sig-network] Services
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should serve multiport endpoints from pods  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-network] Services should serve multiport endpoints from pods  [Conformance]","total":280,"completed":217,"skipped":3635,"failed":0}
+SSSSSSSSSS
 ------------------------------
 [sig-storage] Projected downwardAPI 
-  should provide container's memory request [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:34:10.251: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:46:58.056: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-8939
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9279
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
-[It] should provide container's memory request [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/projected_downwardapi.go:40
+[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Creating a pod to test downward API volume plugin
-Aug 30 17:34:10.436: INFO: Waiting up to 5m0s for pod "downwardapi-volume-b964c188-06eb-4761-aa6a-2421f0cac2cf" in namespace "projected-8939" to be "success or failure"
-Aug 30 17:34:10.444: INFO: Pod "downwardapi-volume-b964c188-06eb-4761-aa6a-2421f0cac2cf": Phase="Pending", Reason="", readiness=false. Elapsed: 8.236189ms
-Aug 30 17:34:12.451: INFO: Pod "downwardapi-volume-b964c188-06eb-4761-aa6a-2421f0cac2cf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015133659s
+Feb 12 10:46:58.242: INFO: Waiting up to 5m0s for pod "downwardapi-volume-377a4708-dacf-4969-8c3e-eb9e3b03e19e" in namespace "projected-9279" to be "success or failure"
+Feb 12 10:46:58.253: INFO: Pod "downwardapi-volume-377a4708-dacf-4969-8c3e-eb9e3b03e19e": Phase="Pending", Reason="", readiness=false. Elapsed: 10.088792ms
+Feb 12 10:47:00.259: INFO: Pod "downwardapi-volume-377a4708-dacf-4969-8c3e-eb9e3b03e19e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016477906s
+Feb 12 10:47:02.265: INFO: Pod "downwardapi-volume-377a4708-dacf-4969-8c3e-eb9e3b03e19e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022365308s
 STEP: Saw pod success
-Aug 30 17:34:12.451: INFO: Pod "downwardapi-volume-b964c188-06eb-4761-aa6a-2421f0cac2cf" satisfied condition "success or failure"
-Aug 30 17:34:12.456: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-b964c188-06eb-4761-aa6a-2421f0cac2cf container client-container: 
+Feb 12 10:47:02.265: INFO: Pod "downwardapi-volume-377a4708-dacf-4969-8c3e-eb9e3b03e19e" satisfied condition "success or failure"
+Feb 12 10:47:02.270: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-377a4708-dacf-4969-8c3e-eb9e3b03e19e container client-container: 
 STEP: delete the pod
-Aug 30 17:34:12.531: INFO: Waiting for pod downwardapi-volume-b964c188-06eb-4761-aa6a-2421f0cac2cf to disappear
-Aug 30 17:34:12.535: INFO: Pod downwardapi-volume-b964c188-06eb-4761-aa6a-2421f0cac2cf no longer exists
+Feb 12 10:47:02.310: INFO: Waiting for pod downwardapi-volume-377a4708-dacf-4969-8c3e-eb9e3b03e19e to disappear
+Feb 12 10:47:02.316: INFO: Pod downwardapi-volume-377a4708-dacf-4969-8c3e-eb9e3b03e19e no longer exists
 [AfterEach] [sig-storage] Projected downwardAPI
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:34:12.535: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-8939" for this suite.
-•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance]","total":280,"completed":240,"skipped":3790,"failed":0}
-SSSSSSSSSSSSSSSSSS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:47:02.316: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9279" for this suite.
+•{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":218,"skipped":3645,"failed":0}
+SSSSSSSSS
 ------------------------------
-[sig-storage] Downward API volume 
-  should update annotations on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] ResourceQuota 
+  should create a ResourceQuota and capture the life of a secret. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:34:12.549: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-4779
+Feb 12 10:47:02.334: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename resourcequota
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-5530
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should update annotations on modification [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating the pod
-Aug 30 17:34:15.278: INFO: Successfully updated pod "annotationupdatefe1e7538-38e4-45ef-bf83-6a52ab850f60"
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:34:17.308: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-4779" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance]","total":280,"completed":241,"skipped":3808,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[It] should create a ResourceQuota and capture the life of a secret. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Discovering how many secrets are in namespace by default
+STEP: Counting existing ResourceQuota
+STEP: Creating a ResourceQuota
+STEP: Ensuring resource quota status is calculated
+STEP: Creating a Secret
+STEP: Ensuring resource quota status captures secret creation
+STEP: Deleting a secret
+STEP: Ensuring resource quota status released usage
+[AfterEach] [sig-api-machinery] ResourceQuota
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:47:19.596: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "resourcequota-5530" for this suite.
+
+• [SLOW TEST:17.280 seconds]
+[sig-api-machinery] ResourceQuota
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should create a ResourceQuota and capture the life of a secret. [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance]","total":280,"completed":219,"skipped":3654,"failed":0}
+SSS
+------------------------------
+[sig-cli] Kubectl client Proxy server 
+  should support proxy with --port 0  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:47:19.614: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-7399
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[It] should support proxy with --port 0  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: starting the proxy server
+Feb 12 10:47:19.785: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-008079477 proxy -p 0 --disable-filter'
+STEP: curling proxy /api/ output
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:47:19.844: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-7399" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support proxy with --port 0  [Conformance]","total":280,"completed":220,"skipped":3657,"failed":0}
+SSSSSSSSSS
 ------------------------------
 [sig-storage] Secrets 
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:34:17.325: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:47:19.863: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-1206
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-655
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name s-test-opt-del-8b9e6561-2bca-43e7-956b-e8b1072527a8
-STEP: Creating secret with name s-test-opt-upd-e763d665-4ec6-411c-8eeb-2f60e173e133
-STEP: Creating the pod
-STEP: Deleting secret s-test-opt-del-8b9e6561-2bca-43e7-956b-e8b1072527a8
-STEP: Updating secret s-test-opt-upd-e763d665-4ec6-411c-8eeb-2f60e173e133
-STEP: Creating secret with name s-test-opt-create-a2014a50-6be1-47b8-8844-0a266b75a54e
-STEP: waiting to observe update in volume
+[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating secret with name secret-test-aa8162e0-a9d9-4b52-b8fe-8033d330559b
+STEP: Creating a pod to test consume secrets
+Feb 12 10:47:20.043: INFO: Waiting up to 5m0s for pod "pod-secrets-a5819aee-2c0c-464d-bade-94e8b915a754" in namespace "secrets-655" to be "success or failure"
+Feb 12 10:47:20.048: INFO: Pod "pod-secrets-a5819aee-2c0c-464d-bade-94e8b915a754": Phase="Pending", Reason="", readiness=false. Elapsed: 4.952036ms
+Feb 12 10:47:22.055: INFO: Pod "pod-secrets-a5819aee-2c0c-464d-bade-94e8b915a754": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012073331s
+STEP: Saw pod success
+Feb 12 10:47:22.055: INFO: Pod "pod-secrets-a5819aee-2c0c-464d-bade-94e8b915a754" satisfied condition "success or failure"
+Feb 12 10:47:22.060: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-secrets-a5819aee-2c0c-464d-bade-94e8b915a754 container secret-volume-test: 
+STEP: delete the pod
+Feb 12 10:47:22.144: INFO: Waiting for pod pod-secrets-a5819aee-2c0c-464d-bade-94e8b915a754 to disappear
+Feb 12 10:47:22.149: INFO: Pod pod-secrets-a5819aee-2c0c-464d-bade-94e8b915a754 no longer exists
 [AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:35:34.676: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-1206" for this suite.
-
-• [SLOW TEST:77.367 seconds]
-[sig-storage] Secrets
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:34
-  optional updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:47:22.149: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-655" for this suite.
+•{"msg":"PASSED [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":280,"completed":221,"skipped":3667,"failed":0}
+SSSSS
 ------------------------------
-{"msg":"PASSED [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":242,"skipped":3875,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[k8s.io] Container Runtime blackbox test on terminated container 
+  should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Container Runtime
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:47:22.172: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-runtime
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-runtime-8622
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: create the container
+STEP: wait for the container to reach Succeeded
+STEP: get the container status
+STEP: the container should be terminated
+STEP: the termination message should be set
+Feb 12 10:47:24.405: INFO: Expected: &{OK} to match Container's Termination Message: OK --
+STEP: delete the container
+[AfterEach] [k8s.io] Container Runtime
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:47:24.428: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-runtime-8622" for this suite.
+•{"msg":"PASSED [k8s.io] Container Runtime blackbox test on terminated container should report termination message [LinuxOnly] from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","total":280,"completed":222,"skipped":3672,"failed":0}
+SS
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-cli] Kubectl client Proxy server 
+  should support --unix-socket=/path  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:35:34.693: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-5835
+Feb 12 10:47:24.446: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-8031
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 17:35:34.881: INFO: Waiting up to 5m0s for pod "downwardapi-volume-de4f28f4-8193-47ea-b5d8-b7c5fc5edcbd" in namespace "downward-api-5835" to be "success or failure"
-Aug 30 17:35:34.895: INFO: Pod "downwardapi-volume-de4f28f4-8193-47ea-b5d8-b7c5fc5edcbd": Phase="Pending", Reason="", readiness=false. Elapsed: 14.022138ms
-Aug 30 17:35:36.901: INFO: Pod "downwardapi-volume-de4f28f4-8193-47ea-b5d8-b7c5fc5edcbd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.020276556s
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[It] should support --unix-socket=/path  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Starting the proxy
+Feb 12 10:47:24.609: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-008079477 proxy --unix-socket=/tmp/kubectl-proxy-unix096297104/test'
+STEP: retrieving proxy /api/ output
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:47:24.670: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-8031" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support --unix-socket=/path  [Conformance]","total":280,"completed":223,"skipped":3674,"failed":0}
+SSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:47:24.688: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-5122
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name configmap-test-volume-map-1b767f15-7141-4004-a0db-aa8e3b58bb0c
+STEP: Creating a pod to test consume configMaps
+Feb 12 10:47:24.885: INFO: Waiting up to 5m0s for pod "pod-configmaps-0db69dec-9db6-4584-9e6b-1e4e29ace37f" in namespace "configmap-5122" to be "success or failure"
+Feb 12 10:47:24.893: INFO: Pod "pod-configmaps-0db69dec-9db6-4584-9e6b-1e4e29ace37f": Phase="Pending", Reason="", readiness=false. Elapsed: 7.883568ms
+Feb 12 10:47:26.901: INFO: Pod "pod-configmaps-0db69dec-9db6-4584-9e6b-1e4e29ace37f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.015327477s
 STEP: Saw pod success
-Aug 30 17:35:36.901: INFO: Pod "downwardapi-volume-de4f28f4-8193-47ea-b5d8-b7c5fc5edcbd" satisfied condition "success or failure"
-Aug 30 17:35:36.905: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-de4f28f4-8193-47ea-b5d8-b7c5fc5edcbd container client-container: 
+Feb 12 10:47:26.901: INFO: Pod "pod-configmaps-0db69dec-9db6-4584-9e6b-1e4e29ace37f" satisfied condition "success or failure"
+Feb 12 10:47:26.906: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-configmaps-0db69dec-9db6-4584-9e6b-1e4e29ace37f container configmap-volume-test: 
 STEP: delete the pod
-Aug 30 17:35:36.974: INFO: Waiting for pod downwardapi-volume-de4f28f4-8193-47ea-b5d8-b7c5fc5edcbd to disappear
-Aug 30 17:35:36.979: INFO: Pod downwardapi-volume-de4f28f4-8193-47ea-b5d8-b7c5fc5edcbd no longer exists
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:35:36.979: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-5835" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":243,"skipped":3905,"failed":0}
-SSSSSSSSSSSSS
+Feb 12 10:47:26.975: INFO: Waiting for pod pod-configmaps-0db69dec-9db6-4584-9e6b-1e4e29ace37f to disappear
+Feb 12 10:47:26.980: INFO: Pod pod-configmaps-0db69dec-9db6-4584-9e6b-1e4e29ace37f no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:47:26.980: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-5122" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":224,"skipped":3694,"failed":0}
+SSSSSSSSSSSS
 ------------------------------
-[sig-apps] Job 
-  should adopt matching orphans and release non-matching pods [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Job
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] InitContainer [NodeConformance] 
+  should invoke init containers on a RestartNever pod [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:35:37.001: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename job
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in job-3698
+Feb 12 10:47:27.002: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename init-container
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-495
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should adopt matching orphans and release non-matching pods [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a job
-STEP: Ensuring active pods == parallelism
-STEP: Orphaning one of the Job's Pods
-Aug 30 17:35:41.695: INFO: Successfully updated pod "adopt-release-kf9tr"
-STEP: Checking that the Job readopts the Pod
-Aug 30 17:35:41.695: INFO: Waiting up to 15m0s for pod "adopt-release-kf9tr" in namespace "job-3698" to be "adopted"
-Aug 30 17:35:41.699: INFO: Pod "adopt-release-kf9tr": Phase="Running", Reason="", readiness=true. Elapsed: 4.269078ms
-Aug 30 17:35:43.705: INFO: Pod "adopt-release-kf9tr": Phase="Running", Reason="", readiness=true. Elapsed: 2.009736213s
-Aug 30 17:35:43.705: INFO: Pod "adopt-release-kf9tr" satisfied condition "adopted"
-STEP: Removing the labels from the Job's Pod
-Aug 30 17:35:44.219: INFO: Successfully updated pod "adopt-release-kf9tr"
-STEP: Checking that the Job releases the Pod
-Aug 30 17:35:44.219: INFO: Waiting up to 15m0s for pod "adopt-release-kf9tr" in namespace "job-3698" to be "released"
-Aug 30 17:35:44.223: INFO: Pod "adopt-release-kf9tr": Phase="Running", Reason="", readiness=true. Elapsed: 4.526787ms
-Aug 30 17:35:46.229: INFO: Pod "adopt-release-kf9tr": Phase="Running", Reason="", readiness=true. Elapsed: 2.010599196s
-Aug 30 17:35:46.230: INFO: Pod "adopt-release-kf9tr" satisfied condition "released"
-[AfterEach] [sig-apps] Job
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:35:46.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "job-3698" for this suite.
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
+[It] should invoke init containers on a RestartNever pod [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating the pod
+Feb 12 10:47:27.167: INFO: PodSpec: initContainers in spec.initContainers
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:47:31.528: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "init-container-495" for this suite.
+•{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance]","total":280,"completed":225,"skipped":3706,"failed":0}
+S
+------------------------------
+[k8s.io] [sig-node] PreStop 
+  should call prestop when killing a pod  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] [sig-node] PreStop
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:47:31.548: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename prestop
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in prestop-9806
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] [sig-node] PreStop
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/node/pre_stop.go:172
+[It] should call prestop when killing a pod  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating server pod server in namespace prestop-9806
+STEP: Waiting for pods to come up.
+STEP: Creating tester pod tester in namespace prestop-9806
+STEP: Deleting pre-stop pod
+Feb 12 10:47:40.882: INFO: Saw: {
+	"Hostname": "server",
+	"Sent": null,
+	"Received": {
+		"prestop": 1
+	},
+	"Errors": null,
+	"Log": [
+		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.",
+		"default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up."
+	],
+	"StillContactingPeers": true
+}
+STEP: Deleting the server pod
+[AfterEach] [k8s.io] [sig-node] PreStop
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:47:40.894: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "prestop-9806" for this suite.
 
-• [SLOW TEST:9.243 seconds]
-[sig-apps] Job
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should adopt matching orphans and release non-matching pods [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:9.371 seconds]
+[k8s.io] [sig-node] PreStop
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should call prestop when killing a pod  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance]","total":280,"completed":244,"skipped":3918,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [k8s.io] [sig-node] PreStop should call prestop when killing a pod  [Conformance]","total":280,"completed":226,"skipped":3707,"failed":0}
+SSSSSS
 ------------------------------
-[sig-api-machinery] Garbage collector 
-  should not be blocked by dependency circle [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] Services 
+  should be able to create a functioning NodePort service [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:35:46.245: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-4640
+Feb 12 10:47:40.920: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename services
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-6353
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not be blocked by dependency circle [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:35:46.447: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"0fd81173-035d-4f31-8d95-ece8eaeead2d", Controller:(*bool)(0xc004d14176), BlockOwnerDeletion:(*bool)(0xc004d14177)}}
-Aug 30 17:35:46.456: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"895a778d-af2c-414f-9eb0-100ed980d8a4", Controller:(*bool)(0xc004e57bc6), BlockOwnerDeletion:(*bool)(0xc004e57bc7)}}
-Aug 30 17:35:46.464: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"05718575-3730-4738-8b7c-8ebc032a58f9", Controller:(*bool)(0xc004d14336), BlockOwnerDeletion:(*bool)(0xc004d14337)}}
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:35:51.490: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-4640" for this suite.
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
+[It] should be able to create a functioning NodePort service [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating service nodeport-test with type=NodePort in namespace services-6353
+STEP: creating replication controller nodeport-test in namespace services-6353
+I0212 10:47:41.121804      21 runners.go:189] Created replication controller with name: nodeport-test, namespace: services-6353, replica count: 2
+I0212 10:47:44.172145      21 runners.go:189] nodeport-test Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+Feb 12 10:47:44.172: INFO: Creating new exec pod
+Feb 12 10:47:47.201: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=services-6353 execpodnrggb -- /bin/sh -x -c nc -zv -t -w 2 nodeport-test 80'
+Feb 12 10:47:47.840: INFO: stderr: "+ nc -zv -t -w 2 nodeport-test 80\nConnection to nodeport-test 80 port [tcp/http] succeeded!\n"
+Feb 12 10:47:47.840: INFO: stdout: ""
+Feb 12 10:47:47.840: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=services-6353 execpodnrggb -- /bin/sh -x -c nc -zv -t -w 2 10.240.17.45 80'
+Feb 12 10:47:48.434: INFO: stderr: "+ nc -zv -t -w 2 10.240.17.45 80\nConnection to 10.240.17.45 80 port [tcp/http] succeeded!\n"
+Feb 12 10:47:48.434: INFO: stdout: ""
+Feb 12 10:47:48.434: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=services-6353 execpodnrggb -- /bin/sh -x -c nc -zv -t -w 2 104.248.134.247 30497'
+Feb 12 10:47:49.032: INFO: stderr: "+ nc -zv -t -w 2 104.248.134.247 30497\nConnection to 104.248.134.247 30497 port [tcp/30497] succeeded!\n"
+Feb 12 10:47:49.032: INFO: stdout: ""
+Feb 12 10:47:49.032: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=services-6353 execpodnrggb -- /bin/sh -x -c nc -zv -t -w 2 139.59.138.127 30497'
+Feb 12 10:47:49.631: INFO: stderr: "+ nc -zv -t -w 2 139.59.138.127 30497\nConnection to 139.59.138.127 30497 port [tcp/30497] succeeded!\n"
+Feb 12 10:47:49.631: INFO: stdout: ""
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:47:49.631: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-6353" for this suite.
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
+
+• [SLOW TEST:8.729 seconds]
+[sig-network] Services
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should be able to create a functioning NodePort service [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-network] Services should be able to create a functioning NodePort service [Conformance]","total":280,"completed":227,"skipped":3713,"failed":0}
+SSSSSSSSSS
+------------------------------
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should deny crd creation [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:47:49.650: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-1378
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Feb 12 10:47:50.434: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:47:52.449: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723669, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723669, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723669, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723669, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Feb 12 10:47:55.476: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should deny crd creation [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Registering the crd webhook via the AdmissionRegistration API
+STEP: Creating a custom resource definition that should be denied by the webhook
+Feb 12 10:47:55.668: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:47:55.907: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-1378" for this suite.
+STEP: Destroying namespace "webhook-1378-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:5.267 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should not be blocked by dependency circle [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:6.348 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should deny crd creation [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]","total":280,"completed":245,"skipped":3957,"failed":0}
-SSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance]","total":280,"completed":228,"skipped":3723,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[k8s.io] Pods 
-  should contain environment variables for services [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Downward API volume 
+  should provide container's cpu request [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:35:51.512: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename pods
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-1918
+Feb 12 10:47:55.998: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-5167
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
-[It] should contain environment variables for services [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:35:53.740: INFO: Waiting up to 5m0s for pod "client-envvars-240c4bf3-28a0-44a8-be0d-77992470667e" in namespace "pods-1918" to be "success or failure"
-Aug 30 17:35:53.748: INFO: Pod "client-envvars-240c4bf3-28a0-44a8-be0d-77992470667e": Phase="Pending", Reason="", readiness=false. Elapsed: 7.411245ms
-Aug 30 17:35:55.754: INFO: Pod "client-envvars-240c4bf3-28a0-44a8-be0d-77992470667e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.013395056s
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should provide container's cpu request [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Feb 12 10:47:56.168: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9b6b8fab-b20d-4997-ba6a-5ae3d9c16ccd" in namespace "downward-api-5167" to be "success or failure"
+Feb 12 10:47:56.179: INFO: Pod "downwardapi-volume-9b6b8fab-b20d-4997-ba6a-5ae3d9c16ccd": Phase="Pending", Reason="", readiness=false. Elapsed: 10.517407ms
+Feb 12 10:47:58.184: INFO: Pod "downwardapi-volume-9b6b8fab-b20d-4997-ba6a-5ae3d9c16ccd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015841715s
+Feb 12 10:48:00.191: INFO: Pod "downwardapi-volume-9b6b8fab-b20d-4997-ba6a-5ae3d9c16ccd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.02292651s
 STEP: Saw pod success
-Aug 30 17:35:55.754: INFO: Pod "client-envvars-240c4bf3-28a0-44a8-be0d-77992470667e" satisfied condition "success or failure"
-Aug 30 17:35:55.759: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod client-envvars-240c4bf3-28a0-44a8-be0d-77992470667e container env3cont: 
+Feb 12 10:48:00.192: INFO: Pod "downwardapi-volume-9b6b8fab-b20d-4997-ba6a-5ae3d9c16ccd" satisfied condition "success or failure"
+Feb 12 10:48:00.197: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-9b6b8fab-b20d-4997-ba6a-5ae3d9c16ccd container client-container: 
 STEP: delete the pod
-Aug 30 17:35:55.797: INFO: Waiting for pod client-envvars-240c4bf3-28a0-44a8-be0d-77992470667e to disappear
-Aug 30 17:35:55.802: INFO: Pod client-envvars-240c4bf3-28a0-44a8-be0d-77992470667e no longer exists
-[AfterEach] [k8s.io] Pods
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:35:55.802: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "pods-1918" for this suite.
-•{"msg":"PASSED [k8s.io] Pods should contain environment variables for services [NodeConformance] [Conformance]","total":280,"completed":246,"skipped":3966,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Feb 12 10:48:00.238: INFO: Waiting for pod downwardapi-volume-9b6b8fab-b20d-4997-ba6a-5ae3d9c16ccd to disappear
+Feb 12 10:48:00.243: INFO: Pod downwardapi-volume-9b6b8fab-b20d-4997-ba6a-5ae3d9c16ccd no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:48:00.243: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-5167" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance]","total":280,"completed":229,"skipped":3754,"failed":0}
+SSSSSSSSS
 ------------------------------
 [sig-storage] Secrets 
   should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:35:55.822: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:48:00.260: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-8349
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-8016
 STEP: Waiting for a default service account to be provisioned in namespace
 [It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secret-namespace-3588
-STEP: Creating secret with name secret-test-fe7ff3ac-d3d7-4105-9ac6-4e5b0212473a
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secret-namespace-7718
+STEP: Creating secret with name secret-test-b43324fc-660c-48cb-b9d2-5656e7adb6ef
 STEP: Creating a pod to test consume secrets
-Aug 30 17:35:56.180: INFO: Waiting up to 5m0s for pod "pod-secrets-8df46728-804f-45f7-b65e-5a6ba7e16195" in namespace "secrets-8349" to be "success or failure"
-Aug 30 17:35:56.187: INFO: Pod "pod-secrets-8df46728-804f-45f7-b65e-5a6ba7e16195": Phase="Pending", Reason="", readiness=false. Elapsed: 7.792178ms
-Aug 30 17:35:58.193: INFO: Pod "pod-secrets-8df46728-804f-45f7-b65e-5a6ba7e16195": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.013364769s
+Feb 12 10:48:00.595: INFO: Waiting up to 5m0s for pod "pod-secrets-1a1b8ec7-48e8-4805-ad48-f82ca7c5f050" in namespace "secrets-8016" to be "success or failure"
+Feb 12 10:48:00.601: INFO: Pod "pod-secrets-1a1b8ec7-48e8-4805-ad48-f82ca7c5f050": Phase="Pending", Reason="", readiness=false. Elapsed: 6.376906ms
+Feb 12 10:48:02.607: INFO: Pod "pod-secrets-1a1b8ec7-48e8-4805-ad48-f82ca7c5f050": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012390705s
+Feb 12 10:48:04.614: INFO: Pod "pod-secrets-1a1b8ec7-48e8-4805-ad48-f82ca7c5f050": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018663825s
 STEP: Saw pod success
-Aug 30 17:35:58.193: INFO: Pod "pod-secrets-8df46728-804f-45f7-b65e-5a6ba7e16195" satisfied condition "success or failure"
-Aug 30 17:35:58.198: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-secrets-8df46728-804f-45f7-b65e-5a6ba7e16195 container secret-volume-test: 
+Feb 12 10:48:04.614: INFO: Pod "pod-secrets-1a1b8ec7-48e8-4805-ad48-f82ca7c5f050" satisfied condition "success or failure"
+Feb 12 10:48:04.620: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-secrets-1a1b8ec7-48e8-4805-ad48-f82ca7c5f050 container secret-volume-test: 
 STEP: delete the pod
-Aug 30 17:35:58.277: INFO: Waiting for pod pod-secrets-8df46728-804f-45f7-b65e-5a6ba7e16195 to disappear
-Aug 30 17:35:58.281: INFO: Pod pod-secrets-8df46728-804f-45f7-b65e-5a6ba7e16195 no longer exists
+Feb 12 10:48:04.692: INFO: Waiting for pod pod-secrets-1a1b8ec7-48e8-4805-ad48-f82ca7c5f050 to disappear
+Feb 12 10:48:04.697: INFO: Pod pod-secrets-1a1b8ec7-48e8-4805-ad48-f82ca7c5f050 no longer exists
 [AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:35:58.281: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-8349" for this suite.
-STEP: Destroying namespace "secret-namespace-3588" for this suite.
-•{"msg":"PASSED [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]","total":280,"completed":247,"skipped":3995,"failed":0}
-SSSSSSSSSSS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:48:04.697: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-8016" for this suite.
+STEP: Destroying namespace "secret-namespace-7718" for this suite.
+•{"msg":"PASSED [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]","total":280,"completed":230,"skipped":3763,"failed":0}
+SSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl run job 
-  should create a job from an image when restart is OnFailure  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  should have a working scale subresource [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:35:58.307: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2218
+Feb 12 10:48:04.723: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-2694
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[BeforeEach] Kubectl run job
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1681
-[It] should create a job from an image when restart is OnFailure  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: running the image docker.io/library/httpd:2.4.38-alpine
-Aug 30 17:35:58.482: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 run e2e-test-httpd-job --restart=OnFailure --generator=job/v1 --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-2218'
-Aug 30 17:35:58.702: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-Aug 30 17:35:58.702: INFO: stdout: "job.batch/e2e-test-httpd-job created\n"
-STEP: verifying the job e2e-test-httpd-job was created
-[AfterEach] Kubectl run job
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1686
-Aug 30 17:35:58.710: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete jobs e2e-test-httpd-job --namespace=kubectl-2218'
-Aug 30 17:35:58.791: INFO: stderr: ""
-Aug 30 17:35:58.791: INFO: stdout: "job.batch \"e2e-test-httpd-job\" deleted\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:35:58.791: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-2218" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run job should create a job from an image when restart is OnFailure  [Conformance]","total":280,"completed":248,"skipped":4006,"failed":0}
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
+STEP: Creating service test in namespace statefulset-2694
+[It] should have a working scale subresource [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating statefulset ss in namespace statefulset-2694
+Feb 12 10:48:04.920: INFO: Found 0 stateful pods, waiting for 1
+Feb 12 10:48:14.930: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
+STEP: getting scale subresource
+STEP: updating a scale subresource
+STEP: verifying the statefulset Spec.Replicas was modified
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
+Feb 12 10:48:14.961: INFO: Deleting all statefulset in ns statefulset-2694
+Feb 12 10:48:14.966: INFO: Scaling statefulset ss to 0
+Feb 12 10:48:34.993: INFO: Waiting for statefulset status.replicas updated to 0
+Feb 12 10:48:34.997: INFO: Deleting statefulset ss
+[AfterEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:48:35.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-2694" for this suite.
 
+• [SLOW TEST:30.315 seconds]
+[sig-apps] StatefulSet
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+    should have a working scale subresource [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-api-machinery] Namespaces [Serial] 
-  should ensure that all services are removed when a namespace is deleted [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Namespaces [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance]","total":280,"completed":231,"skipped":3782,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] EmptyDir volumes 
+  should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:35:58.810: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename namespaces
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in namespaces-634
+Feb 12 10:48:35.040: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-2827
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should ensure that all services are removed when a namespace is deleted [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a test namespace
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-7928
+[It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir 0777 on tmpfs
+Feb 12 10:48:35.208: INFO: Waiting up to 5m0s for pod "pod-efd62020-109c-4e52-b5bc-41cca5d3e6cb" in namespace "emptydir-2827" to be "success or failure"
+Feb 12 10:48:35.219: INFO: Pod "pod-efd62020-109c-4e52-b5bc-41cca5d3e6cb": Phase="Pending", Reason="", readiness=false. Elapsed: 10.473693ms
+Feb 12 10:48:37.267: INFO: Pod "pod-efd62020-109c-4e52-b5bc-41cca5d3e6cb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.058439234s
+STEP: Saw pod success
+Feb 12 10:48:37.267: INFO: Pod "pod-efd62020-109c-4e52-b5bc-41cca5d3e6cb" satisfied condition "success or failure"
+Feb 12 10:48:37.274: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-efd62020-109c-4e52-b5bc-41cca5d3e6cb container test-container: 
+STEP: delete the pod
+Feb 12 10:48:37.306: INFO: Waiting for pod pod-efd62020-109c-4e52-b5bc-41cca5d3e6cb to disappear
+Feb 12 10:48:37.359: INFO: Pod pod-efd62020-109c-4e52-b5bc-41cca5d3e6cb no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:48:37.359: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-2827" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":232,"skipped":3853,"failed":0}
+SSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Watchers 
+  should receive events on concurrent watches in same order [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Watchers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:48:37.378: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename watch
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-8443
 STEP: Waiting for a default service account to be provisioned in namespace
-STEP: Creating a service in the namespace
-STEP: Deleting the namespace
-STEP: Waiting for the namespace to be removed.
-STEP: Recreating the namespace
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in nsdeletetest-5106
-STEP: Verifying there is no service in the namespace
-[AfterEach] [sig-api-machinery] Namespaces [Serial]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:36:05.330: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "namespaces-634" for this suite.
-STEP: Destroying namespace "nsdeletetest-7928" for this suite.
-Aug 30 17:36:05.351: INFO: Namespace nsdeletetest-7928 was already deleted
-STEP: Destroying namespace "nsdeletetest-5106" for this suite.
-
-• [SLOW TEST:6.551 seconds]
-[sig-api-machinery] Namespaces [Serial]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should ensure that all services are removed when a namespace is deleted [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[It] should receive events on concurrent watches in same order [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: starting a background goroutine to produce watch events
+STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order
+[AfterEach] [sig-api-machinery] Watchers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:48:42.581: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-8443" for this suite.
+
+• [SLOW TEST:5.301 seconds]
+[sig-api-machinery] Watchers
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should receive events on concurrent watches in same order [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance]","total":280,"completed":249,"skipped":4006,"failed":0}
-SS
+{"msg":"PASSED [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance]","total":280,"completed":233,"skipped":3864,"failed":0}
+SSSSSSSSSSSSSS
 ------------------------------
-[sig-network] Services 
-  should serve a basic endpoint from pods  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] EmptyDir volumes 
+  should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:36:05.360: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename services
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-6406
+Feb 12 10:48:42.681: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename emptydir
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-6815
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
-[It] should serve a basic endpoint from pods  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating service endpoint-test2 in namespace services-6406
-STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-6406 to expose endpoints map[]
-Aug 30 17:36:05.555: INFO: Get endpoints failed (5.694606ms elapsed, ignoring for 5s): endpoints "endpoint-test2" not found
-Aug 30 17:36:06.570: INFO: successfully validated that service endpoint-test2 in namespace services-6406 exposes endpoints map[] (1.021149707s elapsed)
-STEP: Creating pod pod1 in namespace services-6406
-STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-6406 to expose endpoints map[pod1:[80]]
-Aug 30 17:36:08.623: INFO: successfully validated that service endpoint-test2 in namespace services-6406 exposes endpoints map[pod1:[80]] (2.033486003s elapsed)
-STEP: Creating pod pod2 in namespace services-6406
-STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-6406 to expose endpoints map[pod1:[80] pod2:[80]]
-Aug 30 17:36:10.683: INFO: successfully validated that service endpoint-test2 in namespace services-6406 exposes endpoints map[pod1:[80] pod2:[80]] (2.051096297s elapsed)
-STEP: Deleting pod pod1 in namespace services-6406
-STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-6406 to expose endpoints map[pod2:[80]]
-Aug 30 17:36:11.725: INFO: successfully validated that service endpoint-test2 in namespace services-6406 exposes endpoints map[pod2:[80]] (1.027520825s elapsed)
-STEP: Deleting pod pod2 in namespace services-6406
-STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-6406 to expose endpoints map[]
-Aug 30 17:36:12.747: INFO: successfully validated that service endpoint-test2 in namespace services-6406 exposes endpoints map[] (1.009907265s elapsed)
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:36:12.781: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "services-6406" for this suite.
-[AfterEach] [sig-network] Services
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
+[It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir 0644 on tmpfs
+Feb 12 10:48:42.859: INFO: Waiting up to 5m0s for pod "pod-06b966fe-16ff-434d-ba83-ecf1260a2a15" in namespace "emptydir-6815" to be "success or failure"
+Feb 12 10:48:42.865: INFO: Pod "pod-06b966fe-16ff-434d-ba83-ecf1260a2a15": Phase="Pending", Reason="", readiness=false. Elapsed: 5.850951ms
+Feb 12 10:48:44.870: INFO: Pod "pod-06b966fe-16ff-434d-ba83-ecf1260a2a15": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011512098s
+STEP: Saw pod success
+Feb 12 10:48:44.871: INFO: Pod "pod-06b966fe-16ff-434d-ba83-ecf1260a2a15" satisfied condition "success or failure"
+Feb 12 10:48:44.876: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-06b966fe-16ff-434d-ba83-ecf1260a2a15 container test-container: 
+STEP: delete the pod
+Feb 12 10:48:44.920: INFO: Waiting for pod pod-06b966fe-16ff-434d-ba83-ecf1260a2a15 to disappear
+Feb 12 10:48:44.926: INFO: Pod pod-06b966fe-16ff-434d-ba83-ecf1260a2a15 no longer exists
+[AfterEach] [sig-storage] EmptyDir volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:48:44.926: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-6815" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":234,"skipped":3878,"failed":0}
+SSSSSSSSS
+------------------------------
+[k8s.io] InitContainer [NodeConformance] 
+  should not start app containers if init containers fail on a RestartAlways pod [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:48:44.950: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename init-container
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in init-container-6955
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/init_container.go:153
+[It] should not start app containers if init containers fail on a RestartAlways pod [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating the pod
+Feb 12 10:48:45.105: INFO: PodSpec: initContainers in spec.initContainers
+Feb 12 10:49:28.529: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-b1ed3f5f-7cd3-48ed-af30-9dee3c3ee9b0", GenerateName:"", Namespace:"init-container-6955", SelfLink:"/api/v1/namespaces/init-container-6955/pods/pod-init-b1ed3f5f-7cd3-48ed-af30-9dee3c3ee9b0", UID:"35b3bd45-8b64-424c-906d-35140b04547f", ResourceVersion:"32974", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63748723724, loc:(*time.Location)(0x791c680)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"105076846"}, Annotations:map[string]string{"cni.projectcalico.org/podIP":"172.25.1.217/32"}, OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-mxg2d", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc0070ea000), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-mxg2d", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"docker.io/library/busybox:1.29", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-mxg2d", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-mxg2d", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc003d40078), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"nostalgic-fermat-cdc5d8777-xg74t", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc004ffa000), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc003d400f0)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc003d40110)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc003d40118), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc003d4011c), PreemptionPolicy:(*v1.PreemptionPolicy)(nil), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723724, loc:(*time.Location)(0x791c680)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723724, loc:(*time.Location)(0x791c680)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723724, loc:(*time.Location)(0x791c680)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723724, loc:(*time.Location)(0x791c680)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"138.68.97.197", PodIP:"172.25.1.217", PodIPs:[]v1.PodIP{v1.PodIP{IP:"172.25.1.217"}}, StartTime:(*v1.Time)(0xc003ac6140), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc001fde070)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc001fde0e0)}, Ready:false, RestartCount:3, Image:"busybox:1.29", ImageID:"docker-pullable://busybox@sha256:8ccbac733d19c0dd4d70b4f0c1e12245b5fa3ad24758a11035ee505c629c0796", ContainerID:"docker://555bde11cca2f7974850de54d9d13e876174458699e889a9e2e56099ac76babd", Started:(*bool)(nil)}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc003ac66e0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"docker.io/library/busybox:1.29", ImageID:"", ContainerID:"", Started:(*bool)(nil)}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc003ac6320), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause:3.1", ImageID:"", ContainerID:"", Started:(*bool)(0xc003d4019f)}}, QOSClass:"Burstable", EphemeralContainerStatuses:[]v1.ContainerStatus(nil)}}
+[AfterEach] [k8s.io] InitContainer [NodeConformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:49:28.529: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "init-container-6955" for this suite.
 
-• [SLOW TEST:7.436 seconds]
-[sig-network] Services
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should serve a basic endpoint from pods  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:43.595 seconds]
+[k8s.io] InitContainer [NodeConformance]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should not start app containers if init containers fail on a RestartAlways pod [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Services should serve a basic endpoint from pods  [Conformance]","total":280,"completed":250,"skipped":4008,"failed":0}
-SSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [k8s.io] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance]","total":280,"completed":235,"skipped":3887,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-node] Downward API 
-  should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-node] Downward API
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Downward API volume 
+  should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:36:12.797: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:49:28.545: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-5045
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7801
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward api env vars
-Aug 30 17:36:12.971: INFO: Waiting up to 5m0s for pod "downward-api-91c9c019-1063-4428-be9c-ec96d8acb532" in namespace "downward-api-5045" to be "success or failure"
-Aug 30 17:36:12.979: INFO: Pod "downward-api-91c9c019-1063-4428-be9c-ec96d8acb532": Phase="Pending", Reason="", readiness=false. Elapsed: 8.186921ms
-Aug 30 17:36:14.985: INFO: Pod "downward-api-91c9c019-1063-4428-be9c-ec96d8acb532": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.014521683s
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Feb 12 10:49:28.713: INFO: Waiting up to 5m0s for pod "downwardapi-volume-22995dd7-bbbc-42c3-b87c-8b337bcfa17b" in namespace "downward-api-7801" to be "success or failure"
+Feb 12 10:49:28.719: INFO: Pod "downwardapi-volume-22995dd7-bbbc-42c3-b87c-8b337bcfa17b": Phase="Pending", Reason="", readiness=false. Elapsed: 5.965957ms
+Feb 12 10:49:30.725: INFO: Pod "downwardapi-volume-22995dd7-bbbc-42c3-b87c-8b337bcfa17b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.01208507s
 STEP: Saw pod success
-Aug 30 17:36:14.985: INFO: Pod "downward-api-91c9c019-1063-4428-be9c-ec96d8acb532" satisfied condition "success or failure"
-Aug 30 17:36:14.990: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downward-api-91c9c019-1063-4428-be9c-ec96d8acb532 container dapi-container: 
+Feb 12 10:49:30.725: INFO: Pod "downwardapi-volume-22995dd7-bbbc-42c3-b87c-8b337bcfa17b" satisfied condition "success or failure"
+Feb 12 10:49:30.730: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-22995dd7-bbbc-42c3-b87c-8b337bcfa17b container client-container: 
 STEP: delete the pod
-Aug 30 17:36:15.060: INFO: Waiting for pod downward-api-91c9c019-1063-4428-be9c-ec96d8acb532 to disappear
-Aug 30 17:36:15.065: INFO: Pod downward-api-91c9c019-1063-4428-be9c-ec96d8acb532 no longer exists
-[AfterEach] [sig-node] Downward API
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:36:15.065: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-5045" for this suite.
-•{"msg":"PASSED [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]","total":280,"completed":251,"skipped":4028,"failed":0}
-S
+Feb 12 10:49:30.776: INFO: Waiting for pod downwardapi-volume-22995dd7-bbbc-42c3-b87c-8b337bcfa17b to disappear
+Feb 12 10:49:30.782: INFO: Pod downwardapi-volume-22995dd7-bbbc-42c3-b87c-8b337bcfa17b no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:49:30.782: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-7801" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":236,"skipped":3915,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
-  should mutate custom resource [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] Services 
+  should provide secure master service  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:36:15.082: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename webhook
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-3466
+Feb 12 10:49:30.803: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename services
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-9891
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
-STEP: Setting up server cert
-STEP: Create role binding to let webhook read extension-apiserver-authentication
-STEP: Deploying the webhook pod
-STEP: Wait for the deployment to be ready
-Aug 30 17:36:15.591: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
-Aug 30 17:36:17.609: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405775, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405775, loc:(*time.Location)(0x7925260)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405775, loc:(*time.Location)(0x7925260)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63734405775, loc:(*time.Location)(0x7925260)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
-STEP: Deploying the webhook service
-STEP: Verifying the service has paired with the endpoint
-Aug 30 17:36:20.626: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
-[It] should mutate custom resource [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:36:20.632: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Registering the mutating webhook for custom resource e2e-test-webhook-872-crds.webhook.example.com via the AdmissionRegistration API
-STEP: Creating a custom resource that should be mutated by the webhook
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:36:21.978: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "webhook-3466" for this suite.
-STEP: Destroying namespace "webhook-3466-markers" for this suite.
-[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
-
-• [SLOW TEST:6.989 seconds]
-[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should mutate custom resource [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance]","total":280,"completed":252,"skipped":4029,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
+[It] should provide secure master service  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:49:30.977: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-9891" for this suite.
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
+•{"msg":"PASSED [sig-network] Services should provide secure master service  [Conformance]","total":280,"completed":237,"skipped":3945,"failed":0}
+SSSS
 ------------------------------
-[sig-cli] Kubectl client Update Demo 
-  should scale a replication controller  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] Proxy version v1 
+  should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] version v1
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:36:22.072: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2573
+Feb 12 10:49:30.994: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename proxy
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in proxy-2010
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[BeforeEach] Update Demo
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:324
-[It] should scale a replication controller  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating a replication controller
-Aug 30 17:36:22.255: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-2573'
-Aug 30 17:36:22.603: INFO: stderr: ""
-Aug 30 17:36:22.603: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Aug 30 17:36:22.603: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-2573'
-Aug 30 17:36:22.671: INFO: stderr: ""
-Aug 30 17:36:22.671: INFO: stdout: "update-demo-nautilus-qx7q8 update-demo-nautilus-z5zjr "
-Aug 30 17:36:22.671: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-qx7q8 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2573'
-Aug 30 17:36:22.738: INFO: stderr: ""
-Aug 30 17:36:22.738: INFO: stdout: ""
-Aug 30 17:36:22.738: INFO: update-demo-nautilus-qx7q8 is created but not running
-Aug 30 17:36:27.738: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-2573'
-Aug 30 17:36:27.803: INFO: stderr: ""
-Aug 30 17:36:27.803: INFO: stdout: "update-demo-nautilus-qx7q8 update-demo-nautilus-z5zjr "
-Aug 30 17:36:27.803: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-qx7q8 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2573'
-Aug 30 17:36:27.864: INFO: stderr: ""
-Aug 30 17:36:27.864: INFO: stdout: "true"
-Aug 30 17:36:27.864: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-qx7q8 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-2573'
-Aug 30 17:36:27.937: INFO: stderr: ""
-Aug 30 17:36:27.937: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Aug 30 17:36:27.937: INFO: validating pod update-demo-nautilus-qx7q8
-Aug 30 17:36:28.072: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Aug 30 17:36:28.072: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Aug 30 17:36:28.072: INFO: update-demo-nautilus-qx7q8 is verified up and running
-Aug 30 17:36:28.072: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-z5zjr -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2573'
-Aug 30 17:36:28.145: INFO: stderr: ""
-Aug 30 17:36:28.145: INFO: stdout: "true"
-Aug 30 17:36:28.145: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-z5zjr -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-2573'
-Aug 30 17:36:28.211: INFO: stderr: ""
-Aug 30 17:36:28.211: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Aug 30 17:36:28.211: INFO: validating pod update-demo-nautilus-z5zjr
-Aug 30 17:36:28.344: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Aug 30 17:36:28.344: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Aug 30 17:36:28.344: INFO: update-demo-nautilus-z5zjr is verified up and running
-STEP: scaling down the replication controller
-Aug 30 17:36:28.346: INFO: scanned /root for discovery docs: 
-Aug 30 17:36:28.346: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 scale rc update-demo-nautilus --replicas=1 --timeout=5m --namespace=kubectl-2573'
-Aug 30 17:36:29.440: INFO: stderr: ""
-Aug 30 17:36:29.440: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Aug 30 17:36:29.440: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-2573'
-Aug 30 17:36:29.511: INFO: stderr: ""
-Aug 30 17:36:29.511: INFO: stdout: "update-demo-nautilus-qx7q8 update-demo-nautilus-z5zjr "
-STEP: Replicas for name=update-demo: expected=1 actual=2
-Aug 30 17:36:34.512: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-2573'
-Aug 30 17:36:34.587: INFO: stderr: ""
-Aug 30 17:36:34.587: INFO: stdout: "update-demo-nautilus-qx7q8 "
-Aug 30 17:36:34.588: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-qx7q8 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2573'
-Aug 30 17:36:34.654: INFO: stderr: ""
-Aug 30 17:36:34.654: INFO: stdout: "true"
-Aug 30 17:36:34.654: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-qx7q8 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-2573'
-Aug 30 17:36:34.719: INFO: stderr: ""
-Aug 30 17:36:34.719: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Aug 30 17:36:34.719: INFO: validating pod update-demo-nautilus-qx7q8
-Aug 30 17:36:34.745: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Aug 30 17:36:34.745: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Aug 30 17:36:34.745: INFO: update-demo-nautilus-qx7q8 is verified up and running
-STEP: scaling up the replication controller
-Aug 30 17:36:34.747: INFO: scanned /root for discovery docs: 
-Aug 30 17:36:34.747: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 scale rc update-demo-nautilus --replicas=2 --timeout=5m --namespace=kubectl-2573'
-Aug 30 17:36:35.837: INFO: stderr: ""
-Aug 30 17:36:35.837: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n"
-STEP: waiting for all containers in name=update-demo pods to come up.
-Aug 30 17:36:35.837: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-2573'
-Aug 30 17:36:35.941: INFO: stderr: ""
-Aug 30 17:36:35.941: INFO: stdout: "update-demo-nautilus-kgczl update-demo-nautilus-qx7q8 "
-Aug 30 17:36:35.941: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-kgczl -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2573'
-Aug 30 17:36:36.038: INFO: stderr: ""
-Aug 30 17:36:36.038: INFO: stdout: ""
-Aug 30 17:36:36.038: INFO: update-demo-nautilus-kgczl is created but not running
-Aug 30 17:36:41.038: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=kubectl-2573'
-Aug 30 17:36:41.104: INFO: stderr: ""
-Aug 30 17:36:41.104: INFO: stdout: "update-demo-nautilus-kgczl update-demo-nautilus-qx7q8 "
-Aug 30 17:36:41.104: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-kgczl -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2573'
-Aug 30 17:36:41.165: INFO: stderr: ""
-Aug 30 17:36:41.165: INFO: stdout: "true"
-Aug 30 17:36:41.165: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-kgczl -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-2573'
-Aug 30 17:36:41.235: INFO: stderr: ""
-Aug 30 17:36:41.235: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Aug 30 17:36:41.235: INFO: validating pod update-demo-nautilus-kgczl
-Aug 30 17:36:41.368: INFO: got data: {
-  "image": "nautilus.jpg"
-}
-
-Aug 30 17:36:41.368: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Aug 30 17:36:41.368: INFO: update-demo-nautilus-kgczl is verified up and running
-Aug 30 17:36:41.368: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-qx7q8 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2573'
-Aug 30 17:36:41.436: INFO: stderr: ""
-Aug 30 17:36:41.436: INFO: stdout: "true"
-Aug 30 17:36:41.436: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods update-demo-nautilus-qx7q8 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-2573'
-Aug 30 17:36:41.499: INFO: stderr: ""
-Aug 30 17:36:41.499: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus:1.0"
-Aug 30 17:36:41.499: INFO: validating pod update-demo-nautilus-qx7q8
-Aug 30 17:36:41.508: INFO: got data: {
-  "image": "nautilus.jpg"
-}
+[It] should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:49:31.187: INFO: (0) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 12.153176ms)
+Feb 12 10:49:31.231: INFO: (1) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 44.005877ms)
+Feb 12 10:49:31.241: INFO: (2) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 9.584667ms)
+Feb 12 10:49:31.250: INFO: (3) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 8.757745ms)
+Feb 12 10:49:31.260: INFO: (4) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 10.17732ms)
+Feb 12 10:49:31.271: INFO: (5) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 10.344537ms)
+Feb 12 10:49:31.281: INFO: (6) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 10.355251ms)
+Feb 12 10:49:31.291: INFO: (7) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 10.323213ms)
+Feb 12 10:49:31.301: INFO: (8) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 9.701244ms)
+Feb 12 10:49:31.311: INFO: (9) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 9.128583ms)
+Feb 12 10:49:31.319: INFO: (10) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 8.548691ms)
+Feb 12 10:49:31.334: INFO: (11) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 15.072252ms)
+Feb 12 10:49:31.343: INFO: (12) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 9.011036ms)
+Feb 12 10:49:31.353: INFO: (13) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 9.742471ms)
+Feb 12 10:49:31.367: INFO: (14) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 13.242093ms)
+Feb 12 10:49:31.376: INFO: (15) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 9.286533ms)
+Feb 12 10:49:31.387: INFO: (16) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 10.962561ms)
+Feb 12 10:49:31.396: INFO: (17) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 8.867162ms)
+Feb 12 10:49:31.405: INFO: (18) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 9.308647ms)
+Feb 12 10:49:31.415: INFO: (19) /api/v1/nodes/nostalgic-fermat-cdc5d8777-xg74t:10250/proxy/logs/: 
+alternatives.log
+apt/
+... (200; 9.061491ms)
+[AfterEach] version v1
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:49:31.415: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "proxy-2010" for this suite.
+•{"msg":"PASSED [sig-network] Proxy version v1 should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]","total":280,"completed":238,"skipped":3949,"failed":0}
 
-Aug 30 17:36:41.508: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
-Aug 30 17:36:41.508: INFO: update-demo-nautilus-qx7q8 is verified up and running
-STEP: using delete to clean up resources
-Aug 30 17:36:41.508: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete --grace-period=0 --force -f - --namespace=kubectl-2573'
-Aug 30 17:36:41.597: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Aug 30 17:36:41.597: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n"
-Aug 30 17:36:41.597: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-2573'
-Aug 30 17:36:41.666: INFO: stderr: "No resources found in kubectl-2573 namespace.\n"
-Aug 30 17:36:41.666: INFO: stdout: ""
-Aug 30 17:36:41.666: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -l name=update-demo --namespace=kubectl-2573 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
-Aug 30 17:36:41.728: INFO: stderr: ""
-Aug 30 17:36:41.728: INFO: stdout: "update-demo-nautilus-kgczl\nupdate-demo-nautilus-qx7q8\n"
-Aug 30 17:36:42.228: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get rc,svc -l name=update-demo --no-headers --namespace=kubectl-2573'
-Aug 30 17:36:42.333: INFO: stderr: "No resources found in kubectl-2573 namespace.\n"
-Aug 30 17:36:42.333: INFO: stdout: ""
-Aug 30 17:36:42.333: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -l name=update-demo --namespace=kubectl-2573 -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
-Aug 30 17:36:42.425: INFO: stderr: ""
-Aug 30 17:36:42.425: INFO: stdout: ""
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:36:42.425: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-2573" for this suite.
+------------------------------
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:49:31.435: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename sched-pred
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-249
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
+Feb 12 10:49:31.592: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Feb 12 10:49:31.610: INFO: Waiting for terminating namespaces to be deleted...
+Feb 12 10:49:31.615: INFO: 
+Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-dbhjj before test
+Feb 12 10:49:31.737: INFO: logrotate-4f9dq from kube-system started at 2021-02-12 09:48:36 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.737: INFO: 	Container logrotate ready: true, restart count 0
+Feb 12 10:49:31.737: INFO: node-local-dns-k2khb from kube-system started at 2021-02-12 09:48:36 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.737: INFO: 	Container node-cache ready: true, restart count 0
+Feb 12 10:49:31.737: INFO: kube-proxy-c88c9 from kube-system started at 2021-02-12 09:48:16 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.737: INFO: 	Container kube-proxy ready: true, restart count 0
+Feb 12 10:49:31.738: INFO: canal-k48gh from kube-system started at 2021-02-12 09:48:16 +0000 UTC (2 container statuses recorded)
+Feb 12 10:49:31.738: INFO: 	Container calico-node ready: true, restart count 0
+Feb 12 10:49:31.738: INFO: 	Container kube-flannel ready: true, restart count 0
+Feb 12 10:49:31.738: INFO: sonobuoy-e2e-job-690c418daa0f4403 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded)
+Feb 12 10:49:31.738: INFO: 	Container e2e ready: true, restart count 0
+Feb 12 10:49:31.738: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Feb 12 10:49:31.738: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-vkql5 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded)
+Feb 12 10:49:31.738: INFO: 	Container sonobuoy-worker ready: true, restart count 2
+Feb 12 10:49:31.738: INFO: 	Container systemd-logs ready: true, restart count 0
+Feb 12 10:49:31.738: INFO: openvpn-client-78d595f58b-2pzkd from kube-system started at 2021-02-12 10:38:49 +0000 UTC (2 container statuses recorded)
+Feb 12 10:49:31.738: INFO: 	Container dnat-controller ready: true, restart count 0
+Feb 12 10:49:31.738: INFO: 	Container openvpn-client ready: true, restart count 0
+Feb 12 10:49:31.739: INFO: coredns-6449d494cc-vrdxh from kube-system started at 2021-02-12 10:38:49 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.739: INFO: 	Container coredns ready: true, restart count 0
+Feb 12 10:49:31.739: INFO: user-ssh-keys-agent-8lrpd from kube-system started at 2021-02-12 09:48:16 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.739: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
+Feb 12 10:49:31.739: INFO: 
+Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-r6snm before test
+Feb 12 10:49:31.912: INFO: dashboard-metrics-scraper-59bfc65dc9-rhjf7 from kubernetes-dashboard started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.912: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
+Feb 12 10:49:31.912: INFO: dashboard-metrics-scraper-59bfc65dc9-d8l4b from kubernetes-dashboard started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.912: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
+Feb 12 10:49:31.912: INFO: node-local-dns-fkrhg from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.912: INFO: 	Container node-cache ready: true, restart count 0
+Feb 12 10:49:31.912: INFO: logrotate-7vwrm from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.912: INFO: 	Container logrotate ready: true, restart count 0
+Feb 12 10:49:31.912: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-zp8g7 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded)
+Feb 12 10:49:31.912: INFO: 	Container sonobuoy-worker ready: false, restart count 2
+Feb 12 10:49:31.912: INFO: 	Container systemd-logs ready: true, restart count 0
+Feb 12 10:49:31.912: INFO: kube-proxy-gfdtg from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.912: INFO: 	Container kube-proxy ready: true, restart count 0
+Feb 12 10:49:31.912: INFO: user-ssh-keys-agent-vxhth from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.912: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
+Feb 12 10:49:31.912: INFO: coredns-6449d494cc-psw4s from kube-system started at 2021-02-12 09:48:21 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.912: INFO: 	Container coredns ready: true, restart count 0
+Feb 12 10:49:31.912: INFO: canal-b9q6f from kube-system started at 2021-02-12 09:48:02 +0000 UTC (2 container statuses recorded)
+Feb 12 10:49:31.912: INFO: 	Container calico-node ready: true, restart count 0
+Feb 12 10:49:31.912: INFO: 	Container kube-flannel ready: true, restart count 0
+Feb 12 10:49:31.912: INFO: sonobuoy from sonobuoy started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:31.912: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Feb 12 10:49:31.912: INFO: 
+Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-xg74t before test
+Feb 12 10:49:32.005: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-wbbpd from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded)
+Feb 12 10:49:32.005: INFO: 	Container sonobuoy-worker ready: false, restart count 2
+Feb 12 10:49:32.005: INFO: 	Container systemd-logs ready: true, restart count 0
+Feb 12 10:49:32.005: INFO: logrotate-9f95x from kube-system started at 2021-02-12 10:39:16 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:32.005: INFO: 	Container logrotate ready: true, restart count 0
+Feb 12 10:49:32.005: INFO: pod-init-b1ed3f5f-7cd3-48ed-af30-9dee3c3ee9b0 from init-container-6955 started at 2021-02-12 10:48:44 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:32.005: INFO: 	Container run1 ready: false, restart count 0
+Feb 12 10:49:32.005: INFO: kube-proxy-4zfp6 from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:32.005: INFO: 	Container kube-proxy ready: true, restart count 0
+Feb 12 10:49:32.005: INFO: canal-ww2cz from kube-system started at 2021-02-12 09:48:02 +0000 UTC (2 container statuses recorded)
+Feb 12 10:49:32.005: INFO: 	Container calico-node ready: true, restart count 0
+Feb 12 10:49:32.005: INFO: 	Container kube-flannel ready: true, restart count 0
+Feb 12 10:49:32.006: INFO: user-ssh-keys-agent-kc6zq from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:32.006: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
+Feb 12 10:49:32.006: INFO: node-local-dns-75w47 from kube-system started at 2021-02-12 10:39:16 +0000 UTC (1 container statuses recorded)
+Feb 12 10:49:32.006: INFO: 	Container node-cache ready: true, restart count 0
+[It] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Trying to launch a pod without a label to get a node which can launch it.
+STEP: Explicitly delete pod here to free the resource it takes.
+STEP: Trying to apply a random label on the found node.
+STEP: verifying the node has the label kubernetes.io/e2e-6e6ed808-0c42-46c7-a38b-22d9279b4276 90
+STEP: Trying to create a pod(pod1) with hostport 54321 and hostIP 127.0.0.1 and expect scheduled
+STEP: Trying to create another pod(pod2) with hostport 54321 but hostIP 127.0.0.2 on the node which pod1 resides and expect scheduled
+STEP: Trying to create a third pod(pod3) with hostport 54321, hostIP 127.0.0.2 but use UDP protocol on the node which pod2 resides
+STEP: removing the label kubernetes.io/e2e-6e6ed808-0c42-46c7-a38b-22d9279b4276 off the node nostalgic-fermat-cdc5d8777-xg74t
+STEP: verifying the node doesn't have the label kubernetes.io/e2e-6e6ed808-0c42-46c7-a38b-22d9279b4276
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:49:46.169: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-249" for this suite.
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
 
-• [SLOW TEST:20.373 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Update Demo
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:322
-    should scale a replication controller  [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:14.751 seconds]
+[sig-scheduling] SchedulerPredicates [Serial]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
+  validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Update Demo should scale a replication controller  [Conformance]","total":280,"completed":253,"skipped":4055,"failed":0}
-S
+{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [Conformance]","total":280,"completed":239,"skipped":3949,"failed":0}
+SSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Guestbook application 
-  should create and stop a working application  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  Should recreate evicted statefulset [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:36:42.446: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-8966
+Feb 12 10:49:46.185: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-5571
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[It] should create and stop a working application  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: creating all guestbook components
-Aug 30 17:36:42.608: INFO: apiVersion: v1
-kind: Service
-metadata:
-  name: agnhost-slave
-  labels:
-    app: agnhost
-    role: slave
-    tier: backend
-spec:
-  ports:
-  - port: 6379
-  selector:
-    app: agnhost
-    role: slave
-    tier: backend
-
-Aug 30 17:36:42.608: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-8966'
-Aug 30 17:36:42.834: INFO: stderr: ""
-Aug 30 17:36:42.834: INFO: stdout: "service/agnhost-slave created\n"
-Aug 30 17:36:42.835: INFO: apiVersion: v1
-kind: Service
-metadata:
-  name: agnhost-master
-  labels:
-    app: agnhost
-    role: master
-    tier: backend
-spec:
-  ports:
-  - port: 6379
-    targetPort: 6379
-  selector:
-    app: agnhost
-    role: master
-    tier: backend
-
-Aug 30 17:36:42.835: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-8966'
-Aug 30 17:36:43.006: INFO: stderr: ""
-Aug 30 17:36:43.006: INFO: stdout: "service/agnhost-master created\n"
-Aug 30 17:36:43.006: INFO: apiVersion: v1
-kind: Service
-metadata:
-  name: frontend
-  labels:
-    app: guestbook
-    tier: frontend
-spec:
-  # if your cluster supports it, uncomment the following to automatically create
-  # an external load-balanced IP for the frontend service.
-  # type: LoadBalancer
-  ports:
-  - port: 80
-  selector:
-    app: guestbook
-    tier: frontend
-
-Aug 30 17:36:43.006: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-8966'
-Aug 30 17:36:43.243: INFO: stderr: ""
-Aug 30 17:36:43.243: INFO: stdout: "service/frontend created\n"
-Aug 30 17:36:43.243: INFO: apiVersion: apps/v1
-kind: Deployment
-metadata:
-  name: frontend
-spec:
-  replicas: 3
-  selector:
-    matchLabels:
-      app: guestbook
-      tier: frontend
-  template:
-    metadata:
-      labels:
-        app: guestbook
-        tier: frontend
-    spec:
-      containers:
-      - name: guestbook-frontend
-        image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8
-        args: [ "guestbook", "--backend-port", "6379" ]
-        resources:
-          requests:
-            cpu: 100m
-            memory: 100Mi
-        ports:
-        - containerPort: 80
-
-Aug 30 17:36:43.243: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-8966'
-Aug 30 17:36:43.432: INFO: stderr: ""
-Aug 30 17:36:43.433: INFO: stdout: "deployment.apps/frontend created\n"
-Aug 30 17:36:43.433: INFO: apiVersion: apps/v1
-kind: Deployment
-metadata:
-  name: agnhost-master
-spec:
-  replicas: 1
-  selector:
-    matchLabels:
-      app: agnhost
-      role: master
-      tier: backend
-  template:
-    metadata:
-      labels:
-        app: agnhost
-        role: master
-        tier: backend
-    spec:
-      containers:
-      - name: master
-        image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8
-        args: [ "guestbook", "--http-port", "6379" ]
-        resources:
-          requests:
-            cpu: 100m
-            memory: 100Mi
-        ports:
-        - containerPort: 6379
-
-Aug 30 17:36:43.433: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-8966'
-Aug 30 17:36:43.684: INFO: stderr: ""
-Aug 30 17:36:43.684: INFO: stdout: "deployment.apps/agnhost-master created\n"
-Aug 30 17:36:43.684: INFO: apiVersion: apps/v1
-kind: Deployment
-metadata:
-  name: agnhost-slave
-spec:
-  replicas: 2
-  selector:
-    matchLabels:
-      app: agnhost
-      role: slave
-      tier: backend
-  template:
-    metadata:
-      labels:
-        app: agnhost
-        role: slave
-        tier: backend
-    spec:
-      containers:
-      - name: slave
-        image: gcr.io/kubernetes-e2e-test-images/agnhost:2.8
-        args: [ "guestbook", "--slaveof", "agnhost-master", "--http-port", "6379" ]
-        resources:
-          requests:
-            cpu: 100m
-            memory: 100Mi
-        ports:
-        - containerPort: 6379
-
-Aug 30 17:36:43.684: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 create -f - --namespace=kubectl-8966'
-Aug 30 17:36:43.833: INFO: stderr: ""
-Aug 30 17:36:43.833: INFO: stdout: "deployment.apps/agnhost-slave created\n"
-STEP: validating guestbook app
-Aug 30 17:36:43.833: INFO: Waiting for all frontend pods to be Running.
-Aug 30 17:36:48.884: INFO: Waiting for frontend to serve content.
-Aug 30 17:36:48.993: INFO: Trying to add a new entry to the guestbook.
-Aug 30 17:36:49.125: INFO: Verifying that added entry can be retrieved.
-STEP: using delete to clean up resources
-Aug 30 17:36:49.258: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete --grace-period=0 --force -f - --namespace=kubectl-8966'
-Aug 30 17:36:49.353: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Aug 30 17:36:49.353: INFO: stdout: "service \"agnhost-slave\" force deleted\n"
-STEP: using delete to clean up resources
-Aug 30 17:36:49.354: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete --grace-period=0 --force -f - --namespace=kubectl-8966'
-Aug 30 17:36:49.458: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Aug 30 17:36:49.458: INFO: stdout: "service \"agnhost-master\" force deleted\n"
-STEP: using delete to clean up resources
-Aug 30 17:36:49.458: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete --grace-period=0 --force -f - --namespace=kubectl-8966'
-Aug 30 17:36:49.554: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Aug 30 17:36:49.554: INFO: stdout: "service \"frontend\" force deleted\n"
-STEP: using delete to clean up resources
-Aug 30 17:36:49.555: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete --grace-period=0 --force -f - --namespace=kubectl-8966'
-Aug 30 17:36:49.630: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Aug 30 17:36:49.630: INFO: stdout: "deployment.apps \"frontend\" force deleted\n"
-STEP: using delete to clean up resources
-Aug 30 17:36:49.630: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete --grace-period=0 --force -f - --namespace=kubectl-8966'
-Aug 30 17:36:49.729: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Aug 30 17:36:49.729: INFO: stdout: "deployment.apps \"agnhost-master\" force deleted\n"
-STEP: using delete to clean up resources
-Aug 30 17:36:49.730: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete --grace-period=0 --force -f - --namespace=kubectl-8966'
-Aug 30 17:36:49.834: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
-Aug 30 17:36:49.834: INFO: stdout: "deployment.apps \"agnhost-slave\" force deleted\n"
-[AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:36:49.834: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-8966" for this suite.
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
+STEP: Creating service test in namespace statefulset-5571
+[It] Should recreate evicted statefulset [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Looking for a node to schedule stateful set and pod
+STEP: Creating pod with conflicting port in namespace statefulset-5571
+STEP: Creating statefulset with conflicting port in namespace statefulset-5571
+STEP: Waiting until pod test-pod will start running in namespace statefulset-5571
+STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-5571
+Feb 12 10:49:50.406: INFO: Observed stateful pod in namespace: statefulset-5571, name: ss-0, uid: e6c5e25f-8a6a-45ce-baf5-a1784a6cedc5, status phase: Pending. Waiting for statefulset controller to delete.
+Feb 12 10:49:50.503: INFO: Observed stateful pod in namespace: statefulset-5571, name: ss-0, uid: e6c5e25f-8a6a-45ce-baf5-a1784a6cedc5, status phase: Failed. Waiting for statefulset controller to delete.
+Feb 12 10:49:50.513: INFO: Observed stateful pod in namespace: statefulset-5571, name: ss-0, uid: e6c5e25f-8a6a-45ce-baf5-a1784a6cedc5, status phase: Failed. Waiting for statefulset controller to delete.
+Feb 12 10:49:50.521: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-5571
+STEP: Removing pod with conflicting port in namespace statefulset-5571
+STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-5571 and will be in running state
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
+Feb 12 10:49:56.565: INFO: Deleting all statefulset in ns statefulset-5571
+Feb 12 10:49:56.577: INFO: Scaling statefulset ss to 0
+Feb 12 10:50:06.602: INFO: Waiting for statefulset status.replicas updated to 0
+Feb 12 10:50:06.607: INFO: Deleting statefulset ss
+[AfterEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:50:06.628: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-5571" for this suite.
 
-• [SLOW TEST:7.406 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Guestbook application
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:380
-    should create and stop a working application  [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:20.473 seconds]
+[sig-apps] StatefulSet
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+    Should recreate evicted statefulset [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Guestbook application should create and stop a working application  [Conformance]","total":280,"completed":254,"skipped":4056,"failed":0}
-SSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance]","total":280,"completed":240,"skipped":3966,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] Garbage collector 
-  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] Daemon set [Serial] 
+  should rollback without unnecessary restarts [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:36:49.852: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename gc
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-6535
+Feb 12 10:50:06.664: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename daemonsets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in daemonsets-2363
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: create the rc1
-STEP: create the rc2
-STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well
-STEP: delete the rc simpletest-rc-to-be-deleted
-STEP: wait for the rc to be deleted
-STEP: Gathering metrics
-Aug 30 17:37:00.128: INFO: For apiserver_request_total:
-For apiserver_request_latency_seconds:
-For apiserver_init_events_total:
-For garbage_collector_attempt_to_delete_queue_latency:
-For garbage_collector_attempt_to_delete_work_duration:
-For garbage_collector_attempt_to_orphan_queue_latency:
-For garbage_collector_attempt_to_orphan_work_duration:
-For garbage_collector_dirty_processing_latency_microseconds:
-For garbage_collector_event_processing_latency_microseconds:
-For garbage_collector_graph_changes_queue_latency:
-For garbage_collector_graph_changes_work_duration:
-For garbage_collector_orphan_processing_latency_microseconds:
-For namespace_queue_latency:
-For namespace_queue_latency_sum:
-For namespace_queue_latency_count:
-For namespace_retries:
-For namespace_work_duration:
-For namespace_work_duration_sum:
-For namespace_work_duration_count:
-For function_duration_seconds:
-For errors_total:
-For evicted_pods_total:
+[BeforeEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:133
+[It] should rollback without unnecessary restarts [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:50:07.010: INFO: Create a RollingUpdate DaemonSet
+Feb 12 10:50:07.017: INFO: Check that daemon pods launch on every node of the cluster
+Feb 12 10:50:07.030: INFO: Number of nodes with available pods: 0
+Feb 12 10:50:07.030: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod
+Feb 12 10:50:08.044: INFO: Number of nodes with available pods: 0
+Feb 12 10:50:08.044: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod
+Feb 12 10:50:09.046: INFO: Number of nodes with available pods: 1
+Feb 12 10:50:09.046: INFO: Node nostalgic-fermat-cdc5d8777-dbhjj is running more than one daemon pod
+Feb 12 10:50:10.050: INFO: Number of nodes with available pods: 3
+Feb 12 10:50:10.050: INFO: Number of running nodes: 3, number of available pods: 3
+Feb 12 10:50:10.050: INFO: Update the DaemonSet to trigger a rollout
+Feb 12 10:50:10.063: INFO: Updating DaemonSet daemon-set
+Feb 12 10:50:23.089: INFO: Roll back the DaemonSet before rollout is complete
+Feb 12 10:50:23.101: INFO: Updating DaemonSet daemon-set
+Feb 12 10:50:23.101: INFO: Make sure DaemonSet rollback is complete
+Feb 12 10:50:23.108: INFO: Wrong image for pod: daemon-set-z446f. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent.
+Feb 12 10:50:23.108: INFO: Pod daemon-set-z446f is not available
+Feb 12 10:50:24.123: INFO: Wrong image for pod: daemon-set-z446f. Expected: docker.io/library/httpd:2.4.38-alpine, got: foo:non-existent.
+Feb 12 10:50:24.123: INFO: Pod daemon-set-z446f is not available
+Feb 12 10:50:25.122: INFO: Pod daemon-set-c6jwv is not available
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/daemon_set.go:99
+STEP: Deleting DaemonSet "daemon-set"
+STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-2363, will wait for the garbage collector to delete the pods
+Feb 12 10:50:25.218: INFO: Deleting DaemonSet.extensions daemon-set took: 12.751074ms
+Feb 12 10:50:25.719: INFO: Terminating DaemonSet.extensions daemon-set pods took: 501.052702ms
+Feb 12 10:50:32.325: INFO: Number of nodes with available pods: 0
+Feb 12 10:50:32.325: INFO: Number of running nodes: 0, number of available pods: 0
+Feb 12 10:50:32.332: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"selfLink":"/apis/apps/v1/namespaces/daemonsets-2363/daemonsets","resourceVersion":"33740"},"items":null}
 
-W0830 17:37:00.128645      23 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
-[AfterEach] [sig-api-machinery] Garbage collector
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:37:00.128: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "gc-6535" for this suite.
+Feb 12 10:50:32.337: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"selfLink":"/api/v1/namespaces/daemonsets-2363/pods","resourceVersion":"33740"},"items":null}
 
-• [SLOW TEST:10.291 seconds]
-[sig-api-machinery] Garbage collector
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[AfterEach] [sig-apps] Daemon set [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:50:32.361: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "daemonsets-2363" for this suite.
+
+• [SLOW TEST:25.714 seconds]
+[sig-apps] Daemon set [Serial]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  should rollback without unnecessary restarts [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]","total":280,"completed":255,"skipped":4071,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance]","total":280,"completed":241,"skipped":4001,"failed":0}
+SSSSSSSSSS
 ------------------------------
-[sig-apps] ReplicaSet 
-  should adopt matching pods on creation and release no longer matching pods [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] ReplicaSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Docker Containers 
+  should use the image defaults if command and args are blank [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Docker Containers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:37:00.144: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename replicaset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replicaset-6718
+Feb 12 10:50:32.378: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename containers
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-9671
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should adopt matching pods on creation and release no longer matching pods [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Given a Pod with a 'name' label pod-adoption-release is created
-STEP: When a replicaset with a matching selector is created
-STEP: Then the orphan pod is adopted
-STEP: When the matched label of one of its pods change
-Aug 30 17:37:05.372: INFO: Pod name pod-adoption-release: Found 1 pods out of 1
-STEP: Then the pod is released
-[AfterEach] [sig-apps] ReplicaSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:37:06.393: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replicaset-6718" for this suite.
-
-• [SLOW TEST:6.266 seconds]
-[sig-apps] ReplicaSet
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should adopt matching pods on creation and release no longer matching pods [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance]","total":280,"completed":256,"skipped":4097,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[It] should use the image defaults if command and args are blank [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[AfterEach] [k8s.io] Docker Containers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:50:34.574: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "containers-9671" for this suite.
+•{"msg":"PASSED [k8s.io] Docker Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance]","total":280,"completed":242,"skipped":4011,"failed":0}
+SS
 ------------------------------
-[sig-storage] Projected configMap 
-  updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] ConfigMap 
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:37:06.414: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename projected
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9606
+Feb 12 10:50:34.597: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-3341
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] updates should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating projection with configMap that has name projected-configmap-test-upd-72709657-5fc7-48a9-892f-b29de0edb2fd
+[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name cm-test-opt-del-48ecd61f-b87e-4a13-b032-a85469a21854
+STEP: Creating configMap with name cm-test-opt-upd-f15d81fe-8658-4369-8316-cf699778386a
 STEP: Creating the pod
-STEP: Updating configmap projected-configmap-test-upd-72709657-5fc7-48a9-892f-b29de0edb2fd
+STEP: Deleting configmap cm-test-opt-del-48ecd61f-b87e-4a13-b032-a85469a21854
+STEP: Updating configmap cm-test-opt-upd-f15d81fe-8658-4369-8316-cf699778386a
+STEP: Creating configMap with name cm-test-opt-create-4089aad3-0bf6-44e5-9c11-2eaa129d5dce
 STEP: waiting to observe update in volume
-[AfterEach] [sig-storage] Projected configMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:37:10.782: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "projected-9606" for this suite.
-•{"msg":"PASSED [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":257,"skipped":4136,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:50:39.287: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-3341" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":243,"skipped":4013,"failed":0}
+
 ------------------------------
-[sig-apps] ReplicationController 
-  should adopt matching pods on creation [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should unconditionally reject operations on fail closed webhook [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:37:10.798: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename replication-controller
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-8346
+Feb 12 10:50:39.309: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-2668
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should adopt matching pods on creation [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Given a Pod with a 'name' label pod-adoption is created
-STEP: When a replication controller with a matching selector is created
-STEP: Then the orphan pod is adopted
-[AfterEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:37:16.017: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replication-controller-8346" for this suite.
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Feb 12 10:50:39.886: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:50:41.904: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723839, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723839, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723839, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723839, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Feb 12 10:50:44.928: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should unconditionally reject operations on fail closed webhook [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Registering a webhook that server cannot talk to, with fail closed policy, via the AdmissionRegistration API
+STEP: create a namespace for the webhook
+STEP: create a configmap should be unconditionally rejected by the webhook
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:50:45.211: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-2668" for this suite.
+STEP: Destroying namespace "webhook-2668-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:5.243 seconds]
-[sig-apps] ReplicationController
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should adopt matching pods on creation [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:6.012 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should unconditionally reject operations on fail closed webhook [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] ReplicationController should adopt matching pods on creation [Conformance]","total":280,"completed":258,"skipped":4170,"failed":0}
-S
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]","total":280,"completed":244,"skipped":4013,"failed":0}
+SS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Kubelet when scheduling a read only busybox container 
+  should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:37:16.042: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-1503
+Feb 12 10:50:45.324: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubelet-test
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubelet-test-419
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0777 on node default medium
-Aug 30 17:37:16.222: INFO: Waiting up to 5m0s for pod "pod-c7d76681-90c7-41fc-a8e2-55d5d4726881" in namespace "emptydir-1503" to be "success or failure"
-Aug 30 17:37:16.228: INFO: Pod "pod-c7d76681-90c7-41fc-a8e2-55d5d4726881": Phase="Pending", Reason="", readiness=false. Elapsed: 6.546889ms
-Aug 30 17:37:18.234: INFO: Pod "pod-c7d76681-90c7-41fc-a8e2-55d5d4726881": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.012313754s
-STEP: Saw pod success
-Aug 30 17:37:18.234: INFO: Pod "pod-c7d76681-90c7-41fc-a8e2-55d5d4726881" satisfied condition "success or failure"
-Aug 30 17:37:18.240: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-c7d76681-90c7-41fc-a8e2-55d5d4726881 container test-container: 
-STEP: delete the pod
-Aug 30 17:37:18.330: INFO: Waiting for pod pod-c7d76681-90c7-41fc-a8e2-55d5d4726881 to disappear
-Aug 30 17:37:18.334: INFO: Pod pod-c7d76681-90c7-41fc-a8e2-55d5d4726881 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:37:18.334: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-1503" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":259,"skipped":4171,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] [k8s.io] Kubelet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/kubelet.go:37
+[It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[AfterEach] [k8s.io] Kubelet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:50:47.539: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubelet-test-419" for this suite.
+•{"msg":"PASSED [k8s.io] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":245,"skipped":4015,"failed":0}
+S
 ------------------------------
-[sig-storage] Secrets 
-  should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] Services 
+  should be able to change the type from NodePort to ExternalName [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:37:18.349: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename secrets
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-3225
+Feb 12 10:50:47.561: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename services
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-7064
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating secret with name secret-test-369df46f-3838-4e5f-9b78-a35fb9745d25
-STEP: Creating a pod to test consume secrets
-Aug 30 17:37:18.538: INFO: Waiting up to 5m0s for pod "pod-secrets-015e69ef-bfa4-4c17-8bc3-3b7ada9b01f3" in namespace "secrets-3225" to be "success or failure"
-Aug 30 17:37:18.546: INFO: Pod "pod-secrets-015e69ef-bfa4-4c17-8bc3-3b7ada9b01f3": Phase="Pending", Reason="", readiness=false. Elapsed: 7.602152ms
-Aug 30 17:37:20.552: INFO: Pod "pod-secrets-015e69ef-bfa4-4c17-8bc3-3b7ada9b01f3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.013566666s
-STEP: Saw pod success
-Aug 30 17:37:20.552: INFO: Pod "pod-secrets-015e69ef-bfa4-4c17-8bc3-3b7ada9b01f3" satisfied condition "success or failure"
-Aug 30 17:37:20.557: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-secrets-015e69ef-bfa4-4c17-8bc3-3b7ada9b01f3 container secret-volume-test: 
-STEP: delete the pod
-Aug 30 17:37:20.590: INFO: Waiting for pod pod-secrets-015e69ef-bfa4-4c17-8bc3-3b7ada9b01f3 to disappear
-Aug 30 17:37:20.595: INFO: Pod pod-secrets-015e69ef-bfa4-4c17-8bc3-3b7ada9b01f3 no longer exists
-[AfterEach] [sig-storage] Secrets
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:37:20.595: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "secrets-3225" for this suite.
-•{"msg":"PASSED [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","total":280,"completed":260,"skipped":4200,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
+[It] should be able to change the type from NodePort to ExternalName [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating a service nodeport-service with the type=NodePort in namespace services-7064
+STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service
+STEP: creating service externalsvc in namespace services-7064
+STEP: creating replication controller externalsvc in namespace services-7064
+I0212 10:50:47.797501      21 runners.go:189] Created replication controller with name: externalsvc, namespace: services-7064, replica count: 2
+I0212 10:50:50.847910      21 runners.go:189] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+STEP: changing the NodePort service to type=ExternalName
+Feb 12 10:50:50.889: INFO: Creating new exec pod
+Feb 12 10:50:54.918: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=services-7064 execpod2c5wg -- /bin/sh -x -c nslookup nodeport-service'
+Feb 12 10:50:55.513: INFO: stderr: "+ nslookup nodeport-service\n"
+Feb 12 10:50:55.513: INFO: stdout: "Server:\t\t10.240.16.10\nAddress:\t10.240.16.10#53\n\nnodeport-service.services-7064.svc.cluster.local\tcanonical name = externalsvc.services-7064.svc.cluster.local.\nName:\texternalsvc.services-7064.svc.cluster.local\nAddress: 10.240.17.18\n\n"
+STEP: deleting ReplicationController externalsvc in namespace services-7064, will wait for the garbage collector to delete the pods
+Feb 12 10:50:55.582: INFO: Deleting ReplicationController externalsvc took: 13.882293ms
+Feb 12 10:50:56.083: INFO: Terminating ReplicationController externalsvc pods took: 500.235691ms
+Feb 12 10:51:03.009: INFO: Cleaning up the NodePort to ExternalName test service
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:51:03.032: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-7064" for this suite.
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
+
+• [SLOW TEST:15.491 seconds]
+[sig-network] Services
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should be able to change the type from NodePort to ExternalName [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] Security Context When creating a pod with readOnlyRootFilesystem 
-  should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
-STEP: Creating a kubernetes client
-Aug 30 17:37:20.614: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename security-context-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in security-context-test-5446
-STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39
-[It] should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:37:20.793: INFO: Waiting up to 5m0s for pod "busybox-readonly-false-60482ec4-0546-42b4-91a7-6847253f7ec7" in namespace "security-context-test-5446" to be "success or failure"
-Aug 30 17:37:20.798: INFO: Pod "busybox-readonly-false-60482ec4-0546-42b4-91a7-6847253f7ec7": Phase="Pending", Reason="", readiness=false. Elapsed: 5.161755ms
-Aug 30 17:37:22.803: INFO: Pod "busybox-readonly-false-60482ec4-0546-42b4-91a7-6847253f7ec7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010312737s
-Aug 30 17:37:24.809: INFO: Pod "busybox-readonly-false-60482ec4-0546-42b4-91a7-6847253f7ec7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01588593s
-Aug 30 17:37:24.809: INFO: Pod "busybox-readonly-false-60482ec4-0546-42b4-91a7-6847253f7ec7" satisfied condition "success or failure"
-[AfterEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:37:24.809: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "security-context-test-5446" for this suite.
-•{"msg":"PASSED [k8s.io] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]","total":280,"completed":261,"skipped":4225,"failed":0}
-SSSSSSS
+{"msg":"PASSED [sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance]","total":280,"completed":246,"skipped":4016,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-storage] ConfigMap 
-  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:37:24.825: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:51:03.055: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-5311
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-5447
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name configmap-test-volume-0fef218d-7043-4b5b-b18f-26a2561b97e3
+[It] should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name configmap-test-volume-map-9470d5cf-cc49-4a89-ba90-71d54c7a2dc5
 STEP: Creating a pod to test consume configMaps
-Aug 30 17:37:25.000: INFO: Waiting up to 5m0s for pod "pod-configmaps-7ca31bb2-cc5a-4714-ac93-67e4498e841d" in namespace "configmap-5311" to be "success or failure"
-Aug 30 17:37:25.007: INFO: Pod "pod-configmaps-7ca31bb2-cc5a-4714-ac93-67e4498e841d": Phase="Pending", Reason="", readiness=false. Elapsed: 7.137561ms
-Aug 30 17:37:27.013: INFO: Pod "pod-configmaps-7ca31bb2-cc5a-4714-ac93-67e4498e841d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.013182925s
+Feb 12 10:51:03.231: INFO: Waiting up to 5m0s for pod "pod-configmaps-1104dda8-c295-40be-93ad-49b46f2c458e" in namespace "configmap-5447" to be "success or failure"
+Feb 12 10:51:03.238: INFO: Pod "pod-configmaps-1104dda8-c295-40be-93ad-49b46f2c458e": Phase="Pending", Reason="", readiness=false. Elapsed: 6.178546ms
+Feb 12 10:51:05.243: INFO: Pod "pod-configmaps-1104dda8-c295-40be-93ad-49b46f2c458e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011484857s
+Feb 12 10:51:07.249: INFO: Pod "pod-configmaps-1104dda8-c295-40be-93ad-49b46f2c458e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017514393s
 STEP: Saw pod success
-Aug 30 17:37:27.013: INFO: Pod "pod-configmaps-7ca31bb2-cc5a-4714-ac93-67e4498e841d" satisfied condition "success or failure"
-Aug 30 17:37:27.018: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-configmaps-7ca31bb2-cc5a-4714-ac93-67e4498e841d container configmap-volume-test: 
+Feb 12 10:51:07.249: INFO: Pod "pod-configmaps-1104dda8-c295-40be-93ad-49b46f2c458e" satisfied condition "success or failure"
+Feb 12 10:51:07.254: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-configmaps-1104dda8-c295-40be-93ad-49b46f2c458e container configmap-volume-test: 
 STEP: delete the pod
-Aug 30 17:37:27.094: INFO: Waiting for pod pod-configmaps-7ca31bb2-cc5a-4714-ac93-67e4498e841d to disappear
-Aug 30 17:37:27.100: INFO: Pod pod-configmaps-7ca31bb2-cc5a-4714-ac93-67e4498e841d no longer exists
+Feb 12 10:51:07.288: INFO: Waiting for pod pod-configmaps-1104dda8-c295-40be-93ad-49b46f2c458e to disappear
+Feb 12 10:51:07.293: INFO: Pod pod-configmaps-1104dda8-c295-40be-93ad-49b46f2c458e no longer exists
 [AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:37:27.100: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-5311" for this suite.
-•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":262,"skipped":4232,"failed":0}
-SSS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:51:07.293: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-5447" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":247,"skipped":4045,"failed":0}
+SSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl run default 
-  should create an rc or deployment from an image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[sig-node] Downward API 
+  should provide pod UID as env vars [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-node] Downward API
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:51:07.314: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-5204
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide pod UID as env vars [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward api env vars
+Feb 12 10:51:07.509: INFO: Waiting up to 5m0s for pod "downward-api-8ddc4b9e-7344-4150-b589-6917a484c3c3" in namespace "downward-api-5204" to be "success or failure"
+Feb 12 10:51:07.518: INFO: Pod "downward-api-8ddc4b9e-7344-4150-b589-6917a484c3c3": Phase="Pending", Reason="", readiness=false. Elapsed: 9.372503ms
+Feb 12 10:51:09.524: INFO: Pod "downward-api-8ddc4b9e-7344-4150-b589-6917a484c3c3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015275388s
+Feb 12 10:51:11.532: INFO: Pod "downward-api-8ddc4b9e-7344-4150-b589-6917a484c3c3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023141684s
+STEP: Saw pod success
+Feb 12 10:51:11.532: INFO: Pod "downward-api-8ddc4b9e-7344-4150-b589-6917a484c3c3" satisfied condition "success or failure"
+Feb 12 10:51:11.538: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downward-api-8ddc4b9e-7344-4150-b589-6917a484c3c3 container dapi-container: 
+STEP: delete the pod
+Feb 12 10:51:11.622: INFO: Waiting for pod downward-api-8ddc4b9e-7344-4150-b589-6917a484c3c3 to disappear
+Feb 12 10:51:11.627: INFO: Pod downward-api-8ddc4b9e-7344-4150-b589-6917a484c3c3 no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:51:11.627: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-5204" for this suite.
+•{"msg":"PASSED [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance]","total":280,"completed":248,"skipped":4049,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-cli] Kubectl client Kubectl run job 
+  should create a job from an image when restart is OnFailure [Deprecated] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:37:27.118: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:51:11.646: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-1921
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2223
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[BeforeEach] Kubectl run default
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1489
-[It] should create an rc or deployment from an image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[BeforeEach] Kubectl run job
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1685
+[It] should create a job from an image when restart is OnFailure [Deprecated] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: running the image docker.io/library/httpd:2.4.38-alpine
-Aug 30 17:37:27.291: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 run e2e-test-httpd-deployment --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-1921'
-Aug 30 17:37:27.375: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-Aug 30 17:37:27.375: INFO: stdout: "deployment.apps/e2e-test-httpd-deployment created\n"
-STEP: verifying the pod controlled by e2e-test-httpd-deployment gets created
-[AfterEach] Kubectl run default
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1495
-Aug 30 17:37:29.388: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete deployment e2e-test-httpd-deployment --namespace=kubectl-1921'
-Aug 30 17:37:29.467: INFO: stderr: ""
-Aug 30 17:37:29.467: INFO: stdout: "deployment.apps \"e2e-test-httpd-deployment\" deleted\n"
+Feb 12 10:51:11.854: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 run e2e-test-httpd-job --restart=OnFailure --generator=job/v1 --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-2223'
+Feb 12 10:51:12.064: INFO: stderr: "kubectl run --generator=job/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Feb 12 10:51:12.064: INFO: stdout: "job.batch/e2e-test-httpd-job created\n"
+STEP: verifying the job e2e-test-httpd-job was created
+[AfterEach] Kubectl run job
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1690
+Feb 12 10:51:12.071: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete jobs e2e-test-httpd-job --namespace=kubectl-2223'
+Feb 12 10:51:12.172: INFO: stderr: ""
+Feb 12 10:51:12.172: INFO: stdout: "job.batch \"e2e-test-httpd-job\" deleted\n"
 [AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:37:29.467: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-1921" for this suite.
-•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run default should create an rc or deployment from an image  [Conformance]","total":280,"completed":263,"skipped":4235,"failed":0}
-SSSSSSSSSSSSSSSSS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:51:12.172: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-2223" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run job should create a job from an image when restart is OnFailure [Deprecated] [Conformance]","total":280,"completed":249,"skipped":4073,"failed":0}
+SSSSSSSSS
 ------------------------------
-[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch 
-  watch on custom resource definition objects [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Projected secret 
+  should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Projected secret
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:37:29.487: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename crd-watch
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in crd-watch-3983
+Feb 12 10:51:12.192: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename projected
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in projected-9946
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] watch on custom resource definition objects [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:37:29.669: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Creating first CR 
-Aug 30 17:37:30.299: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-08-30T17:37:30Z generation:1 name:name1 resourceVersion:31564 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:1acf2841-9429-43d2-b5a0-d7ece4423026] num:map[num1:9223372036854775807 num2:1000000]]}
-STEP: Creating second CR
-Aug 30 17:37:40.309: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-08-30T17:37:40Z generation:1 name:name2 resourceVersion:31631 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:0c57ab91-98d6-40f9-adec-66e3d0a78f78] num:map[num1:9223372036854775807 num2:1000000]]}
-STEP: Modifying first CR
-Aug 30 17:37:50.318: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-08-30T17:37:30Z generation:2 name:name1 resourceVersion:31670 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:1acf2841-9429-43d2-b5a0-d7ece4423026] num:map[num1:9223372036854775807 num2:1000000]]}
-STEP: Modifying second CR
-Aug 30 17:38:00.329: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-08-30T17:37:40Z generation:2 name:name2 resourceVersion:31707 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:0c57ab91-98d6-40f9-adec-66e3d0a78f78] num:map[num1:9223372036854775807 num2:1000000]]}
-STEP: Deleting first CR
-Aug 30 17:38:10.342: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-08-30T17:37:30Z generation:2 name:name1 resourceVersion:31742 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name1 uid:1acf2841-9429-43d2-b5a0-d7ece4423026] num:map[num1:9223372036854775807 num2:1000000]]}
-STEP: Deleting second CR
-Aug 30 17:38:20.355: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2020-08-30T17:37:40Z generation:2 name:name2 resourceVersion:31777 selfLink:/apis/mygroup.example.com/v1beta1/noxus/name2 uid:0c57ab91-98d6-40f9-adec-66e3d0a78f78] num:map[num1:9223372036854775807 num2:1000000]]}
-[AfterEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:38:30.875: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "crd-watch-3983" for this suite.
-
-• [SLOW TEST:61.405 seconds]
-[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin]
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  CustomResourceDefinition Watch
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/crd_watch.go:41
-    watch on custom resource definition objects [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating projection with secret that has name projected-secret-test-1cc4ad1a-4cde-40f6-96f3-651f98945696
+STEP: Creating a pod to test consume secrets
+Feb 12 10:51:12.377: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-bc69e4ce-0dcf-4b44-8921-a8cd4138dad4" in namespace "projected-9946" to be "success or failure"
+Feb 12 10:51:12.382: INFO: Pod "pod-projected-secrets-bc69e4ce-0dcf-4b44-8921-a8cd4138dad4": Phase="Pending", Reason="", readiness=false. Elapsed: 4.769193ms
+Feb 12 10:51:14.388: INFO: Pod "pod-projected-secrets-bc69e4ce-0dcf-4b44-8921-a8cd4138dad4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011015544s
+STEP: Saw pod success
+Feb 12 10:51:14.388: INFO: Pod "pod-projected-secrets-bc69e4ce-0dcf-4b44-8921-a8cd4138dad4" satisfied condition "success or failure"
+Feb 12 10:51:14.393: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-projected-secrets-bc69e4ce-0dcf-4b44-8921-a8cd4138dad4 container projected-secret-volume-test: 
+STEP: delete the pod
+Feb 12 10:51:14.446: INFO: Waiting for pod pod-projected-secrets-bc69e4ce-0dcf-4b44-8921-a8cd4138dad4 to disappear
+Feb 12 10:51:14.451: INFO: Pod pod-projected-secrets-bc69e4ce-0dcf-4b44-8921-a8cd4138dad4 no longer exists
+[AfterEach] [sig-storage] Projected secret
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:51:14.451: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "projected-9946" for this suite.
+•{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":250,"skipped":4082,"failed":0}
+SSSSSSSSSSSS
 ------------------------------
-{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance]","total":280,"completed":264,"skipped":4252,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSS
+[sig-node] Downward API 
+  should provide host IP as an env var [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-node] Downward API
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:51:14.470: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7884
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should provide host IP as an env var [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward api env vars
+Feb 12 10:51:14.639: INFO: Waiting up to 5m0s for pod "downward-api-d5438849-b716-4872-9b2d-87baa82839a9" in namespace "downward-api-7884" to be "success or failure"
+Feb 12 10:51:14.652: INFO: Pod "downward-api-d5438849-b716-4872-9b2d-87baa82839a9": Phase="Pending", Reason="", readiness=false. Elapsed: 13.293979ms
+Feb 12 10:51:16.660: INFO: Pod "downward-api-d5438849-b716-4872-9b2d-87baa82839a9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.021050244s
+STEP: Saw pod success
+Feb 12 10:51:16.660: INFO: Pod "downward-api-d5438849-b716-4872-9b2d-87baa82839a9" satisfied condition "success or failure"
+Feb 12 10:51:16.666: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downward-api-d5438849-b716-4872-9b2d-87baa82839a9 container dapi-container: 
+STEP: delete the pod
+Feb 12 10:51:16.705: INFO: Waiting for pod downward-api-d5438849-b716-4872-9b2d-87baa82839a9 to disappear
+Feb 12 10:51:16.710: INFO: Pod downward-api-d5438849-b716-4872-9b2d-87baa82839a9 no longer exists
+[AfterEach] [sig-node] Downward API
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:51:16.710: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-7884" for this suite.
+•{"msg":"PASSED [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance]","total":280,"completed":251,"skipped":4094,"failed":0}
+SSSSS
 ------------------------------
-[sig-apps] Job 
-  should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] Job
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-network] Services 
+  should be able to change the type from ExternalName to ClusterIP [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:38:30.892: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename job
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in job-6171
+Feb 12 10:51:16.729: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename services
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in services-3158
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a job
-STEP: Ensuring job reaches completions
-[AfterEach] [sig-apps] Job
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:38:39.080: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "job-6171" for this suite.
+[BeforeEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:139
+[It] should be able to change the type from ExternalName to ClusterIP [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating a service externalname-service with the type=ExternalName in namespace services-3158
+STEP: changing the ExternalName service to type=ClusterIP
+STEP: creating replication controller externalname-service in namespace services-3158
+I0212 10:51:16.926348      21 runners.go:189] Created replication controller with name: externalname-service, namespace: services-3158, replica count: 2
+I0212 10:51:19.976686      21 runners.go:189] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
+Feb 12 10:51:19.976: INFO: Creating new exec pod
+Feb 12 10:51:22.998: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=services-3158 execpodns9gx -- /bin/sh -x -c nc -zv -t -w 2 externalname-service 80'
+Feb 12 10:51:23.610: INFO: stderr: "+ nc -zv -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n"
+Feb 12 10:51:23.610: INFO: stdout: ""
+Feb 12 10:51:23.611: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=services-3158 execpodns9gx -- /bin/sh -x -c nc -zv -t -w 2 10.240.28.157 80'
+Feb 12 10:51:24.178: INFO: stderr: "+ nc -zv -t -w 2 10.240.28.157 80\nConnection to 10.240.28.157 80 port [tcp/http] succeeded!\n"
+Feb 12 10:51:24.178: INFO: stdout: ""
+Feb 12 10:51:24.178: INFO: Cleaning up the ExternalName to ClusterIP test service
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:51:24.203: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "services-3158" for this suite.
+[AfterEach] [sig-network] Services
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/service.go:143
 
-• [SLOW TEST:8.204 seconds]
-[sig-apps] Job
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:7.493 seconds]
+[sig-network] Services
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+  should be able to change the type from ExternalName to ClusterIP [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]","total":280,"completed":265,"skipped":4277,"failed":0}
-SSSS
+{"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance]","total":280,"completed":252,"skipped":4099,"failed":0}
+SSSSSSSS
 ------------------------------
-[k8s.io] Security Context when creating containers with AllowPrivilegeEscalation 
-  should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Subpath
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:38:39.098: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename security-context-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in security-context-test-8501
+Feb 12 10:51:24.222: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-3561
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39
-[It] should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:38:39.271: INFO: Waiting up to 5m0s for pod "alpine-nnp-false-35c87b39-92ea-4597-bc68-92203237e640" in namespace "security-context-test-8501" to be "success or failure"
-Aug 30 17:38:39.278: INFO: Pod "alpine-nnp-false-35c87b39-92ea-4597-bc68-92203237e640": Phase="Pending", Reason="", readiness=false. Elapsed: 6.470966ms
-Aug 30 17:38:41.284: INFO: Pod "alpine-nnp-false-35c87b39-92ea-4597-bc68-92203237e640": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012415597s
-Aug 30 17:38:43.290: INFO: Pod "alpine-nnp-false-35c87b39-92ea-4597-bc68-92203237e640": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018962621s
-Aug 30 17:38:43.291: INFO: Pod "alpine-nnp-false-35c87b39-92ea-4597-bc68-92203237e640" satisfied condition "success or failure"
-[AfterEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:38:43.382: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "security-context-test-8501" for this suite.
-•{"msg":"PASSED [k8s.io] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":266,"skipped":4281,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] Atomic writer volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
+STEP: Setting up data
+[It] should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating pod pod-subpath-test-configmap-g82r
+STEP: Creating a pod to test atomic-volume-subpath
+Feb 12 10:51:24.400: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-g82r" in namespace "subpath-3561" to be "success or failure"
+Feb 12 10:51:24.405: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Pending", Reason="", readiness=false. Elapsed: 4.853878ms
+Feb 12 10:51:26.411: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Running", Reason="", readiness=true. Elapsed: 2.010829034s
+Feb 12 10:51:28.417: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Running", Reason="", readiness=true. Elapsed: 4.017027053s
+Feb 12 10:51:30.424: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Running", Reason="", readiness=true. Elapsed: 6.023929965s
+Feb 12 10:51:32.432: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Running", Reason="", readiness=true. Elapsed: 8.031841165s
+Feb 12 10:51:34.438: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Running", Reason="", readiness=true. Elapsed: 10.03771399s
+Feb 12 10:51:36.444: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Running", Reason="", readiness=true. Elapsed: 12.044038053s
+Feb 12 10:51:38.451: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Running", Reason="", readiness=true. Elapsed: 14.050840873s
+Feb 12 10:51:40.456: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Running", Reason="", readiness=true. Elapsed: 16.056325218s
+Feb 12 10:51:42.464: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Running", Reason="", readiness=true. Elapsed: 18.063904677s
+Feb 12 10:51:44.471: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Running", Reason="", readiness=true. Elapsed: 20.071309529s
+Feb 12 10:51:46.478: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Running", Reason="", readiness=true. Elapsed: 22.077683429s
+Feb 12 10:51:48.484: INFO: Pod "pod-subpath-test-configmap-g82r": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.08423917s
+STEP: Saw pod success
+Feb 12 10:51:48.484: INFO: Pod "pod-subpath-test-configmap-g82r" satisfied condition "success or failure"
+Feb 12 10:51:48.491: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-subpath-test-configmap-g82r container test-container-subpath-configmap-g82r: 
+STEP: delete the pod
+Feb 12 10:51:48.547: INFO: Waiting for pod pod-subpath-test-configmap-g82r to disappear
+Feb 12 10:51:48.553: INFO: Pod pod-subpath-test-configmap-g82r no longer exists
+STEP: Deleting pod pod-subpath-test-configmap-g82r
+Feb 12 10:51:48.553: INFO: Deleting pod "pod-subpath-test-configmap-g82r" in namespace "subpath-3561"
+[AfterEach] [sig-storage] Subpath
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:51:48.575: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-3561" for this suite.
+
+• [SLOW TEST:24.374 seconds]
+[sig-storage] Subpath
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
+  Atomic writer volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
+    should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] ConfigMap 
-  binary data should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [LinuxOnly] [Conformance]","total":280,"completed":253,"skipped":4107,"failed":0}
+SSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-auth] ServiceAccounts 
+  should mount an API token into pods  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-auth] ServiceAccounts
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:38:43.400: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename configmap
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-1609
+Feb 12 10:51:48.597: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename svcaccounts
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svcaccounts-5492
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] binary data should be reflected in volume [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating configMap with name configmap-test-upd-6cc38b18-7471-41f6-9765-12ddb0b8e6b6
-STEP: Creating the pod
-STEP: Waiting for pod with text data
-STEP: Waiting for pod with binary data
-[AfterEach] [sig-storage] ConfigMap
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:38:45.711: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "configmap-1609" for this suite.
-•{"msg":"PASSED [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":267,"skipped":4322,"failed":0}
-SSSSSS
+[It] should mount an API token into pods  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: getting the auto-created API token
+STEP: reading a file in the container
+Feb 12 10:51:51.340: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-5492 pod-service-account-3e1d66e8-476b-4d9f-82ee-4c3fc33c775e -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token'
+STEP: reading a file in the container
+Feb 12 10:51:51.950: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-5492 pod-service-account-3e1d66e8-476b-4d9f-82ee-4c3fc33c775e -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt'
+STEP: reading a file in the container
+Feb 12 10:51:52.542: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-5492 pod-service-account-3e1d66e8-476b-4d9f-82ee-4c3fc33c775e -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace'
+[AfterEach] [sig-auth] ServiceAccounts
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:51:53.145: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "svcaccounts-5492" for this suite.
+•{"msg":"PASSED [sig-auth] ServiceAccounts should mount an API token into pods  [Conformance]","total":280,"completed":254,"skipped":4125,"failed":0}
+SSSSSSSSSSSSSSSS
 ------------------------------
-[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
-  Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-apps] Deployment 
+  deployment should support proportional scaling [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:38:45.730: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename statefulset
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-8598
+Feb 12 10:51:53.164: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename deployment
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-2447
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
-[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
-STEP: Creating service test in namespace statefulset-8598
-[It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Initializing watcher for selector baz=blah,foo=bar
-STEP: Creating stateful set ss in namespace statefulset-8598
-STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-8598
-Aug 30 17:38:45.911: INFO: Found 0 stateful pods, waiting for 1
-Aug 30 17:38:55.920: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod
-Aug 30 17:38:55.931: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8598 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Aug 30 17:38:56.468: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Aug 30 17:38:56.468: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Aug 30 17:38:56.468: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Aug 30 17:38:56.476: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true
-Aug 30 17:39:06.485: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
-Aug 30 17:39:06.485: INFO: Waiting for statefulset status.replicas updated to 0
-Aug 30 17:39:06.520: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999999771s
-Aug 30 17:39:07.527: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.994253541s
-Aug 30 17:39:08.543: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.987658623s
-Aug 30 17:39:09.549: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.972045034s
-Aug 30 17:39:10.555: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.966007102s
-Aug 30 17:39:11.560: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.960124494s
-Aug 30 17:39:12.566: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.954608542s
-Aug 30 17:39:13.573: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.948536464s
-Aug 30 17:39:14.579: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.942007675s
-Aug 30 17:39:15.585: INFO: Verifying statefulset ss doesn't scale past 1 for another 936.046112ms
-STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-8598
-Aug 30 17:39:16.592: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8598 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:39:17.135: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Aug 30 17:39:17.135: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Aug 30 17:39:17.135: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Aug 30 17:39:17.140: INFO: Found 1 stateful pods, waiting for 3
-Aug 30 17:39:27.147: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true
-Aug 30 17:39:27.147: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true
-Aug 30 17:39:27.147: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true
-STEP: Verifying that stateful set ss was scaled up in order
-STEP: Scale down will halt with unhealthy stateful pod
-Aug 30 17:39:27.155: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8598 ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Aug 30 17:39:27.689: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Aug 30 17:39:27.689: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Aug 30 17:39:27.689: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Aug 30 17:39:27.689: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8598 ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Aug 30 17:39:28.272: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Aug 30 17:39:28.272: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Aug 30 17:39:28.272: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Aug 30 17:39:28.272: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8598 ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
-Aug 30 17:39:28.846: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
-Aug 30 17:39:28.846: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
-Aug 30 17:39:28.846: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
-
-Aug 30 17:39:28.846: INFO: Waiting for statefulset status.replicas updated to 0
-Aug 30 17:39:28.851: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2
-Aug 30 17:39:38.865: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false
-Aug 30 17:39:38.865: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false
-Aug 30 17:39:38.865: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false
-Aug 30 17:39:38.884: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999999809s
-Aug 30 17:39:39.893: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.992858964s
-Aug 30 17:39:40.900: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.986062848s
-Aug 30 17:39:41.906: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.979501863s
-Aug 30 17:39:42.912: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.972627621s
-Aug 30 17:39:43.919: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.967248185s
-Aug 30 17:39:44.925: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.960562097s
-Aug 30 17:39:45.931: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.954152895s
-Aug 30 17:39:46.937: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.948441166s
-Aug 30 17:39:47.944: INFO: Verifying statefulset ss doesn't scale past 3 for another 941.642874ms
-STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-8598
-Aug 30 17:39:48.952: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8598 ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:39:49.523: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Aug 30 17:39:49.523: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Aug 30 17:39:49.523: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Aug 30 17:39:49.523: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8598 ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:39:50.070: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Aug 30 17:39:50.070: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Aug 30 17:39:50.070: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Aug 30 17:39:50.070: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 exec --namespace=statefulset-8598 ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
-Aug 30 17:39:50.621: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
-Aug 30 17:39:50.621: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
-Aug 30 17:39:50.621: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
-
-Aug 30 17:39:50.621: INFO: Scaling statefulset ss to 0
-STEP: Verifying that stateful set ss was scaled down in reverse order
-[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
-Aug 30 17:40:10.647: INFO: Deleting all statefulset in ns statefulset-8598
-Aug 30 17:40:10.652: INFO: Scaling statefulset ss to 0
-Aug 30 17:40:10.666: INFO: Waiting for statefulset status.replicas updated to 0
-Aug 30 17:40:10.670: INFO: Deleting statefulset ss
-[AfterEach] [sig-apps] StatefulSet
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:40:10.689: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "statefulset-8598" for this suite.
+[BeforeEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
+[It] deployment should support proportional scaling [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:51:53.329: INFO: Creating deployment "webserver-deployment"
+Feb 12 10:51:53.338: INFO: Waiting for observed generation 1
+Feb 12 10:51:55.359: INFO: Waiting for all required pods to come up
+Feb 12 10:51:55.368: INFO: Pod name httpd: Found 10 pods out of 10
+STEP: ensuring each pod is running
+Feb 12 10:51:57.386: INFO: Waiting for deployment "webserver-deployment" to complete
+Feb 12 10:51:57.399: INFO: Updating deployment "webserver-deployment" with a non-existent image
+Feb 12 10:51:57.416: INFO: Updating deployment webserver-deployment
+Feb 12 10:51:57.416: INFO: Waiting for observed generation 2
+Feb 12 10:51:59.436: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8
+Feb 12 10:51:59.443: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8
+Feb 12 10:51:59.451: INFO: Waiting for the first rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas
+Feb 12 10:51:59.471: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0
+Feb 12 10:51:59.471: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5
+Feb 12 10:51:59.477: INFO: Waiting for the second rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas
+Feb 12 10:51:59.488: INFO: Verifying that deployment "webserver-deployment" has minimum required number of available replicas
+Feb 12 10:51:59.489: INFO: Scaling up the deployment "webserver-deployment" from 10 to 30
+Feb 12 10:51:59.502: INFO: Updating deployment webserver-deployment
+Feb 12 10:51:59.502: INFO: Waiting for the replicasets of deployment "webserver-deployment" to have desired number of replicas
+Feb 12 10:51:59.516: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20
+Feb 12 10:51:59.537: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13
+[AfterEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
+Feb 12 10:51:59.567: INFO: Deployment "webserver-deployment":
+&Deployment{ObjectMeta:{webserver-deployment  deployment-2447 /apis/apps/v1/namespaces/deployment-2447/deployments/webserver-deployment 173b3c61-080a-42af-aeee-73acd72cc1cb 35061 3 2021-02-12 10:51:52 +0000 UTC   map[name:httpd] map[deployment.kubernetes.io/revision:2] [] []  []},Spec:DeploymentSpec{Replicas:*30,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:httpd] map[] [] []  []} {[] [] [{httpd webserver:404 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc000f4d788  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:13,UpdatedReplicas:5,AvailableReplicas:8,UnavailableReplicas:5,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "webserver-deployment-c7997dcc8" is progressing.,LastUpdateTime:2021-02-12 10:51:56 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,},DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2021-02-12 10:51:58 +0000 UTC,LastTransitionTime:2021-02-12 10:51:58 +0000 UTC,},},ReadyReplicas:8,CollisionCount:nil,},}
+
+Feb 12 10:51:59.592: INFO: New ReplicaSet "webserver-deployment-c7997dcc8" of Deployment "webserver-deployment":
+&ReplicaSet{ObjectMeta:{webserver-deployment-c7997dcc8  deployment-2447 /apis/apps/v1/namespaces/deployment-2447/replicasets/webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 35056 3 2021-02-12 10:51:56 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment webserver-deployment 173b3c61-080a-42af-aeee-73acd72cc1cb 0xc000f4de07 0xc000f4de08}] []  []},Spec:ReplicaSetSpec{Replicas:*13,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: c7997dcc8,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [] []  []} {[] [] [{httpd webserver:404 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc000f4de78  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:5,FullyLabeledReplicas:5,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},}
+Feb 12 10:51:59.592: INFO: All old ReplicaSets of Deployment "webserver-deployment":
+Feb 12 10:51:59.592: INFO: &ReplicaSet{ObjectMeta:{webserver-deployment-595b5b9587  deployment-2447 /apis/apps/v1/namespaces/deployment-2447/replicasets/webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 35054 3 2021-02-12 10:51:52 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment webserver-deployment 173b3c61-080a-42af-aeee-73acd72cc1cb 0xc000f4dd47 0xc000f4dd48}] []  []},Spec:ReplicaSetSpec{Replicas:*20,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 595b5b9587,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [] []  []} {[] [] [{httpd docker.io/library/httpd:2.4.38-alpine [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc000f4dda8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:8,FullyLabeledReplicas:8,ObservedGeneration:2,ReadyReplicas:8,AvailableReplicas:8,Conditions:[]ReplicaSetCondition{},},}
+Feb 12 10:51:59.632: INFO: Pod "webserver-deployment-595b5b9587-2vg2g" is not available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-2vg2g webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-2vg2g c0dec276-9c6c-4d32-9181-91fd9f1e245c 35071 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059ac3d7 0xc0059ac3d8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-xg74t,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:58 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.633: INFO: Pod "webserver-deployment-595b5b9587-5qh8l" is available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-5qh8l webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-5qh8l 181ea002-4b40-45e5-b902-ed462d5f4c88 34935 0 2021-02-12 10:51:52 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.0.50/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059ac4f0 0xc0059ac4f1}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-r6snm,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:139.59.138.127,PodIP:172.25.0.50,StartTime:2021-02-12 10:51:52 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-02-12 10:51:54 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://96439633097f7f9c2c0d54b4cf7fc1e47975786c4abf6aa6e77cb0ffc65cca52,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.0.50,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.633: INFO: Pod "webserver-deployment-595b5b9587-ddfnl" is available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-ddfnl webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-ddfnl 2a7a4856-2a35-4dcf-a6c2-4d471e80aa9f 34900 0 2021-02-12 10:51:52 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.2.74/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059ac670 0xc0059ac671}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-dbhjj,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:53 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:54 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:54 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:104.248.134.247,PodIP:172.25.2.74,StartTime:2021-02-12 10:51:53 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-02-12 10:51:54 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://3b254e572972adbb9d5fb30437b9bb32062e80b1d50af7f2ad1d8462d09cda52,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.2.74,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.633: INFO: Pod "webserver-deployment-595b5b9587-dp5s4" is available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-dp5s4 webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-dp5s4 cf51d294-e371-475c-aeb8-197283449af9 34929 0 2021-02-12 10:51:52 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.0.49/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059ac7e0 0xc0059ac7e1}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-r6snm,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:139.59.138.127,PodIP:172.25.0.49,StartTime:2021-02-12 10:51:52 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-02-12 10:51:53 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://b7c54b1878f5a46ca68afd654c8dd790623a94c298aa677783686cb9ab9537e5,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.0.49,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.634: INFO: Pod "webserver-deployment-595b5b9587-drqmp" is not available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-drqmp webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-drqmp 52346600-aa24-4c24-98d8-f0ac822416f8 35077 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059ac940 0xc0059ac941}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.634: INFO: Pod "webserver-deployment-595b5b9587-dv5x2" is available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-dv5x2 webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-dv5x2 ef9cc300-d6f2-4b65-8db9-d2c152e89291 34946 0 2021-02-12 10:51:52 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.1.239/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059aca37 0xc0059aca38}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-xg74t,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:138.68.97.197,PodIP:172.25.1.239,StartTime:2021-02-12 10:51:52 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-02-12 10:51:54 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://308972d1854e8bdea59b54488f00d596253e65996d68480734a60b590a7f0c50,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.239,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.634: INFO: Pod "webserver-deployment-595b5b9587-g8skq" is not available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-g8skq webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-g8skq 0023b873-2ee6-461c-ac75-b137b1ad488b 35075 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059acba7 0xc0059acba8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.634: INFO: Pod "webserver-deployment-595b5b9587-h9f6b" is not available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-h9f6b webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-h9f6b f852e4b4-fe70-4b85-9062-5aa61feaf625 35079 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059acc97 0xc0059acc98}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-xg74t,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:58 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:58 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:58 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:58 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:138.68.97.197,PodIP:,StartTime:2021-02-12 10:51:58 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:docker.io/library/httpd:2.4.38-alpine,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.635: INFO: Pod "webserver-deployment-595b5b9587-jxtxv" is available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-jxtxv webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-jxtxv 17fce55b-3b4f-482c-a8f2-a51520d6cccf 34932 0 2021-02-12 10:51:52 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.0.51/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059acdf7 0xc0059acdf8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-r6snm,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:139.59.138.127,PodIP:172.25.0.51,StartTime:2021-02-12 10:51:52 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-02-12 10:51:54 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://d4fbe6038eb0d50054c9bd97c00f1da569309e7e20dfcefdc18c5d38bc294601,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.0.51,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.636: INFO: Pod "webserver-deployment-595b5b9587-m2w5m" is not available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-m2w5m webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-m2w5m 9cf27560-08c8-4cc1-a169-d33ecf759565 35064 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059acf60 0xc0059acf61}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-r6snm,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:58 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.636: INFO: Pod "webserver-deployment-595b5b9587-n5dbt" is available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-n5dbt webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-n5dbt aaaa829a-608c-4945-9118-02713212ac59 34952 0 2021-02-12 10:51:52 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.1.240/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059ad070 0xc0059ad071}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-xg74t,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:138.68.97.197,PodIP:172.25.1.240,StartTime:2021-02-12 10:51:52 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-02-12 10:51:54 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://f3396a0b8b4c9af70d5405502f8878ad8fb8483373c666a7b3aff458eb182349,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.240,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.637: INFO: Pod "webserver-deployment-595b5b9587-s6gjc" is available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-s6gjc webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-s6gjc 822f7d29-3bc3-4006-a4e8-165470132265 34941 0 2021-02-12 10:51:52 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.2.76/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059ad1e7 0xc0059ad1e8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-dbhjj,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:53 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:104.248.134.247,PodIP:172.25.2.76,StartTime:2021-02-12 10:51:53 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-02-12 10:51:55 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://aa87ec6e78c7f025eb0b5ebcc6ed96fb15655ed6aede472e453559ed38d0589b,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.2.76,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.637: INFO: Pod "webserver-deployment-595b5b9587-sjzz5" is not available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-sjzz5 webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-sjzz5 095a15b0-45fd-4190-81b8-8569d0fb40da 35087 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059ad350 0xc0059ad351}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-dbhjj,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:58 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.637: INFO: Pod "webserver-deployment-595b5b9587-xd4n4" is available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-xd4n4 webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-xd4n4 f280502c-34b0-4dd3-b93e-f85ed47326b1 34938 0 2021-02-12 10:51:52 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[cni.projectcalico.org/podIP:172.25.2.75/32] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059ad460 0xc0059ad461}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-dbhjj,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:53 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:52 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:104.248.134.247,PodIP:172.25.2.75,StartTime:2021-02-12 10:51:53 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-02-12 10:51:55 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:httpd:2.4.38-alpine,ImageID:docker-pullable://httpd@sha256:eb8ccf084cf3e80eece1add239effefd171eb39adbc154d33c14260d905d4060,ContainerID:docker://2f66343f5c106fc422a5df281050f38ad5bf78a6b5d3ff7caf381e689e70b787,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.2.75,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.637: INFO: Pod "webserver-deployment-595b5b9587-zxpsf" is not available:
+&Pod{ObjectMeta:{webserver-deployment-595b5b9587-zxpsf webserver-deployment-595b5b9587- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-595b5b9587-zxpsf bf3ca80a-075c-4b2b-97aa-b55a19d66180 35083 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:595b5b9587] map[] [{apps/v1 ReplicaSet webserver-deployment-595b5b9587 5fc8596e-5569-4d3e-bf1d-47b991c62f77 0xc0059ad5c0 0xc0059ad5c1}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:docker.io/library/httpd:2.4.38-alpine,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-xg74t,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:58 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.637: INFO: Pod "webserver-deployment-c7997dcc8-2sgkl" is not available:
+&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-2sgkl webserver-deployment-c7997dcc8- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-c7997dcc8-2sgkl 8dd190ea-09b1-42d1-a1ba-381e9e125b59 35045 0 2021-02-12 10:51:56 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.1.243/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 0xc0059ad6d0 0xc0059ad6d1}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-xg74t,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:138.68.97.197,PodIP:,StartTime:2021-02-12 10:51:56 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.638: INFO: Pod "webserver-deployment-c7997dcc8-2wk7m" is not available:
+&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-2wk7m webserver-deployment-c7997dcc8- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-c7997dcc8-2wk7m 0e1268be-72a3-4f94-941a-31b23fa3eb12 35048 0 2021-02-12 10:51:56 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.1.244/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 0xc0059ad857 0xc0059ad858}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-xg74t,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:138.68.97.197,PodIP:,StartTime:2021-02-12 10:51:56 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.638: INFO: Pod "webserver-deployment-c7997dcc8-5phf7" is not available:
+&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-5phf7 webserver-deployment-c7997dcc8- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-c7997dcc8-5phf7 d9fff096-e65d-4587-88cd-c1ff33bf4297 35031 0 2021-02-12 10:51:56 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.0.52/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 0xc0059ad9e7 0xc0059ad9e8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-r6snm,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:139.59.138.127,PodIP:,StartTime:2021-02-12 10:51:56 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.639: INFO: Pod "webserver-deployment-c7997dcc8-96mm7" is not available:
+&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-96mm7 webserver-deployment-c7997dcc8- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-c7997dcc8-96mm7 4b6ecead-ce01-4e9a-8521-f25507bd6da8 35044 0 2021-02-12 10:51:56 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.1.242/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 0xc0059adb67 0xc0059adb68}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-xg74t,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:138.68.97.197,PodIP:,StartTime:2021-02-12 10:51:56 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.639: INFO: Pod "webserver-deployment-c7997dcc8-b45g7" is not available:
+&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-b45g7 webserver-deployment-c7997dcc8- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-c7997dcc8-b45g7 a42e516e-811a-4683-b674-54bf8c348275 35084 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 0xc0059adce7 0xc0059adce8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.640: INFO: Pod "webserver-deployment-c7997dcc8-lklkc" is not available:
+&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-lklkc webserver-deployment-c7997dcc8- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-c7997dcc8-lklkc bd6db79d-a297-4a20-a68b-b2293dac908d 35088 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 0xc0059adde7 0xc0059adde8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.640: INFO: Pod "webserver-deployment-c7997dcc8-qkrk5" is not available:
+&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-qkrk5 webserver-deployment-c7997dcc8- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-c7997dcc8-qkrk5 8240b16f-fd38-4395-8008-57c6e4e176ec 35082 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 0xc0059adef7 0xc0059adef8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.640: INFO: Pod "webserver-deployment-c7997dcc8-qmdq4" is not available:
+&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-qmdq4 webserver-deployment-c7997dcc8- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-c7997dcc8-qmdq4 04ca2855-d065-404a-8565-b0097ec5fb52 35076 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 0xc0059adff7 0xc0059adff8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-dbhjj,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:58 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.641: INFO: Pod "webserver-deployment-c7997dcc8-tmtts" is not available:
+&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-tmtts webserver-deployment-c7997dcc8- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-c7997dcc8-tmtts cf5e0ff8-0a52-4fe0-b214-2744e4c7120b 35069 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 0xc0035e2110 0xc0035e2111}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-r6snm,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:58 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.641: INFO: Pod "webserver-deployment-c7997dcc8-vrb2q" is not available:
+&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-vrb2q webserver-deployment-c7997dcc8- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-c7997dcc8-vrb2q 6e29fd2b-7559-4d03-9c66-cad2dfcb8a18 35032 0 2021-02-12 10:51:56 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[cni.projectcalico.org/podIP:172.25.2.77/32] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 0xc0035e2240 0xc0035e2241}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-dbhjj,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:57 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:57 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:57 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:56 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:104.248.134.247,PodIP:,StartTime:2021-02-12 10:51:57 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.641: INFO: Pod "webserver-deployment-c7997dcc8-xjxkz" is not available:
+&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-xjxkz webserver-deployment-c7997dcc8- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-c7997dcc8-xjxkz c4fd42bf-400c-4e9b-8b2b-f3baab8cd8e0 35080 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 0xc0035e23b7 0xc0035e23b8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+Feb 12 10:51:59.641: INFO: Pod "webserver-deployment-c7997dcc8-z7sgg" is not available:
+&Pod{ObjectMeta:{webserver-deployment-c7997dcc8-z7sgg webserver-deployment-c7997dcc8- deployment-2447 /api/v1/namespaces/deployment-2447/pods/webserver-deployment-c7997dcc8-z7sgg 91f35ced-bcff-466f-aa8a-563ad8e1648a 35086 0 2021-02-12 10:51:58 +0000 UTC   map[name:httpd pod-template-hash:c7997dcc8] map[] [{apps/v1 ReplicaSet webserver-deployment-c7997dcc8 b5c3b96f-ef98-4e5a-a390-f902161ccfa4 0xc0035e24c7 0xc0035e24c8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-kg6fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-kg6fb,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-kg6fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-dbhjj,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 10:51:58 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},}
+[AfterEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:51:59.641: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-2447" for this suite.
 
-• [SLOW TEST:84.975 seconds]
-[sig-apps] StatefulSet
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-    Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:6.547 seconds]
+[sig-apps] Deployment
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  deployment should support proportional scaling [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]","total":280,"completed":268,"skipped":4328,"failed":0}
-SSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-apps] Deployment deployment should support proportional scaling [Conformance]","total":280,"completed":255,"skipped":4141,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [sig-network] DNS 
   should provide DNS for ExternalName services [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:40:10.707: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:51:59.711: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename dns
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-4076
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in dns-1333
 STEP: Waiting for a default service account to be provisioned in namespace
 [It] should provide DNS for ExternalName services [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: Creating a test externalName service
-STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-4076.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-4076.svc.cluster.local; sleep 1; done
+STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-1333.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-1333.svc.cluster.local; sleep 1; done
 
-STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-4076.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-4076.svc.cluster.local; sleep 1; done
+STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-1333.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-1333.svc.cluster.local; sleep 1; done
 
 STEP: creating a pod to probe DNS
 STEP: submitting the pod to kubernetes
 STEP: retrieving the pod
 STEP: looking for the results for each expected name from probers
-Aug 30 17:40:13.126: INFO: DNS probes using dns-test-faf4e455-ff42-4394-affa-751eafbe802c succeeded
+Feb 12 10:52:06.168: INFO: DNS probes using dns-test-0a5e456e-e545-494e-ae54-71d5278df9db succeeded
 
 STEP: deleting the pod
 STEP: changing the externalName to bar.example.com
-STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-4076.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-4076.svc.cluster.local; sleep 1; done
+STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-1333.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-1333.svc.cluster.local; sleep 1; done
 
-STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-4076.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-4076.svc.cluster.local; sleep 1; done
+STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-1333.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-1333.svc.cluster.local; sleep 1; done
 
 STEP: creating a second pod to probe DNS
 STEP: submitting the pod to kubernetes
 STEP: retrieving the pod
 STEP: looking for the results for each expected name from probers
-Aug 30 17:40:15.313: INFO: File wheezy_udp@dns-test-service-3.dns-4076.svc.cluster.local from pod  dns-4076/dns-test-771385ad-1462-40b3-b5e5-4f57f49672bc contains 'foo.example.com.
-' instead of 'bar.example.com.'
-Aug 30 17:40:15.402: INFO: File jessie_udp@dns-test-service-3.dns-4076.svc.cluster.local from pod  dns-4076/dns-test-771385ad-1462-40b3-b5e5-4f57f49672bc contains 'foo.example.com.
-' instead of 'bar.example.com.'
-Aug 30 17:40:15.402: INFO: Lookups using dns-4076/dns-test-771385ad-1462-40b3-b5e5-4f57f49672bc failed for: [wheezy_udp@dns-test-service-3.dns-4076.svc.cluster.local jessie_udp@dns-test-service-3.dns-4076.svc.cluster.local]
-
-Aug 30 17:40:20.499: INFO: DNS probes using dns-test-771385ad-1462-40b3-b5e5-4f57f49672bc succeeded
+Feb 12 10:52:10.474: INFO: DNS probes using dns-test-64a84da4-f3f4-4c55-aa22-53b95bba7ce4 succeeded
 
 STEP: deleting the pod
 STEP: changing the service to type=ClusterIP
-STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-4076.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-4076.svc.cluster.local; sleep 1; done
+STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-1333.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-1333.svc.cluster.local; sleep 1; done
 
-STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-4076.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-4076.svc.cluster.local; sleep 1; done
+STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-1333.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-1333.svc.cluster.local; sleep 1; done
 
 STEP: creating a third pod to probe DNS
 STEP: submitting the pod to kubernetes
 STEP: retrieving the pod
 STEP: looking for the results for each expected name from probers
-Aug 30 17:40:22.791: INFO: DNS probes using dns-test-68d45c06-4090-4960-a0ae-9e2a8ad4c0f7 succeeded
+Feb 12 10:52:12.760: INFO: DNS probes using dns-test-8b550732-70e6-4364-9933-dcaf7b9dda2a succeeded
 
 STEP: deleting the pod
 STEP: deleting the test externalName service
 [AfterEach] [sig-network] DNS
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:40:22.829: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "dns-4076" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:52:12.800: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "dns-1333" for this suite.
 
-• [SLOW TEST:12.140 seconds]
+• [SLOW TEST:13.110 seconds]
 [sig-network] DNS
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
   should provide DNS for ExternalName services [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-network] DNS should provide DNS for ExternalName services [Conformance]","total":280,"completed":256,"skipped":4164,"failed":0}
+SSSSSSS
+------------------------------
+[sig-storage] Downward API volume 
+  should provide podname only [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:52:12.821: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-1648
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should provide podname only [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Feb 12 10:52:12.988: INFO: Waiting up to 5m0s for pod "downwardapi-volume-0e5146cd-2d89-4389-a60b-3415d6920bf1" in namespace "downward-api-1648" to be "success or failure"
+Feb 12 10:52:12.993: INFO: Pod "downwardapi-volume-0e5146cd-2d89-4389-a60b-3415d6920bf1": Phase="Pending", Reason="", readiness=false. Elapsed: 4.282281ms
+Feb 12 10:52:15.000: INFO: Pod "downwardapi-volume-0e5146cd-2d89-4389-a60b-3415d6920bf1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011227556s
+STEP: Saw pod success
+Feb 12 10:52:15.000: INFO: Pod "downwardapi-volume-0e5146cd-2d89-4389-a60b-3415d6920bf1" satisfied condition "success or failure"
+Feb 12 10:52:15.005: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-0e5146cd-2d89-4389-a60b-3415d6920bf1 container client-container: 
+STEP: delete the pod
+Feb 12 10:52:15.043: INFO: Waiting for pod downwardapi-volume-0e5146cd-2d89-4389-a60b-3415d6920bf1 to disappear
+Feb 12 10:52:15.048: INFO: Pod downwardapi-volume-0e5146cd-2d89-4389-a60b-3415d6920bf1 no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:52:15.048: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-1648" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance]","total":280,"completed":257,"skipped":4171,"failed":0}
+SSS
+------------------------------
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should be able to deny pod and configmap creation [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:52:15.072: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-5028
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Feb 12 10:52:15.493: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:52:17.512: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723934, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723934, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723934, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723934, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Feb 12 10:52:20.537: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should be able to deny pod and configmap creation [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Registering the webhook via the AdmissionRegistration API
+STEP: create a pod that should be denied by the webhook
+STEP: create a pod that causes the webhook to hang
+STEP: create a configmap that should be denied by the webhook
+STEP: create a configmap that should be admitted by the webhook
+STEP: update (PUT) the admitted configmap to a non-compliant one should be rejected by the webhook
+STEP: update (PATCH) the admitted configmap to a non-compliant one should be rejected by the webhook
+STEP: create a namespace that bypass the webhook
+STEP: create a configmap that violates the webhook policy but is in a whitelisted namespace
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:52:31.503: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-5028" for this suite.
+STEP: Destroying namespace "webhook-5028-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
+
+• [SLOW TEST:16.537 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should be able to deny pod and configmap creation [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance]","total":280,"completed":258,"skipped":4174,"failed":0}
+SSSSSSSSSSSSSSS
 ------------------------------
-{"msg":"PASSED [sig-network] DNS should provide DNS for ExternalName services [Conformance]","total":280,"completed":269,"skipped":4345,"failed":0}
+[sig-cli] Kubectl client Kubectl run default 
+  should create an rc or deployment from an image  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:52:31.611: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename kubectl
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-6535
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[BeforeEach] Kubectl run default
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1490
+[It] should create an rc or deployment from an image  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: running the image docker.io/library/httpd:2.4.38-alpine
+Feb 12 10:52:31.772: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 run e2e-test-httpd-deployment --image=docker.io/library/httpd:2.4.38-alpine --namespace=kubectl-6535'
+Feb 12 10:52:31.896: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Feb 12 10:52:31.896: INFO: stdout: "deployment.apps/e2e-test-httpd-deployment created\n"
+STEP: verifying the pod controlled by e2e-test-httpd-deployment gets created
+[AfterEach] Kubectl run default
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1496
+Feb 12 10:52:33.914: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete deployment e2e-test-httpd-deployment --namespace=kubectl-6535'
+Feb 12 10:52:34.018: INFO: stderr: ""
+Feb 12 10:52:34.018: INFO: stdout: "deployment.apps \"e2e-test-httpd-deployment\" deleted\n"
+[AfterEach] [sig-cli] Kubectl client
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:52:34.018: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-6535" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run default should create an rc or deployment from an image  [Conformance]","total":280,"completed":259,"skipped":4189,"failed":0}
 SSSSSS
 ------------------------------
-[sig-network] Service endpoints latency 
-  should not be very high  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-network] Service endpoints latency
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[k8s.io] Docker Containers 
+  should be able to override the image's default command and arguments [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Docker Containers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:40:22.848: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename svc-latency
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in svc-latency-1197
+Feb 12 10:52:34.039: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename containers
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-3673
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should not be very high  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:40:23.021: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: creating replication controller svc-latency-rc in namespace svc-latency-1197
-I0830 17:40:23.035532      23 runners.go:189] Created replication controller with name: svc-latency-rc, namespace: svc-latency-1197, replica count: 1
-I0830 17:40:24.085900      23 runners.go:189] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-I0830 17:40:25.086162      23 runners.go:189] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady 
-Aug 30 17:40:25.201: INFO: Created: latency-svc-mtcsf
-Aug 30 17:40:25.207: INFO: Got endpoints: latency-svc-mtcsf [21.027577ms]
-Aug 30 17:40:25.220: INFO: Created: latency-svc-sfp86
-Aug 30 17:40:25.227: INFO: Got endpoints: latency-svc-sfp86 [19.747092ms]
-Aug 30 17:40:25.229: INFO: Created: latency-svc-wb9wk
-Aug 30 17:40:25.235: INFO: Got endpoints: latency-svc-wb9wk [27.371439ms]
-Aug 30 17:40:25.235: INFO: Created: latency-svc-z2jqv
-Aug 30 17:40:25.241: INFO: Got endpoints: latency-svc-z2jqv [32.864282ms]
-Aug 30 17:40:25.242: INFO: Created: latency-svc-fzptm
-Aug 30 17:40:25.247: INFO: Got endpoints: latency-svc-fzptm [38.99426ms]
-Aug 30 17:40:25.250: INFO: Created: latency-svc-xwt6d
-Aug 30 17:40:25.254: INFO: Got endpoints: latency-svc-xwt6d [45.109874ms]
-Aug 30 17:40:25.263: INFO: Created: latency-svc-57ls2
-Aug 30 17:40:25.265: INFO: Got endpoints: latency-svc-57ls2 [56.536062ms]
-Aug 30 17:40:25.268: INFO: Created: latency-svc-mqnzh
-Aug 30 17:40:25.269: INFO: Got endpoints: latency-svc-mqnzh [59.951364ms]
-Aug 30 17:40:25.272: INFO: Created: latency-svc-jm8bc
-Aug 30 17:40:25.278: INFO: Got endpoints: latency-svc-jm8bc [68.91451ms]
-Aug 30 17:40:25.284: INFO: Created: latency-svc-cknnk
-Aug 30 17:40:25.290: INFO: Got endpoints: latency-svc-cknnk [80.728323ms]
-Aug 30 17:40:25.290: INFO: Created: latency-svc-6z86n
-Aug 30 17:40:25.295: INFO: Created: latency-svc-8hzmb
-Aug 30 17:40:25.297: INFO: Got endpoints: latency-svc-6z86n [87.855098ms]
-Aug 30 17:40:25.302: INFO: Got endpoints: latency-svc-8hzmb [92.146731ms]
-Aug 30 17:40:25.306: INFO: Created: latency-svc-4jphk
-Aug 30 17:40:25.314: INFO: Got endpoints: latency-svc-4jphk [104.801016ms]
-Aug 30 17:40:25.315: INFO: Created: latency-svc-9c8fj
-Aug 30 17:40:25.320: INFO: Got endpoints: latency-svc-9c8fj [111.190115ms]
-Aug 30 17:40:25.321: INFO: Created: latency-svc-zbmth
-Aug 30 17:40:25.325: INFO: Got endpoints: latency-svc-zbmth [115.842441ms]
-Aug 30 17:40:25.328: INFO: Created: latency-svc-v8xml
-Aug 30 17:40:25.334: INFO: Got endpoints: latency-svc-v8xml [125.143967ms]
-Aug 30 17:40:25.336: INFO: Created: latency-svc-tnzqm
-Aug 30 17:40:25.342: INFO: Created: latency-svc-mclht
-Aug 30 17:40:25.342: INFO: Got endpoints: latency-svc-tnzqm [115.39452ms]
-Aug 30 17:40:25.348: INFO: Got endpoints: latency-svc-mclht [112.566162ms]
-Aug 30 17:40:25.353: INFO: Created: latency-svc-lfgsm
-Aug 30 17:40:25.355: INFO: Created: latency-svc-bbdhl
-Aug 30 17:40:25.356: INFO: Got endpoints: latency-svc-lfgsm [115.218468ms]
-Aug 30 17:40:25.362: INFO: Got endpoints: latency-svc-bbdhl [114.388375ms]
-Aug 30 17:40:25.364: INFO: Created: latency-svc-xtmxj
-Aug 30 17:40:25.370: INFO: Got endpoints: latency-svc-xtmxj [116.696427ms]
-Aug 30 17:40:25.371: INFO: Created: latency-svc-bthd6
-Aug 30 17:40:25.378: INFO: Got endpoints: latency-svc-bthd6 [112.429517ms]
-Aug 30 17:40:25.384: INFO: Created: latency-svc-2br4g
-Aug 30 17:40:25.390: INFO: Got endpoints: latency-svc-2br4g [121.471705ms]
-Aug 30 17:40:25.394: INFO: Created: latency-svc-h6s9q
-Aug 30 17:40:25.402: INFO: Created: latency-svc-vtpst
-Aug 30 17:40:25.402: INFO: Got endpoints: latency-svc-h6s9q [123.994493ms]
-Aug 30 17:40:25.410: INFO: Created: latency-svc-brjr8
-Aug 30 17:40:25.410: INFO: Got endpoints: latency-svc-vtpst [120.110218ms]
-Aug 30 17:40:25.415: INFO: Got endpoints: latency-svc-brjr8 [118.111443ms]
-Aug 30 17:40:25.419: INFO: Created: latency-svc-v5k8f
-Aug 30 17:40:25.426: INFO: Got endpoints: latency-svc-v5k8f [124.344694ms]
-Aug 30 17:40:25.431: INFO: Created: latency-svc-cbdrq
-Aug 30 17:40:25.437: INFO: Got endpoints: latency-svc-cbdrq [122.689519ms]
-Aug 30 17:40:25.444: INFO: Created: latency-svc-gmpwd
-Aug 30 17:40:25.452: INFO: Got endpoints: latency-svc-gmpwd [131.640602ms]
-Aug 30 17:40:25.456: INFO: Created: latency-svc-pcgps
-Aug 30 17:40:25.462: INFO: Got endpoints: latency-svc-pcgps [136.218889ms]
-Aug 30 17:40:25.465: INFO: Created: latency-svc-k9fwk
-Aug 30 17:40:25.474: INFO: Created: latency-svc-x5c5d
-Aug 30 17:40:25.474: INFO: Got endpoints: latency-svc-k9fwk [140.739548ms]
-Aug 30 17:40:25.480: INFO: Got endpoints: latency-svc-x5c5d [137.277606ms]
-Aug 30 17:40:25.481: INFO: Created: latency-svc-js5dv
-Aug 30 17:40:25.488: INFO: Got endpoints: latency-svc-js5dv [140.030187ms]
-Aug 30 17:40:25.490: INFO: Created: latency-svc-dfk59
-Aug 30 17:40:25.495: INFO: Got endpoints: latency-svc-dfk59 [138.72036ms]
-Aug 30 17:40:25.499: INFO: Created: latency-svc-4qxgq
-Aug 30 17:40:25.504: INFO: Got endpoints: latency-svc-4qxgq [142.367773ms]
-Aug 30 17:40:25.508: INFO: Created: latency-svc-7v2bd
-Aug 30 17:40:25.516: INFO: Got endpoints: latency-svc-7v2bd [145.804483ms]
-Aug 30 17:40:25.518: INFO: Created: latency-svc-fksd5
-Aug 30 17:40:25.526: INFO: Created: latency-svc-q4cnm
-Aug 30 17:40:25.533: INFO: Created: latency-svc-q27d6
-Aug 30 17:40:25.542: INFO: Created: latency-svc-dggbp
-Aug 30 17:40:25.550: INFO: Created: latency-svc-phqlt
-Aug 30 17:40:25.556: INFO: Created: latency-svc-cgrl4
-Aug 30 17:40:25.558: INFO: Got endpoints: latency-svc-fksd5 [180.239938ms]
-Aug 30 17:40:25.566: INFO: Created: latency-svc-z44ms
-Aug 30 17:40:25.579: INFO: Created: latency-svc-n64v6
-Aug 30 17:40:25.591: INFO: Created: latency-svc-pt9g4
-Aug 30 17:40:25.600: INFO: Created: latency-svc-btkv9
-Aug 30 17:40:25.606: INFO: Got endpoints: latency-svc-q4cnm [216.12899ms]
-Aug 30 17:40:25.607: INFO: Created: latency-svc-p7wmx
-Aug 30 17:40:25.616: INFO: Created: latency-svc-9qcgx
-Aug 30 17:40:25.625: INFO: Created: latency-svc-mrl8t
-Aug 30 17:40:25.630: INFO: Created: latency-svc-dfcnj
-Aug 30 17:40:25.642: INFO: Created: latency-svc-qzklq
-Aug 30 17:40:25.651: INFO: Created: latency-svc-dkvbz
-Aug 30 17:40:25.656: INFO: Created: latency-svc-zmqvh
-Aug 30 17:40:25.657: INFO: Got endpoints: latency-svc-q27d6 [255.051452ms]
-Aug 30 17:40:25.671: INFO: Created: latency-svc-69rrm
-Aug 30 17:40:25.708: INFO: Got endpoints: latency-svc-dggbp [297.633383ms]
-Aug 30 17:40:25.725: INFO: Created: latency-svc-5v7m8
-Aug 30 17:40:25.757: INFO: Got endpoints: latency-svc-phqlt [342.137437ms]
-Aug 30 17:40:25.775: INFO: Created: latency-svc-26xw6
-Aug 30 17:40:25.808: INFO: Got endpoints: latency-svc-cgrl4 [382.163721ms]
-Aug 30 17:40:25.823: INFO: Created: latency-svc-2gzpg
-Aug 30 17:40:25.858: INFO: Got endpoints: latency-svc-z44ms [420.88148ms]
-Aug 30 17:40:25.880: INFO: Created: latency-svc-8g4sx
-Aug 30 17:40:25.907: INFO: Got endpoints: latency-svc-n64v6 [455.264055ms]
-Aug 30 17:40:25.922: INFO: Created: latency-svc-zqtcn
-Aug 30 17:40:25.960: INFO: Got endpoints: latency-svc-pt9g4 [498.679ms]
-Aug 30 17:40:25.979: INFO: Created: latency-svc-jpswr
-Aug 30 17:40:26.014: INFO: Got endpoints: latency-svc-btkv9 [539.593052ms]
-Aug 30 17:40:26.031: INFO: Created: latency-svc-w2bh8
-Aug 30 17:40:26.057: INFO: Got endpoints: latency-svc-p7wmx [577.231985ms]
-Aug 30 17:40:26.071: INFO: Created: latency-svc-cdh8m
-Aug 30 17:40:26.107: INFO: Got endpoints: latency-svc-9qcgx [618.651758ms]
-Aug 30 17:40:26.120: INFO: Created: latency-svc-jdg7f
-Aug 30 17:40:26.158: INFO: Got endpoints: latency-svc-mrl8t [662.216785ms]
-Aug 30 17:40:26.174: INFO: Created: latency-svc-d8f7n
-Aug 30 17:40:26.207: INFO: Got endpoints: latency-svc-dfcnj [703.188085ms]
-Aug 30 17:40:26.220: INFO: Created: latency-svc-wgr88
-Aug 30 17:40:26.258: INFO: Got endpoints: latency-svc-qzklq [741.808567ms]
-Aug 30 17:40:26.274: INFO: Created: latency-svc-zwn69
-Aug 30 17:40:26.307: INFO: Got endpoints: latency-svc-dkvbz [748.918449ms]
-Aug 30 17:40:26.321: INFO: Created: latency-svc-pm98t
-Aug 30 17:40:26.357: INFO: Got endpoints: latency-svc-zmqvh [750.355651ms]
-Aug 30 17:40:26.370: INFO: Created: latency-svc-6f8b2
-Aug 30 17:40:26.409: INFO: Got endpoints: latency-svc-69rrm [751.780504ms]
-Aug 30 17:40:26.422: INFO: Created: latency-svc-xdp56
-Aug 30 17:40:26.457: INFO: Got endpoints: latency-svc-5v7m8 [749.808028ms]
-Aug 30 17:40:26.472: INFO: Created: latency-svc-t8mg6
-Aug 30 17:40:26.508: INFO: Got endpoints: latency-svc-26xw6 [750.848902ms]
-Aug 30 17:40:26.520: INFO: Created: latency-svc-l2lqk
-Aug 30 17:40:26.557: INFO: Got endpoints: latency-svc-2gzpg [748.923618ms]
-Aug 30 17:40:26.572: INFO: Created: latency-svc-5hhfh
-Aug 30 17:40:26.608: INFO: Got endpoints: latency-svc-8g4sx [750.013375ms]
-Aug 30 17:40:26.621: INFO: Created: latency-svc-68lxv
-Aug 30 17:40:26.658: INFO: Got endpoints: latency-svc-zqtcn [750.353553ms]
-Aug 30 17:40:26.671: INFO: Created: latency-svc-6chlq
-Aug 30 17:40:26.707: INFO: Got endpoints: latency-svc-jpswr [746.741792ms]
-Aug 30 17:40:26.721: INFO: Created: latency-svc-2d7pz
-Aug 30 17:40:26.758: INFO: Got endpoints: latency-svc-w2bh8 [743.699187ms]
-Aug 30 17:40:26.771: INFO: Created: latency-svc-fmzj5
-Aug 30 17:40:26.808: INFO: Got endpoints: latency-svc-cdh8m [751.340517ms]
-Aug 30 17:40:26.822: INFO: Created: latency-svc-qpmtb
-Aug 30 17:40:26.862: INFO: Got endpoints: latency-svc-jdg7f [755.155608ms]
-Aug 30 17:40:26.877: INFO: Created: latency-svc-2bxff
-Aug 30 17:40:26.908: INFO: Got endpoints: latency-svc-d8f7n [750.1407ms]
-Aug 30 17:40:26.923: INFO: Created: latency-svc-qgrqp
-Aug 30 17:40:26.959: INFO: Got endpoints: latency-svc-wgr88 [751.392303ms]
-Aug 30 17:40:26.974: INFO: Created: latency-svc-xgxc6
-Aug 30 17:40:27.016: INFO: Got endpoints: latency-svc-zwn69 [758.094973ms]
-Aug 30 17:40:27.032: INFO: Created: latency-svc-jsbqr
-Aug 30 17:40:27.058: INFO: Got endpoints: latency-svc-pm98t [750.507079ms]
-Aug 30 17:40:27.076: INFO: Created: latency-svc-gcdgz
-Aug 30 17:40:27.107: INFO: Got endpoints: latency-svc-6f8b2 [750.519771ms]
-Aug 30 17:40:27.120: INFO: Created: latency-svc-6rlqj
-Aug 30 17:40:27.157: INFO: Got endpoints: latency-svc-xdp56 [748.374491ms]
-Aug 30 17:40:27.172: INFO: Created: latency-svc-tnpch
-Aug 30 17:40:27.208: INFO: Got endpoints: latency-svc-t8mg6 [750.625412ms]
-Aug 30 17:40:27.224: INFO: Created: latency-svc-vrq9b
-Aug 30 17:40:27.259: INFO: Got endpoints: latency-svc-l2lqk [751.191465ms]
-Aug 30 17:40:27.274: INFO: Created: latency-svc-f5dqn
-Aug 30 17:40:27.311: INFO: Got endpoints: latency-svc-5hhfh [753.593291ms]
-Aug 30 17:40:27.326: INFO: Created: latency-svc-8flb9
-Aug 30 17:40:27.357: INFO: Got endpoints: latency-svc-68lxv [749.095614ms]
-Aug 30 17:40:27.371: INFO: Created: latency-svc-2ndv8
-Aug 30 17:40:27.410: INFO: Got endpoints: latency-svc-6chlq [751.791899ms]
-Aug 30 17:40:27.422: INFO: Created: latency-svc-wpnkz
-Aug 30 17:40:27.458: INFO: Got endpoints: latency-svc-2d7pz [750.373818ms]
-Aug 30 17:40:27.472: INFO: Created: latency-svc-lmt5b
-Aug 30 17:40:27.507: INFO: Got endpoints: latency-svc-fmzj5 [749.044054ms]
-Aug 30 17:40:27.522: INFO: Created: latency-svc-stblj
-Aug 30 17:40:27.557: INFO: Got endpoints: latency-svc-qpmtb [748.828823ms]
-Aug 30 17:40:27.569: INFO: Created: latency-svc-ktgmp
-Aug 30 17:40:27.608: INFO: Got endpoints: latency-svc-2bxff [746.104044ms]
-Aug 30 17:40:27.623: INFO: Created: latency-svc-ztskb
-Aug 30 17:40:27.657: INFO: Got endpoints: latency-svc-qgrqp [749.533985ms]
-Aug 30 17:40:27.681: INFO: Created: latency-svc-rjdkj
-Aug 30 17:40:27.707: INFO: Got endpoints: latency-svc-xgxc6 [748.491534ms]
-Aug 30 17:40:27.721: INFO: Created: latency-svc-vvft2
-Aug 30 17:40:27.757: INFO: Got endpoints: latency-svc-jsbqr [740.597973ms]
-Aug 30 17:40:27.772: INFO: Created: latency-svc-qx2jf
-Aug 30 17:40:27.807: INFO: Got endpoints: latency-svc-gcdgz [749.026397ms]
-Aug 30 17:40:27.820: INFO: Created: latency-svc-6z5qg
-Aug 30 17:40:27.858: INFO: Got endpoints: latency-svc-6rlqj [750.037082ms]
-Aug 30 17:40:27.872: INFO: Created: latency-svc-hjzhl
-Aug 30 17:40:27.908: INFO: Got endpoints: latency-svc-tnpch [751.162759ms]
-Aug 30 17:40:27.925: INFO: Created: latency-svc-7lcz6
-Aug 30 17:40:27.958: INFO: Got endpoints: latency-svc-vrq9b [749.509216ms]
-Aug 30 17:40:27.973: INFO: Created: latency-svc-474tq
-Aug 30 17:40:28.014: INFO: Got endpoints: latency-svc-f5dqn [754.438236ms]
-Aug 30 17:40:28.026: INFO: Created: latency-svc-4msw7
-Aug 30 17:40:28.057: INFO: Got endpoints: latency-svc-8flb9 [746.617233ms]
-Aug 30 17:40:28.074: INFO: Created: latency-svc-pgk7h
-Aug 30 17:40:28.107: INFO: Got endpoints: latency-svc-2ndv8 [750.552746ms]
-Aug 30 17:40:28.120: INFO: Created: latency-svc-2fj6k
-Aug 30 17:40:28.157: INFO: Got endpoints: latency-svc-wpnkz [746.906536ms]
-Aug 30 17:40:28.174: INFO: Created: latency-svc-s62m6
-Aug 30 17:40:28.206: INFO: Got endpoints: latency-svc-lmt5b [748.893136ms]
-Aug 30 17:40:28.220: INFO: Created: latency-svc-qs98c
-Aug 30 17:40:28.258: INFO: Got endpoints: latency-svc-stblj [750.811925ms]
-Aug 30 17:40:28.272: INFO: Created: latency-svc-hjxmk
-Aug 30 17:40:28.307: INFO: Got endpoints: latency-svc-ktgmp [749.902398ms]
-Aug 30 17:40:28.321: INFO: Created: latency-svc-q488f
-Aug 30 17:40:28.357: INFO: Got endpoints: latency-svc-ztskb [748.258222ms]
-Aug 30 17:40:28.372: INFO: Created: latency-svc-vn9hv
-Aug 30 17:40:28.408: INFO: Got endpoints: latency-svc-rjdkj [750.481405ms]
-Aug 30 17:40:28.421: INFO: Created: latency-svc-6fdbx
-Aug 30 17:40:28.457: INFO: Got endpoints: latency-svc-vvft2 [749.275855ms]
-Aug 30 17:40:28.475: INFO: Created: latency-svc-lm2v7
-Aug 30 17:40:28.511: INFO: Got endpoints: latency-svc-qx2jf [754.211249ms]
-Aug 30 17:40:28.525: INFO: Created: latency-svc-cqtqc
-Aug 30 17:40:28.556: INFO: Got endpoints: latency-svc-6z5qg [748.977457ms]
-Aug 30 17:40:28.570: INFO: Created: latency-svc-s8rw9
-Aug 30 17:40:28.608: INFO: Got endpoints: latency-svc-hjzhl [750.163781ms]
-Aug 30 17:40:28.621: INFO: Created: latency-svc-w8xtm
-Aug 30 17:40:28.658: INFO: Got endpoints: latency-svc-7lcz6 [749.715529ms]
-Aug 30 17:40:28.673: INFO: Created: latency-svc-d92xt
-Aug 30 17:40:28.708: INFO: Got endpoints: latency-svc-474tq [750.096147ms]
-Aug 30 17:40:28.721: INFO: Created: latency-svc-wjpmj
-Aug 30 17:40:28.757: INFO: Got endpoints: latency-svc-4msw7 [743.59917ms]
-Aug 30 17:40:28.772: INFO: Created: latency-svc-hdwp7
-Aug 30 17:40:28.807: INFO: Got endpoints: latency-svc-pgk7h [749.957765ms]
-Aug 30 17:40:28.822: INFO: Created: latency-svc-d5hsn
-Aug 30 17:40:28.857: INFO: Got endpoints: latency-svc-2fj6k [749.780399ms]
-Aug 30 17:40:28.871: INFO: Created: latency-svc-95xdz
-Aug 30 17:40:28.908: INFO: Got endpoints: latency-svc-s62m6 [751.568163ms]
-Aug 30 17:40:28.924: INFO: Created: latency-svc-qbjzc
-Aug 30 17:40:28.957: INFO: Got endpoints: latency-svc-qs98c [750.935842ms]
-Aug 30 17:40:28.970: INFO: Created: latency-svc-9qwpg
-Aug 30 17:40:29.007: INFO: Got endpoints: latency-svc-hjxmk [748.45108ms]
-Aug 30 17:40:29.019: INFO: Created: latency-svc-mpv2r
-Aug 30 17:40:29.058: INFO: Got endpoints: latency-svc-q488f [750.959362ms]
-Aug 30 17:40:29.076: INFO: Created: latency-svc-c27lx
-Aug 30 17:40:29.112: INFO: Got endpoints: latency-svc-vn9hv [754.824827ms]
-Aug 30 17:40:29.128: INFO: Created: latency-svc-rrdnh
-Aug 30 17:40:29.169: INFO: Got endpoints: latency-svc-6fdbx [760.736085ms]
-Aug 30 17:40:29.194: INFO: Created: latency-svc-q4kbf
-Aug 30 17:40:29.207: INFO: Got endpoints: latency-svc-lm2v7 [750.785818ms]
-Aug 30 17:40:29.220: INFO: Created: latency-svc-cvqzh
-Aug 30 17:40:29.259: INFO: Got endpoints: latency-svc-cqtqc [747.415829ms]
-Aug 30 17:40:29.274: INFO: Created: latency-svc-q6hjh
-Aug 30 17:40:29.307: INFO: Got endpoints: latency-svc-s8rw9 [751.456671ms]
-Aug 30 17:40:29.321: INFO: Created: latency-svc-kg8c2
-Aug 30 17:40:29.359: INFO: Got endpoints: latency-svc-w8xtm [750.82407ms]
-Aug 30 17:40:29.375: INFO: Created: latency-svc-gsbkb
-Aug 30 17:40:29.407: INFO: Got endpoints: latency-svc-d92xt [749.336362ms]
-Aug 30 17:40:29.422: INFO: Created: latency-svc-ldc7g
-Aug 30 17:40:29.458: INFO: Got endpoints: latency-svc-wjpmj [750.219094ms]
-Aug 30 17:40:29.473: INFO: Created: latency-svc-mxpmh
-Aug 30 17:40:29.508: INFO: Got endpoints: latency-svc-hdwp7 [750.501449ms]
-Aug 30 17:40:29.521: INFO: Created: latency-svc-pltwv
-Aug 30 17:40:29.559: INFO: Got endpoints: latency-svc-d5hsn [751.393725ms]
-Aug 30 17:40:29.572: INFO: Created: latency-svc-74wbm
-Aug 30 17:40:29.608: INFO: Got endpoints: latency-svc-95xdz [750.009318ms]
-Aug 30 17:40:29.622: INFO: Created: latency-svc-wlgj4
-Aug 30 17:40:29.657: INFO: Got endpoints: latency-svc-qbjzc [749.184806ms]
-Aug 30 17:40:29.670: INFO: Created: latency-svc-kbcvr
-Aug 30 17:40:29.711: INFO: Got endpoints: latency-svc-9qwpg [753.198298ms]
-Aug 30 17:40:29.725: INFO: Created: latency-svc-2kbxz
-Aug 30 17:40:29.759: INFO: Got endpoints: latency-svc-mpv2r [752.380471ms]
-Aug 30 17:40:29.778: INFO: Created: latency-svc-s7bjh
-Aug 30 17:40:29.808: INFO: Got endpoints: latency-svc-c27lx [749.376372ms]
-Aug 30 17:40:29.828: INFO: Created: latency-svc-s28hx
-Aug 30 17:40:29.858: INFO: Got endpoints: latency-svc-rrdnh [746.743075ms]
-Aug 30 17:40:29.873: INFO: Created: latency-svc-5mql7
-Aug 30 17:40:29.907: INFO: Got endpoints: latency-svc-q4kbf [738.457575ms]
-Aug 30 17:40:29.921: INFO: Created: latency-svc-bjzbv
-Aug 30 17:40:29.962: INFO: Got endpoints: latency-svc-cvqzh [754.036872ms]
-Aug 30 17:40:29.975: INFO: Created: latency-svc-rv52m
-Aug 30 17:40:30.007: INFO: Got endpoints: latency-svc-q6hjh [748.334898ms]
-Aug 30 17:40:30.022: INFO: Created: latency-svc-w6rmt
-Aug 30 17:40:30.058: INFO: Got endpoints: latency-svc-kg8c2 [750.714656ms]
-Aug 30 17:40:30.072: INFO: Created: latency-svc-sf9b2
-Aug 30 17:40:30.107: INFO: Got endpoints: latency-svc-gsbkb [748.668967ms]
-Aug 30 17:40:30.122: INFO: Created: latency-svc-wn7z6
-Aug 30 17:40:30.157: INFO: Got endpoints: latency-svc-ldc7g [749.785173ms]
-Aug 30 17:40:30.173: INFO: Created: latency-svc-54qd4
-Aug 30 17:40:30.208: INFO: Got endpoints: latency-svc-mxpmh [749.402901ms]
-Aug 30 17:40:30.221: INFO: Created: latency-svc-tkj68
-Aug 30 17:40:30.259: INFO: Got endpoints: latency-svc-pltwv [750.735277ms]
-Aug 30 17:40:30.274: INFO: Created: latency-svc-zsffq
-Aug 30 17:40:30.308: INFO: Got endpoints: latency-svc-74wbm [749.12238ms]
-Aug 30 17:40:30.321: INFO: Created: latency-svc-nhbjk
-Aug 30 17:40:30.360: INFO: Got endpoints: latency-svc-wlgj4 [752.201013ms]
-Aug 30 17:40:30.375: INFO: Created: latency-svc-vrd58
-Aug 30 17:40:30.408: INFO: Got endpoints: latency-svc-kbcvr [750.758051ms]
-Aug 30 17:40:30.423: INFO: Created: latency-svc-fbbqw
-Aug 30 17:40:30.457: INFO: Got endpoints: latency-svc-2kbxz [746.344645ms]
-Aug 30 17:40:30.472: INFO: Created: latency-svc-l8krb
-Aug 30 17:40:30.507: INFO: Got endpoints: latency-svc-s7bjh [747.733484ms]
-Aug 30 17:40:30.520: INFO: Created: latency-svc-svxrl
-Aug 30 17:40:30.559: INFO: Got endpoints: latency-svc-s28hx [751.770098ms]
-Aug 30 17:40:30.571: INFO: Created: latency-svc-pw5lt
-Aug 30 17:40:30.607: INFO: Got endpoints: latency-svc-5mql7 [749.006624ms]
-Aug 30 17:40:30.621: INFO: Created: latency-svc-r8ldv
-Aug 30 17:40:30.659: INFO: Got endpoints: latency-svc-bjzbv [751.450032ms]
-Aug 30 17:40:30.673: INFO: Created: latency-svc-9cvwl
-Aug 30 17:40:30.708: INFO: Got endpoints: latency-svc-rv52m [746.23213ms]
-Aug 30 17:40:30.723: INFO: Created: latency-svc-8jmbf
-Aug 30 17:40:30.759: INFO: Got endpoints: latency-svc-w6rmt [751.444774ms]
-Aug 30 17:40:30.776: INFO: Created: latency-svc-f5tbf
-Aug 30 17:40:30.808: INFO: Got endpoints: latency-svc-sf9b2 [749.820774ms]
-Aug 30 17:40:30.822: INFO: Created: latency-svc-trrhn
-Aug 30 17:40:30.860: INFO: Got endpoints: latency-svc-wn7z6 [752.208863ms]
-Aug 30 17:40:30.876: INFO: Created: latency-svc-gvjjr
-Aug 30 17:40:30.907: INFO: Got endpoints: latency-svc-54qd4 [749.897046ms]
-Aug 30 17:40:30.926: INFO: Created: latency-svc-fwfgp
-Aug 30 17:40:30.958: INFO: Got endpoints: latency-svc-tkj68 [750.032985ms]
-Aug 30 17:40:30.972: INFO: Created: latency-svc-cv8q4
-Aug 30 17:40:31.013: INFO: Got endpoints: latency-svc-zsffq [753.942733ms]
-Aug 30 17:40:31.026: INFO: Created: latency-svc-6jp95
-Aug 30 17:40:31.060: INFO: Got endpoints: latency-svc-nhbjk [751.581256ms]
-Aug 30 17:40:31.074: INFO: Created: latency-svc-z6w94
-Aug 30 17:40:31.108: INFO: Got endpoints: latency-svc-vrd58 [747.721953ms]
-Aug 30 17:40:31.122: INFO: Created: latency-svc-4qmh5
-Aug 30 17:40:31.159: INFO: Got endpoints: latency-svc-fbbqw [750.188758ms]
-Aug 30 17:40:31.174: INFO: Created: latency-svc-nhwpg
-Aug 30 17:40:31.207: INFO: Got endpoints: latency-svc-l8krb [750.398426ms]
-Aug 30 17:40:31.221: INFO: Created: latency-svc-7zbp4
-Aug 30 17:40:31.258: INFO: Got endpoints: latency-svc-svxrl [751.345062ms]
-Aug 30 17:40:31.272: INFO: Created: latency-svc-6q2jv
-Aug 30 17:40:31.313: INFO: Got endpoints: latency-svc-pw5lt [753.327947ms]
-Aug 30 17:40:31.349: INFO: Created: latency-svc-zhrr7
-Aug 30 17:40:31.362: INFO: Got endpoints: latency-svc-r8ldv [755.054112ms]
-Aug 30 17:40:31.386: INFO: Created: latency-svc-5rfdk
-Aug 30 17:40:31.407: INFO: Got endpoints: latency-svc-9cvwl [748.469815ms]
-Aug 30 17:40:31.422: INFO: Created: latency-svc-krskz
-Aug 30 17:40:31.459: INFO: Got endpoints: latency-svc-8jmbf [750.816982ms]
-Aug 30 17:40:31.473: INFO: Created: latency-svc-stcfq
-Aug 30 17:40:31.507: INFO: Got endpoints: latency-svc-f5tbf [748.355561ms]
-Aug 30 17:40:31.521: INFO: Created: latency-svc-h92cs
-Aug 30 17:40:31.558: INFO: Got endpoints: latency-svc-trrhn [750.053366ms]
-Aug 30 17:40:31.573: INFO: Created: latency-svc-plcm6
-Aug 30 17:40:31.607: INFO: Got endpoints: latency-svc-gvjjr [747.846015ms]
-Aug 30 17:40:31.621: INFO: Created: latency-svc-gbhxp
-Aug 30 17:40:31.657: INFO: Got endpoints: latency-svc-fwfgp [750.228857ms]
-Aug 30 17:40:31.671: INFO: Created: latency-svc-vf54h
-Aug 30 17:40:31.710: INFO: Got endpoints: latency-svc-cv8q4 [751.908599ms]
-Aug 30 17:40:31.723: INFO: Created: latency-svc-b8xm2
-Aug 30 17:40:31.763: INFO: Got endpoints: latency-svc-6jp95 [749.954908ms]
-Aug 30 17:40:31.777: INFO: Created: latency-svc-hw5vk
-Aug 30 17:40:31.808: INFO: Got endpoints: latency-svc-z6w94 [747.929809ms]
-Aug 30 17:40:31.824: INFO: Created: latency-svc-wgfdm
-Aug 30 17:40:31.858: INFO: Got endpoints: latency-svc-4qmh5 [750.773634ms]
-Aug 30 17:40:31.873: INFO: Created: latency-svc-svcgd
-Aug 30 17:40:31.909: INFO: Got endpoints: latency-svc-nhwpg [749.965167ms]
-Aug 30 17:40:31.923: INFO: Created: latency-svc-4vp7g
-Aug 30 17:40:31.958: INFO: Got endpoints: latency-svc-7zbp4 [750.36664ms]
-Aug 30 17:40:31.974: INFO: Created: latency-svc-2rcdp
-Aug 30 17:40:32.017: INFO: Got endpoints: latency-svc-6q2jv [758.455567ms]
-Aug 30 17:40:32.032: INFO: Created: latency-svc-b2lxg
-Aug 30 17:40:32.057: INFO: Got endpoints: latency-svc-zhrr7 [744.587165ms]
-Aug 30 17:40:32.077: INFO: Created: latency-svc-ghkxx
-Aug 30 17:40:32.108: INFO: Got endpoints: latency-svc-5rfdk [745.373197ms]
-Aug 30 17:40:32.125: INFO: Created: latency-svc-4464s
-Aug 30 17:40:32.159: INFO: Got endpoints: latency-svc-krskz [751.960873ms]
-Aug 30 17:40:32.173: INFO: Created: latency-svc-wn4fj
-Aug 30 17:40:32.209: INFO: Got endpoints: latency-svc-stcfq [750.143944ms]
-Aug 30 17:40:32.228: INFO: Created: latency-svc-glmcc
-Aug 30 17:40:32.257: INFO: Got endpoints: latency-svc-h92cs [749.938407ms]
-Aug 30 17:40:32.272: INFO: Created: latency-svc-f9xfq
-Aug 30 17:40:32.307: INFO: Got endpoints: latency-svc-plcm6 [749.151887ms]
-Aug 30 17:40:32.322: INFO: Created: latency-svc-6cj7x
-Aug 30 17:40:32.360: INFO: Got endpoints: latency-svc-gbhxp [752.067327ms]
-Aug 30 17:40:32.373: INFO: Created: latency-svc-rcn9b
-Aug 30 17:40:32.407: INFO: Got endpoints: latency-svc-vf54h [749.711859ms]
-Aug 30 17:40:32.422: INFO: Created: latency-svc-mwnwc
-Aug 30 17:40:32.459: INFO: Got endpoints: latency-svc-b8xm2 [748.931835ms]
-Aug 30 17:40:32.472: INFO: Created: latency-svc-wtrfc
-Aug 30 17:40:32.506: INFO: Got endpoints: latency-svc-hw5vk [743.195028ms]
-Aug 30 17:40:32.522: INFO: Created: latency-svc-dvs9f
-Aug 30 17:40:32.558: INFO: Got endpoints: latency-svc-wgfdm [750.2298ms]
-Aug 30 17:40:32.571: INFO: Created: latency-svc-zl7jm
-Aug 30 17:40:32.608: INFO: Got endpoints: latency-svc-svcgd [749.413394ms]
-Aug 30 17:40:32.622: INFO: Created: latency-svc-95tkg
-Aug 30 17:40:32.657: INFO: Got endpoints: latency-svc-4vp7g [748.807499ms]
-Aug 30 17:40:32.671: INFO: Created: latency-svc-7jm9t
-Aug 30 17:40:32.708: INFO: Got endpoints: latency-svc-2rcdp [749.75062ms]
-Aug 30 17:40:32.722: INFO: Created: latency-svc-cdkgx
-Aug 30 17:40:32.760: INFO: Got endpoints: latency-svc-b2lxg [743.402254ms]
-Aug 30 17:40:32.773: INFO: Created: latency-svc-ndnpd
-Aug 30 17:40:32.807: INFO: Got endpoints: latency-svc-ghkxx [749.829467ms]
-Aug 30 17:40:32.822: INFO: Created: latency-svc-bcvn9
-Aug 30 17:40:32.858: INFO: Got endpoints: latency-svc-4464s [749.665511ms]
-Aug 30 17:40:32.874: INFO: Created: latency-svc-xk49v
-Aug 30 17:40:32.908: INFO: Got endpoints: latency-svc-wn4fj [748.620725ms]
-Aug 30 17:40:32.924: INFO: Created: latency-svc-m267r
-Aug 30 17:40:32.958: INFO: Got endpoints: latency-svc-glmcc [748.983877ms]
-Aug 30 17:40:32.973: INFO: Created: latency-svc-xfl2r
-Aug 30 17:40:33.010: INFO: Got endpoints: latency-svc-f9xfq [752.876983ms]
-Aug 30 17:40:33.023: INFO: Created: latency-svc-vpdzx
-Aug 30 17:40:33.058: INFO: Got endpoints: latency-svc-6cj7x [750.94493ms]
-Aug 30 17:40:33.108: INFO: Got endpoints: latency-svc-rcn9b [748.509201ms]
-Aug 30 17:40:33.158: INFO: Got endpoints: latency-svc-mwnwc [750.30676ms]
-Aug 30 17:40:33.209: INFO: Got endpoints: latency-svc-wtrfc [749.840268ms]
-Aug 30 17:40:33.258: INFO: Got endpoints: latency-svc-dvs9f [751.532256ms]
-Aug 30 17:40:33.307: INFO: Got endpoints: latency-svc-zl7jm [749.509709ms]
-Aug 30 17:40:33.358: INFO: Got endpoints: latency-svc-95tkg [749.700255ms]
-Aug 30 17:40:33.408: INFO: Got endpoints: latency-svc-7jm9t [750.275171ms]
-Aug 30 17:40:33.458: INFO: Got endpoints: latency-svc-cdkgx [750.530946ms]
-Aug 30 17:40:33.508: INFO: Got endpoints: latency-svc-ndnpd [747.745505ms]
-Aug 30 17:40:33.559: INFO: Got endpoints: latency-svc-bcvn9 [751.921701ms]
-Aug 30 17:40:33.608: INFO: Got endpoints: latency-svc-xk49v [750.447661ms]
-Aug 30 17:40:33.658: INFO: Got endpoints: latency-svc-m267r [750.019235ms]
-Aug 30 17:40:33.708: INFO: Got endpoints: latency-svc-xfl2r [749.935921ms]
-Aug 30 17:40:33.758: INFO: Got endpoints: latency-svc-vpdzx [748.105029ms]
-Aug 30 17:40:33.758: INFO: Latencies: [19.747092ms 27.371439ms 32.864282ms 38.99426ms 45.109874ms 56.536062ms 59.951364ms 68.91451ms 80.728323ms 87.855098ms 92.146731ms 104.801016ms 111.190115ms 112.429517ms 112.566162ms 114.388375ms 115.218468ms 115.39452ms 115.842441ms 116.696427ms 118.111443ms 120.110218ms 121.471705ms 122.689519ms 123.994493ms 124.344694ms 125.143967ms 131.640602ms 136.218889ms 137.277606ms 138.72036ms 140.030187ms 140.739548ms 142.367773ms 145.804483ms 180.239938ms 216.12899ms 255.051452ms 297.633383ms 342.137437ms 382.163721ms 420.88148ms 455.264055ms 498.679ms 539.593052ms 577.231985ms 618.651758ms 662.216785ms 703.188085ms 738.457575ms 740.597973ms 741.808567ms 743.195028ms 743.402254ms 743.59917ms 743.699187ms 744.587165ms 745.373197ms 746.104044ms 746.23213ms 746.344645ms 746.617233ms 746.741792ms 746.743075ms 746.906536ms 747.415829ms 747.721953ms 747.733484ms 747.745505ms 747.846015ms 747.929809ms 748.105029ms 748.258222ms 748.334898ms 748.355561ms 748.374491ms 748.45108ms 748.469815ms 748.491534ms 748.509201ms 748.620725ms 748.668967ms 748.807499ms 748.828823ms 748.893136ms 748.918449ms 748.923618ms 748.931835ms 748.977457ms 748.983877ms 749.006624ms 749.026397ms 749.044054ms 749.095614ms 749.12238ms 749.151887ms 749.184806ms 749.275855ms 749.336362ms 749.376372ms 749.402901ms 749.413394ms 749.509216ms 749.509709ms 749.533985ms 749.665511ms 749.700255ms 749.711859ms 749.715529ms 749.75062ms 749.780399ms 749.785173ms 749.808028ms 749.820774ms 749.829467ms 749.840268ms 749.897046ms 749.902398ms 749.935921ms 749.938407ms 749.954908ms 749.957765ms 749.965167ms 750.009318ms 750.013375ms 750.019235ms 750.032985ms 750.037082ms 750.053366ms 750.096147ms 750.1407ms 750.143944ms 750.163781ms 750.188758ms 750.219094ms 750.228857ms 750.2298ms 750.275171ms 750.30676ms 750.353553ms 750.355651ms 750.36664ms 750.373818ms 750.398426ms 750.447661ms 750.481405ms 750.501449ms 750.507079ms 750.519771ms 750.530946ms 750.552746ms 750.625412ms 750.714656ms 750.735277ms 750.758051ms 750.773634ms 750.785818ms 750.811925ms 750.816982ms 750.82407ms 750.848902ms 750.935842ms 750.94493ms 750.959362ms 751.162759ms 751.191465ms 751.340517ms 751.345062ms 751.392303ms 751.393725ms 751.444774ms 751.450032ms 751.456671ms 751.532256ms 751.568163ms 751.581256ms 751.770098ms 751.780504ms 751.791899ms 751.908599ms 751.921701ms 751.960873ms 752.067327ms 752.201013ms 752.208863ms 752.380471ms 752.876983ms 753.198298ms 753.327947ms 753.593291ms 753.942733ms 754.036872ms 754.211249ms 754.438236ms 754.824827ms 755.054112ms 755.155608ms 758.094973ms 758.455567ms 760.736085ms]
-Aug 30 17:40:33.759: INFO: 50 %ile: 749.402901ms
-Aug 30 17:40:33.759: INFO: 90 %ile: 751.921701ms
-Aug 30 17:40:33.759: INFO: 99 %ile: 758.455567ms
-Aug 30 17:40:33.759: INFO: Total sample count: 200
-[AfterEach] [sig-network] Service endpoints latency
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:40:33.759: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "svc-latency-1197" for this suite.
+[It] should be able to override the image's default command and arguments [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test override all
+Feb 12 10:52:34.208: INFO: Waiting up to 5m0s for pod "client-containers-9ba25ad4-deb1-43ae-9a4c-3add4484897d" in namespace "containers-3673" to be "success or failure"
+Feb 12 10:52:34.217: INFO: Pod "client-containers-9ba25ad4-deb1-43ae-9a4c-3add4484897d": Phase="Pending", Reason="", readiness=false. Elapsed: 8.682619ms
+Feb 12 10:52:36.222: INFO: Pod "client-containers-9ba25ad4-deb1-43ae-9a4c-3add4484897d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.013944931s
+STEP: Saw pod success
+Feb 12 10:52:36.222: INFO: Pod "client-containers-9ba25ad4-deb1-43ae-9a4c-3add4484897d" satisfied condition "success or failure"
+Feb 12 10:52:36.227: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod client-containers-9ba25ad4-deb1-43ae-9a4c-3add4484897d container test-container: 
+STEP: delete the pod
+Feb 12 10:52:36.261: INFO: Waiting for pod client-containers-9ba25ad4-deb1-43ae-9a4c-3add4484897d to disappear
+Feb 12 10:52:36.266: INFO: Pod client-containers-9ba25ad4-deb1-43ae-9a4c-3add4484897d no longer exists
+[AfterEach] [k8s.io] Docker Containers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:52:36.266: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "containers-3673" for this suite.
+•{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance]","total":280,"completed":260,"skipped":4195,"failed":0}
+SSSSSSSS
+------------------------------
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] 
+  should mutate configmap [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:52:36.289: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename webhook
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in webhook-1847
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:87
+STEP: Setting up server cert
+STEP: Create role binding to let webhook read extension-apiserver-authentication
+STEP: Deploying the webhook pod
+STEP: Wait for the deployment to be ready
+Feb 12 10:52:37.065: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set
+Feb 12 10:52:39.082: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723956, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723956, loc:(*time.Location)(0x791c680)}}, Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723956, loc:(*time.Location)(0x791c680)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63748723956, loc:(*time.Location)(0x791c680)}}, Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5f65f8c764\" is progressing."}}, CollisionCount:(*int32)(nil)}
+STEP: Deploying the webhook service
+STEP: Verifying the service has paired with the endpoint
+Feb 12 10:52:42.105: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1
+[It] should mutate configmap [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Registering the mutating configmap webhook via the AdmissionRegistration API
+STEP: create a configmap that should be updated by the webhook
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:52:42.453: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "webhook-1847" for this suite.
+STEP: Destroying namespace "webhook-1847-markers" for this suite.
+[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/webhook.go:102
 
-• [SLOW TEST:10.927 seconds]
-[sig-network] Service endpoints latency
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:23
-  should not be very high  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:6.261 seconds]
+[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should mutate configmap [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-network] Service endpoints latency should not be very high  [Conformance]","total":280,"completed":270,"skipped":4351,"failed":0}
-S
+{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance]","total":280,"completed":261,"skipped":4203,"failed":0}
+[sig-node] ConfigMap 
+  should be consumable via environment variable [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-node] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:52:42.551: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-6966
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable via environment variable [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap configmap-6966/configmap-test-4d41e8cc-331f-4173-a64a-331faead4085
+STEP: Creating a pod to test consume configMaps
+Feb 12 10:52:42.730: INFO: Waiting up to 5m0s for pod "pod-configmaps-38827e73-4469-4f98-a022-16728950807e" in namespace "configmap-6966" to be "success or failure"
+Feb 12 10:52:42.737: INFO: Pod "pod-configmaps-38827e73-4469-4f98-a022-16728950807e": Phase="Pending", Reason="", readiness=false. Elapsed: 7.04386ms
+Feb 12 10:52:44.744: INFO: Pod "pod-configmaps-38827e73-4469-4f98-a022-16728950807e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.013315204s
+STEP: Saw pod success
+Feb 12 10:52:44.744: INFO: Pod "pod-configmaps-38827e73-4469-4f98-a022-16728950807e" satisfied condition "success or failure"
+Feb 12 10:52:44.749: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-configmaps-38827e73-4469-4f98-a022-16728950807e container env-test: 
+STEP: delete the pod
+Feb 12 10:52:44.786: INFO: Waiting for pod pod-configmaps-38827e73-4469-4f98-a022-16728950807e to disappear
+Feb 12 10:52:44.790: INFO: Pod pod-configmaps-38827e73-4469-4f98-a022-16728950807e no longer exists
+[AfterEach] [sig-node] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:52:44.790: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-6966" for this suite.
+•{"msg":"PASSED [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance]","total":280,"completed":262,"skipped":4203,"failed":0}
+SSSSSSS
 ------------------------------
-[sig-apps] ReplicationController 
-  should release no longer matching pods [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-api-machinery] Watchers 
+  should observe add, update, and delete watch notifications on configmaps [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Watchers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:40:33.776: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename replication-controller
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-6685
+Feb 12 10:52:44.812: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename watch
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in watch-544
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should release no longer matching pods [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Given a ReplicationController is created
-STEP: When the matched label of one of its pods change
-Aug 30 17:40:33.955: INFO: Pod name pod-release: Found 0 pods out of 1
-Aug 30 17:40:38.969: INFO: Pod name pod-release: Found 1 pods out of 1
-STEP: Then the pod is released
-[AfterEach] [sig-apps] ReplicationController
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:40:39.995: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "replication-controller-6685" for this suite.
+[It] should observe add, update, and delete watch notifications on configmaps [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating a watch on configmaps with label A
+STEP: creating a watch on configmaps with label B
+STEP: creating a watch on configmaps with label A or B
+STEP: creating a configmap with label A and ensuring the correct watchers observe the notification
+Feb 12 10:52:44.985: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-544 /api/v1/namespaces/watch-544/configmaps/e2e-watch-test-configmap-a 67b331a6-a95e-4ee5-9f78-623d946e7342 35989 0 2021-02-12 10:52:44 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Feb 12 10:52:44.985: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-544 /api/v1/namespaces/watch-544/configmaps/e2e-watch-test-configmap-a 67b331a6-a95e-4ee5-9f78-623d946e7342 35989 0 2021-02-12 10:52:44 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
+STEP: modifying configmap A and ensuring the correct watchers observe the notification
+Feb 12 10:52:55.007: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-544 /api/v1/namespaces/watch-544/configmaps/e2e-watch-test-configmap-a 67b331a6-a95e-4ee5-9f78-623d946e7342 36071 0 2021-02-12 10:52:44 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+Feb 12 10:52:55.008: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-544 /api/v1/namespaces/watch-544/configmaps/e2e-watch-test-configmap-a 67b331a6-a95e-4ee5-9f78-623d946e7342 36071 0 2021-02-12 10:52:44 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},}
+STEP: modifying configmap A again and ensuring the correct watchers observe the notification
+Feb 12 10:53:05.025: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-544 /api/v1/namespaces/watch-544/configmaps/e2e-watch-test-configmap-a 67b331a6-a95e-4ee5-9f78-623d946e7342 36118 0 2021-02-12 10:52:44 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Feb 12 10:53:05.025: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-544 /api/v1/namespaces/watch-544/configmaps/e2e-watch-test-configmap-a 67b331a6-a95e-4ee5-9f78-623d946e7342 36118 0 2021-02-12 10:52:44 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+STEP: deleting configmap A and ensuring the correct watchers observe the notification
+Feb 12 10:53:15.042: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-544 /api/v1/namespaces/watch-544/configmaps/e2e-watch-test-configmap-a 67b331a6-a95e-4ee5-9f78-623d946e7342 36164 0 2021-02-12 10:52:44 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+Feb 12 10:53:15.042: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a  watch-544 /api/v1/namespaces/watch-544/configmaps/e2e-watch-test-configmap-a 67b331a6-a95e-4ee5-9f78-623d946e7342 36164 0 2021-02-12 10:52:44 +0000 UTC   map[watch-this-configmap:multiple-watchers-A] map[] [] []  []},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},}
+STEP: creating a configmap with label B and ensuring the correct watchers observe the notification
+Feb 12 10:53:25.054: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b  watch-544 /api/v1/namespaces/watch-544/configmaps/e2e-watch-test-configmap-b 83dc4f1f-fe85-46b4-8bbe-21d1d2e8c41d 36210 0 2021-02-12 10:53:24 +0000 UTC   map[watch-this-configmap:multiple-watchers-B] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Feb 12 10:53:25.054: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b  watch-544 /api/v1/namespaces/watch-544/configmaps/e2e-watch-test-configmap-b 83dc4f1f-fe85-46b4-8bbe-21d1d2e8c41d 36210 0 2021-02-12 10:53:24 +0000 UTC   map[watch-this-configmap:multiple-watchers-B] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
+STEP: deleting configmap B and ensuring the correct watchers observe the notification
+Feb 12 10:53:35.091: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b  watch-544 /api/v1/namespaces/watch-544/configmaps/e2e-watch-test-configmap-b 83dc4f1f-fe85-46b4-8bbe-21d1d2e8c41d 36256 0 2021-02-12 10:53:24 +0000 UTC   map[watch-this-configmap:multiple-watchers-B] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
+Feb 12 10:53:35.091: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b  watch-544 /api/v1/namespaces/watch-544/configmaps/e2e-watch-test-configmap-b 83dc4f1f-fe85-46b4-8bbe-21d1d2e8c41d 36256 0 2021-02-12 10:53:24 +0000 UTC   map[watch-this-configmap:multiple-watchers-B] map[] [] []  []},Data:map[string]string{},BinaryData:map[string][]byte{},}
+[AfterEach] [sig-api-machinery] Watchers
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:53:45.092: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "watch-544" for this suite.
 
-• [SLOW TEST:6.241 seconds]
-[sig-apps] ReplicationController
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
-  should release no longer matching pods [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:60.298 seconds]
+[sig-api-machinery] Watchers
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should observe add, update, and delete watch notifications on configmaps [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-apps] ReplicationController should release no longer matching pods [Conformance]","total":280,"completed":271,"skipped":4352,"failed":0}
-SSSSSSSSSSSSSSSSSSSS
+{"msg":"PASSED [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance]","total":280,"completed":263,"skipped":4210,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-cli] Kubectl client Kubectl rolling-update 
-  should support rolling-update to same image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[sig-storage] Downward API volume 
+  should update labels on modification [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:53:45.111: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-6129
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should update labels on modification [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating the pod
+Feb 12 10:53:47.832: INFO: Successfully updated pod "labelsupdate3871662e-3ebc-4c99-998a-cb0e28937398"
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:53:49.867: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-6129" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance]","total":280,"completed":264,"skipped":4244,"failed":0}
+SS
+------------------------------
+[sig-cli] Kubectl client Kubectl run deployment 
+  should create a deployment from an image [Deprecated] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:40:40.017: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:53:49.885: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename kubectl
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-2938
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in kubectl-5016
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:272
-[BeforeEach] Kubectl rolling-update
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1585
-[It] should support rolling-update to same image  [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:273
+[BeforeEach] Kubectl run deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1629
+[It] should create a deployment from an image [Deprecated] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 STEP: running the image docker.io/library/httpd:2.4.38-alpine
-Aug 30 17:40:40.189: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 run e2e-test-httpd-rc --image=docker.io/library/httpd:2.4.38-alpine --generator=run/v1 --namespace=kubectl-2938'
-Aug 30 17:40:40.271: INFO: stderr: "kubectl run --generator=run/v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
-Aug 30 17:40:40.271: INFO: stdout: "replicationcontroller/e2e-test-httpd-rc created\n"
-STEP: verifying the rc e2e-test-httpd-rc was created
-Aug 30 17:40:40.285: INFO: Waiting for rc e2e-test-httpd-rc to stabilize, generation 1 observed generation 0 spec.replicas 1 status.replicas 0
-Aug 30 17:40:40.291: INFO: Waiting for rc e2e-test-httpd-rc to stabilize, generation 1 observed generation 1 spec.replicas 1 status.replicas 0
-STEP: rolling-update to same image controller
-Aug 30 17:40:40.303: INFO: scanned /root for discovery docs: 
-Aug 30 17:40:40.303: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 rolling-update e2e-test-httpd-rc --update-period=1s --image=docker.io/library/httpd:2.4.38-alpine --image-pull-policy=IfNotPresent --namespace=kubectl-2938'
-Aug 30 17:40:56.151: INFO: stderr: "Command \"rolling-update\" is deprecated, use \"rollout\" instead\n"
-Aug 30 17:40:56.151: INFO: stdout: "Created e2e-test-httpd-rc-b0ac9a6d7e4559c2ed5f4aab59e66e91\nScaling up e2e-test-httpd-rc-b0ac9a6d7e4559c2ed5f4aab59e66e91 from 0 to 1, scaling down e2e-test-httpd-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-httpd-rc-b0ac9a6d7e4559c2ed5f4aab59e66e91 up to 1\nScaling e2e-test-httpd-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-httpd-rc\nRenaming e2e-test-httpd-rc-b0ac9a6d7e4559c2ed5f4aab59e66e91 to e2e-test-httpd-rc\nreplicationcontroller/e2e-test-httpd-rc rolling updated\n"
-Aug 30 17:40:56.151: INFO: stdout: "Created e2e-test-httpd-rc-b0ac9a6d7e4559c2ed5f4aab59e66e91\nScaling up e2e-test-httpd-rc-b0ac9a6d7e4559c2ed5f4aab59e66e91 from 0 to 1, scaling down e2e-test-httpd-rc from 1 to 0 (keep 1 pods available, don't exceed 2 pods)\nScaling e2e-test-httpd-rc-b0ac9a6d7e4559c2ed5f4aab59e66e91 up to 1\nScaling e2e-test-httpd-rc down to 0\nUpdate succeeded. Deleting old controller: e2e-test-httpd-rc\nRenaming e2e-test-httpd-rc-b0ac9a6d7e4559c2ed5f4aab59e66e91 to e2e-test-httpd-rc\nreplicationcontroller/e2e-test-httpd-rc rolling updated\n"
-STEP: waiting for all containers in run=e2e-test-httpd-rc pods to come up.
-Aug 30 17:40:56.151: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l run=e2e-test-httpd-rc --namespace=kubectl-2938'
-Aug 30 17:40:56.224: INFO: stderr: ""
-Aug 30 17:40:56.224: INFO: stdout: "e2e-test-httpd-rc-b0ac9a6d7e4559c2ed5f4aab59e66e91-dcp2d "
-Aug 30 17:40:56.224: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods e2e-test-httpd-rc-b0ac9a6d7e4559c2ed5f4aab59e66e91-dcp2d -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "e2e-test-httpd-rc") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=kubectl-2938'
-Aug 30 17:40:56.290: INFO: stderr: ""
-Aug 30 17:40:56.290: INFO: stdout: "true"
-Aug 30 17:40:56.290: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 get pods e2e-test-httpd-rc-b0ac9a6d7e4559c2ed5f4aab59e66e91-dcp2d -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "e2e-test-httpd-rc"}}{{.image}}{{end}}{{end}}{{end}} --namespace=kubectl-2938'
-Aug 30 17:40:56.352: INFO: stderr: ""
-Aug 30 17:40:56.352: INFO: stdout: "docker.io/library/httpd:2.4.38-alpine"
-Aug 30 17:40:56.352: INFO: e2e-test-httpd-rc-b0ac9a6d7e4559c2ed5f4aab59e66e91-dcp2d is verified up and running
-[AfterEach] Kubectl rolling-update
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1591
-Aug 30 17:40:56.352: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-962893509 delete rc e2e-test-httpd-rc --namespace=kubectl-2938'
-Aug 30 17:40:56.434: INFO: stderr: ""
-Aug 30 17:40:56.434: INFO: stdout: "replicationcontroller \"e2e-test-httpd-rc\" deleted\n"
+Feb 12 10:53:50.056: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 run e2e-test-httpd-deployment --image=docker.io/library/httpd:2.4.38-alpine --generator=deployment/apps.v1 --namespace=kubectl-5016'
+Feb 12 10:53:50.178: INFO: stderr: "kubectl run --generator=deployment/apps.v1 is DEPRECATED and will be removed in a future version. Use kubectl run --generator=run-pod/v1 or kubectl create instead.\n"
+Feb 12 10:53:50.178: INFO: stdout: "deployment.apps/e2e-test-httpd-deployment created\n"
+STEP: verifying the deployment e2e-test-httpd-deployment was created
+STEP: verifying the pod controlled by deployment e2e-test-httpd-deployment was created
+[AfterEach] Kubectl run deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1634
+Feb 12 10:53:54.198: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 delete deployment e2e-test-httpd-deployment --namespace=kubectl-5016'
+Feb 12 10:53:54.311: INFO: stderr: ""
+Feb 12 10:53:54.311: INFO: stdout: "deployment.apps \"e2e-test-httpd-deployment\" deleted\n"
 [AfterEach] [sig-cli] Kubectl client
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:40:56.434: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "kubectl-2938" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:53:54.311: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "kubectl-5016" for this suite.
+•{"msg":"PASSED [sig-cli] Kubectl client Kubectl run deployment should create a deployment from an image [Deprecated] [Conformance]","total":280,"completed":265,"skipped":4246,"failed":0}
+SSSSSSSSSS
+------------------------------
+[k8s.io] Pods 
+  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:53:54.332: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename pods
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in pods-3000
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:177
+[It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: creating the pod
+STEP: submitting the pod to kubernetes
+STEP: verifying the pod is in kubernetes
+STEP: updating the pod
+Feb 12 10:53:59.060: INFO: Successfully updated pod "pod-update-activedeadlineseconds-18c58d50-a0e3-48c8-9694-201dd92a53c1"
+Feb 12 10:53:59.060: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-18c58d50-a0e3-48c8-9694-201dd92a53c1" in namespace "pods-3000" to be "terminated due to deadline exceeded"
+Feb 12 10:53:59.066: INFO: Pod "pod-update-activedeadlineseconds-18c58d50-a0e3-48c8-9694-201dd92a53c1": Phase="Running", Reason="", readiness=true. Elapsed: 5.910261ms
+Feb 12 10:54:01.072: INFO: Pod "pod-update-activedeadlineseconds-18c58d50-a0e3-48c8-9694-201dd92a53c1": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 2.011949814s
+Feb 12 10:54:01.072: INFO: Pod "pod-update-activedeadlineseconds-18c58d50-a0e3-48c8-9694-201dd92a53c1" satisfied condition "terminated due to deadline exceeded"
+[AfterEach] [k8s.io] Pods
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:54:01.072: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "pods-3000" for this suite.
 
-• [SLOW TEST:16.440 seconds]
-[sig-cli] Kubectl client
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:23
-  Kubectl rolling-update
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:1580
-    should support rolling-update to same image  [Conformance]
-    /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+• [SLOW TEST:6.758 seconds]
+[k8s.io] Pods
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [k8s.io] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]","total":280,"completed":266,"skipped":4256,"failed":0}
+SSSSSSSSS
+------------------------------
+[k8s.io] Probing container 
+  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:54:01.090: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename container-probe
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-3077
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating pod liveness-72768890-f898-419f-ab83-a36783610b92 in namespace container-probe-3077
+Feb 12 10:54:03.282: INFO: Started pod liveness-72768890-f898-419f-ab83-a36783610b92 in namespace container-probe-3077
+STEP: checking the pod's current state and verifying that restartCount is present
+Feb 12 10:54:03.288: INFO: Initial restart count of pod liveness-72768890-f898-419f-ab83-a36783610b92 is 0
+Feb 12 10:54:19.341: INFO: Restart count of pod container-probe-3077/liveness-72768890-f898-419f-ab83-a36783610b92 is now 1 (16.053590215s elapsed)
+STEP: deleting the pod
+[AfterEach] [k8s.io] Probing container
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:54:19.364: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-3077" for this suite.
+
+• [SLOW TEST:18.291 seconds]
+[k8s.io] Probing container
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [k8s.io] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","total":280,"completed":267,"skipped":4265,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] ConfigMap 
+  should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:54:19.390: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename configmap
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in configmap-7778
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating configMap with name configmap-test-volume-0a5345a5-e56b-4631-b651-327edbd0b3d3
+STEP: Creating a pod to test consume configMaps
+Feb 12 10:54:19.577: INFO: Waiting up to 5m0s for pod "pod-configmaps-cc1535bb-452d-4d27-bf80-4a0fa3bfa6b1" in namespace "configmap-7778" to be "success or failure"
+Feb 12 10:54:19.583: INFO: Pod "pod-configmaps-cc1535bb-452d-4d27-bf80-4a0fa3bfa6b1": Phase="Pending", Reason="", readiness=false. Elapsed: 5.653462ms
+Feb 12 10:54:21.589: INFO: Pod "pod-configmaps-cc1535bb-452d-4d27-bf80-4a0fa3bfa6b1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011648584s
+Feb 12 10:54:23.594: INFO: Pod "pod-configmaps-cc1535bb-452d-4d27-bf80-4a0fa3bfa6b1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01752876s
+STEP: Saw pod success
+Feb 12 10:54:23.594: INFO: Pod "pod-configmaps-cc1535bb-452d-4d27-bf80-4a0fa3bfa6b1" satisfied condition "success or failure"
+Feb 12 10:54:23.600: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod pod-configmaps-cc1535bb-452d-4d27-bf80-4a0fa3bfa6b1 container configmap-volume-test: 
+STEP: delete the pod
+Feb 12 10:54:23.645: INFO: Waiting for pod pod-configmaps-cc1535bb-452d-4d27-bf80-4a0fa3bfa6b1 to disappear
+Feb 12 10:54:23.650: INFO: Pod pod-configmaps-cc1535bb-452d-4d27-bf80-4a0fa3bfa6b1 no longer exists
+[AfterEach] [sig-storage] ConfigMap
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:54:23.650: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "configmap-7778" for this suite.
+•{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume as non-root [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":268,"skipped":4339,"failed":0}
+S
+------------------------------
+[sig-apps] ReplicationController 
+  should surface a failure condition on a common issue like exceeded quota [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] ReplicationController
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:54:23.668: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename replication-controller
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in replication-controller-2751
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should surface a failure condition on a common issue like exceeded quota [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 10:54:23.827: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace
+STEP: Creating rc "condition-test" that asks for more than the allowed pod quota
+STEP: Checking rc "condition-test" has the desired failure condition set
+STEP: Scaling down rc "condition-test" to satisfy pod quota
+Feb 12 10:54:25.881: INFO: Updating replication controller "condition-test"
+STEP: Checking rc "condition-test" has no failure condition set
+[AfterEach] [sig-apps] ReplicationController
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:54:26.897: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "replication-controller-2751" for this suite.
+•{"msg":"PASSED [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance]","total":280,"completed":269,"skipped":4340,"failed":0}
+S
+------------------------------
+[sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] 
+  should perform rolling updates and roll backs of template modifications [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 10:54:26.916: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename statefulset
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in statefulset-6228
+STEP: Waiting for a default service account to be provisioned in namespace
+[BeforeEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:64
+[BeforeEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:79
+STEP: Creating service test in namespace statefulset-6228
+[It] should perform rolling updates and roll backs of template modifications [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a new StatefulSet
+Feb 12 10:54:27.109: INFO: Found 0 stateful pods, waiting for 3
+Feb 12 10:54:37.120: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true
+Feb 12 10:54:37.120: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true
+Feb 12 10:54:37.120: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true
+Feb 12 10:54:37.138: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6228 ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Feb 12 10:54:37.779: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Feb 12 10:54:37.779: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Feb 12 10:54:37.779: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+STEP: Updating StatefulSet template: update image from docker.io/library/httpd:2.4.38-alpine to docker.io/library/httpd:2.4.39-alpine
+Feb 12 10:54:47.826: INFO: Updating stateful set ss2
+STEP: Creating a new revision
+STEP: Updating Pods in reverse ordinal order
+Feb 12 10:54:57.858: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6228 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:54:58.433: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
+Feb 12 10:54:58.433: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
+Feb 12 10:54:58.433: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
+
+Feb 12 10:55:18.466: INFO: Waiting for StatefulSet statefulset-6228/ss2 to complete update
+Feb 12 10:55:18.466: INFO: Waiting for Pod statefulset-6228/ss2-0 to have revision ss2-84f9d6bf57 update revision ss2-65c7964b94
+STEP: Rolling back to a previous revision
+Feb 12 10:55:28.480: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6228 ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true'
+Feb 12 10:55:29.061: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n"
+Feb 12 10:55:29.061: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n"
+Feb 12 10:55:29.061: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'
+
+Feb 12 10:55:39.109: INFO: Updating stateful set ss2
+STEP: Rolling back update in reverse ordinal order
+Feb 12 10:55:49.139: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-008079477 exec --namespace=statefulset-6228 ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true'
+Feb 12 10:55:49.698: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n"
+Feb 12 10:55:49.698: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n"
+Feb 12 10:55:49.698: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'
+
+Feb 12 10:56:09.737: INFO: Waiting for StatefulSet statefulset-6228/ss2 to complete update
+Feb 12 10:56:09.737: INFO: Waiting for Pod statefulset-6228/ss2-0 to have revision ss2-65c7964b94 update revision ss2-84f9d6bf57
+[AfterEach] [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/statefulset.go:90
+Feb 12 10:56:19.749: INFO: Deleting all statefulset in ns statefulset-6228
+Feb 12 10:56:19.755: INFO: Scaling statefulset ss2 to 0
+Feb 12 10:57:43.397: INFO: Waiting for statefulset status.replicas updated to 0
+Feb 12 10:57:43.403: INFO: Deleting statefulset ss2
+[AfterEach] [sig-apps] StatefulSet
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:57:43.424: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "statefulset-6228" for this suite.
+
+• [SLOW TEST:197.365 seconds]
+[sig-apps] StatefulSet
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  [k8s.io] Basic StatefulSet functionality [StatefulSetBasic]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+    should perform rolling updates and roll backs of template modifications [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [sig-cli] Kubectl client Kubectl rolling-update should support rolling-update to same image  [Conformance]","total":280,"completed":272,"skipped":4372,"failed":0}
-SSSSS
+{"msg":"PASSED [sig-apps] StatefulSet [k8s.io] Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance]","total":280,"completed":270,"skipped":4341,"failed":0}
+SSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-api-machinery] ResourceQuota 
-  should create a ResourceQuota and capture the life of a secret. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Downward API volume 
+  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:40:56.457: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename resourcequota
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in resourcequota-7823
+Feb 12 10:57:43.446: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename downward-api
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-6998
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should create a ResourceQuota and capture the life of a secret. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Discovering how many secrets are in namespace by default
-STEP: Counting existing ResourceQuota
-STEP: Creating a ResourceQuota
-STEP: Ensuring resource quota status is calculated
-STEP: Creating a Secret
-STEP: Ensuring resource quota status captures secret creation
-STEP: Deleting a secret
-STEP: Ensuring resource quota status released usage
-[AfterEach] [sig-api-machinery] ResourceQuota
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:41:13.689: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "resourcequota-7823" for this suite.
-
-• [SLOW TEST:17.247 seconds]
-[sig-api-machinery] ResourceQuota
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
-  should create a ResourceQuota and capture the life of a secret. [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-------------------------------
-{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance]","total":280,"completed":273,"skipped":4377,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
+[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test downward API volume plugin
+Feb 12 10:57:43.629: INFO: Waiting up to 5m0s for pod "downwardapi-volume-5b4b3ff0-3929-425a-8253-b5ea91fc682c" in namespace "downward-api-6998" to be "success or failure"
+Feb 12 10:57:43.634: INFO: Pod "downwardapi-volume-5b4b3ff0-3929-425a-8253-b5ea91fc682c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.691075ms
+Feb 12 10:57:45.639: INFO: Pod "downwardapi-volume-5b4b3ff0-3929-425a-8253-b5ea91fc682c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010039225s
+STEP: Saw pod success
+Feb 12 10:57:45.639: INFO: Pod "downwardapi-volume-5b4b3ff0-3929-425a-8253-b5ea91fc682c" satisfied condition "success or failure"
+Feb 12 10:57:45.645: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod downwardapi-volume-5b4b3ff0-3929-425a-8253-b5ea91fc682c container client-container: 
+STEP: delete the pod
+Feb 12 10:57:45.718: INFO: Waiting for pod downwardapi-volume-5b4b3ff0-3929-425a-8253-b5ea91fc682c to disappear
+Feb 12 10:57:45.722: INFO: Pod downwardapi-volume-5b4b3ff0-3929-425a-8253-b5ea91fc682c no longer exists
+[AfterEach] [sig-storage] Downward API volume
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 10:57:45.722: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "downward-api-6998" for this suite.
+•{"msg":"PASSED [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":271,"skipped":4359,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
 [k8s.io] Probing container 
-  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:41:13.705: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 10:57:45.754: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename container-probe
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-9194
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in container-probe-6803
 STEP: Waiting for a default service account to be provisioned in namespace
 [BeforeEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
-[It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:51
+[It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating pod busybox-d14a48f3-7e75-4b9a-895a-9e13d4ee9544 in namespace container-probe-6803
+Feb 12 10:57:47.939: INFO: Started pod busybox-d14a48f3-7e75-4b9a-895a-9e13d4ee9544 in namespace container-probe-6803
+STEP: checking the pod's current state and verifying that restartCount is present
+Feb 12 10:57:47.945: INFO: Initial restart count of pod busybox-d14a48f3-7e75-4b9a-895a-9e13d4ee9544 is 0
+STEP: deleting the pod
 [AfterEach] [k8s.io] Probing container
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:42:13.890: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "container-probe-9194" for this suite.
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 11:01:48.874: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "container-probe-6803" for this suite.
 
-• [SLOW TEST:60.202 seconds]
+• [SLOW TEST:243.137 seconds]
 [k8s.io] Probing container
-/workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
-  with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:716
+  should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-{"msg":"PASSED [k8s.io] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]","total":280,"completed":274,"skipped":4408,"failed":0}
+{"msg":"PASSED [k8s.io] Probing container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","total":280,"completed":272,"skipped":4386,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-storage] Secrets 
+  optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 11:01:48.891: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename secrets
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in secrets-7588
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] optional updates should be reflected in volume [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating secret with name s-test-opt-del-1f86ad0b-232e-4275-9c54-06b54246539f
+STEP: Creating secret with name s-test-opt-upd-bf3c0156-8fc6-4eb9-8209-d13e0c59169d
+STEP: Creating the pod
+STEP: Deleting secret s-test-opt-del-1f86ad0b-232e-4275-9c54-06b54246539f
+STEP: Updating secret s-test-opt-upd-bf3c0156-8fc6-4eb9-8209-d13e0c59169d
+STEP: Creating secret with name s-test-opt-create-ced9435d-a566-459a-a987-3a43423259eb
+STEP: waiting to observe update in volume
+[AfterEach] [sig-storage] Secrets
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 11:01:53.591: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "secrets-7588" for this suite.
+•{"msg":"PASSED [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance]","total":280,"completed":273,"skipped":4436,"failed":0}
 SSS
 ------------------------------
 [sig-storage] EmptyDir volumes 
-  should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+  should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 [BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:42:13.907: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
+Feb 12 11:01:53.611: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
 STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-4579
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-2191
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0644 on node default medium
-Aug 30 17:42:14.086: INFO: Waiting up to 5m0s for pod "pod-82cea4cc-255e-44fe-8e6a-8ab320a026c7" in namespace "emptydir-4579" to be "success or failure"
-Aug 30 17:42:14.093: INFO: Pod "pod-82cea4cc-255e-44fe-8e6a-8ab320a026c7": Phase="Pending", Reason="", readiness=false. Elapsed: 6.689636ms
-Aug 30 17:42:16.099: INFO: Pod "pod-82cea4cc-255e-44fe-8e6a-8ab320a026c7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012748754s
-Aug 30 17:42:18.105: INFO: Pod "pod-82cea4cc-255e-44fe-8e6a-8ab320a026c7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018817136s
+[It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test emptydir 0666 on tmpfs
+Feb 12 11:01:53.782: INFO: Waiting up to 5m0s for pod "pod-8b150aa2-0bee-4a33-8eae-c4f2f606045d" in namespace "emptydir-2191" to be "success or failure"
+Feb 12 11:01:53.787: INFO: Pod "pod-8b150aa2-0bee-4a33-8eae-c4f2f606045d": Phase="Pending", Reason="", readiness=false. Elapsed: 5.13099ms
+Feb 12 11:01:55.795: INFO: Pod "pod-8b150aa2-0bee-4a33-8eae-c4f2f606045d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.013312682s
 STEP: Saw pod success
-Aug 30 17:42:18.105: INFO: Pod "pod-82cea4cc-255e-44fe-8e6a-8ab320a026c7" satisfied condition "success or failure"
-Aug 30 17:42:18.111: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-82cea4cc-255e-44fe-8e6a-8ab320a026c7 container test-container: 
+Feb 12 11:01:55.795: INFO: Pod "pod-8b150aa2-0bee-4a33-8eae-c4f2f606045d" satisfied condition "success or failure"
+Feb 12 11:01:55.801: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-r6snm pod pod-8b150aa2-0bee-4a33-8eae-c4f2f606045d container test-container: 
 STEP: delete the pod
-Aug 30 17:42:18.182: INFO: Waiting for pod pod-82cea4cc-255e-44fe-8e6a-8ab320a026c7 to disappear
-Aug 30 17:42:18.186: INFO: Pod pod-82cea4cc-255e-44fe-8e6a-8ab320a026c7 no longer exists
+Feb 12 11:01:55.878: INFO: Waiting for pod pod-8b150aa2-0bee-4a33-8eae-c4f2f606045d to disappear
+Feb 12 11:01:55.882: INFO: Pod pod-8b150aa2-0bee-4a33-8eae-c4f2f606045d no longer exists
 [AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:42:18.186: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-4579" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":275,"skipped":4411,"failed":0}
-SSSSSSSSSSSSSS
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 11:01:55.882: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "emptydir-2191" for this suite.
+•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":274,"skipped":4439,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
 ------------------------------
-[sig-storage] EmptyDir volumes 
-  should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-storage] Subpath Atomic writer volumes 
+  should support subpaths with configmap pod [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-storage] Subpath
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:42:18.203: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename emptydir
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in emptydir-8645
+Feb 12 11:01:55.910: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename subpath
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in subpath-859
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test emptydir 0644 on tmpfs
-Aug 30 17:42:18.376: INFO: Waiting up to 5m0s for pod "pod-af5693d2-8657-4599-b376-aa92cd67e162" in namespace "emptydir-8645" to be "success or failure"
-Aug 30 17:42:18.383: INFO: Pod "pod-af5693d2-8657-4599-b376-aa92cd67e162": Phase="Pending", Reason="", readiness=false. Elapsed: 6.364121ms
-Aug 30 17:42:20.390: INFO: Pod "pod-af5693d2-8657-4599-b376-aa92cd67e162": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013512984s
-Aug 30 17:42:22.397: INFO: Pod "pod-af5693d2-8657-4599-b376-aa92cd67e162": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020056611s
+[BeforeEach] Atomic writer volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:37
+STEP: Setting up data
+[It] should support subpaths with configmap pod [LinuxOnly] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating pod pod-subpath-test-configmap-qk9m
+STEP: Creating a pod to test atomic-volume-subpath
+Feb 12 11:01:56.098: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-qk9m" in namespace "subpath-859" to be "success or failure"
+Feb 12 11:01:56.103: INFO: Pod "pod-subpath-test-configmap-qk9m": Phase="Pending", Reason="", readiness=false. Elapsed: 4.882084ms
+Feb 12 11:01:58.111: INFO: Pod "pod-subpath-test-configmap-qk9m": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012690159s
+Feb 12 11:02:00.117: INFO: Pod "pod-subpath-test-configmap-qk9m": Phase="Running", Reason="", readiness=true. Elapsed: 4.019231486s
+Feb 12 11:02:02.127: INFO: Pod "pod-subpath-test-configmap-qk9m": Phase="Running", Reason="", readiness=true. Elapsed: 6.029009872s
+Feb 12 11:02:04.133: INFO: Pod "pod-subpath-test-configmap-qk9m": Phase="Running", Reason="", readiness=true. Elapsed: 8.0351824s
+Feb 12 11:02:06.145: INFO: Pod "pod-subpath-test-configmap-qk9m": Phase="Running", Reason="", readiness=true. Elapsed: 10.047523333s
+Feb 12 11:02:08.151: INFO: Pod "pod-subpath-test-configmap-qk9m": Phase="Running", Reason="", readiness=true. Elapsed: 12.05319778s
+Feb 12 11:02:10.157: INFO: Pod "pod-subpath-test-configmap-qk9m": Phase="Running", Reason="", readiness=true. Elapsed: 14.058830018s
+Feb 12 11:02:12.163: INFO: Pod "pod-subpath-test-configmap-qk9m": Phase="Running", Reason="", readiness=true. Elapsed: 16.064879179s
+Feb 12 11:02:14.169: INFO: Pod "pod-subpath-test-configmap-qk9m": Phase="Running", Reason="", readiness=true. Elapsed: 18.07126417s
+Feb 12 11:02:16.176: INFO: Pod "pod-subpath-test-configmap-qk9m": Phase="Running", Reason="", readiness=true. Elapsed: 20.077749669s
+Feb 12 11:02:18.183: INFO: Pod "pod-subpath-test-configmap-qk9m": Phase="Succeeded", Reason="", readiness=false. Elapsed: 22.085045902s
 STEP: Saw pod success
-Aug 30 17:42:22.397: INFO: Pod "pod-af5693d2-8657-4599-b376-aa92cd67e162" satisfied condition "success or failure"
-Aug 30 17:42:22.402: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod pod-af5693d2-8657-4599-b376-aa92cd67e162 container test-container: 
+Feb 12 11:02:18.183: INFO: Pod "pod-subpath-test-configmap-qk9m" satisfied condition "success or failure"
+Feb 12 11:02:18.194: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-r6snm pod pod-subpath-test-configmap-qk9m container test-container-subpath-configmap-qk9m: 
 STEP: delete the pod
-Aug 30 17:42:22.481: INFO: Waiting for pod pod-af5693d2-8657-4599-b376-aa92cd67e162 to disappear
-Aug 30 17:42:22.487: INFO: Pod pod-af5693d2-8657-4599-b376-aa92cd67e162 no longer exists
-[AfterEach] [sig-storage] EmptyDir volumes
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:42:22.487: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "emptydir-8645" for this suite.
-•{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":276,"skipped":4425,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+Feb 12 11:02:18.237: INFO: Waiting for pod pod-subpath-test-configmap-qk9m to disappear
+Feb 12 11:02:18.242: INFO: Pod pod-subpath-test-configmap-qk9m no longer exists
+STEP: Deleting pod pod-subpath-test-configmap-qk9m
+Feb 12 11:02:18.242: INFO: Deleting pod "pod-subpath-test-configmap-qk9m" in namespace "subpath-859"
+[AfterEach] [sig-storage] Subpath
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 11:02:18.249: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "subpath-859" for this suite.
+
+• [SLOW TEST:22.361 seconds]
+[sig-storage] Subpath
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/utils/framework.go:23
+  Atomic writer volumes
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/storage/subpath.go:33
+    should support subpaths with configmap pod [LinuxOnly] [Conformance]
+    /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Downward API volume 
-  should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [LinuxOnly] [Conformance]","total":280,"completed":275,"skipped":4470,"failed":0}
+SSSSSSS
+------------------------------
+[sig-api-machinery] Garbage collector 
+  should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:42:22.503: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-7290
+Feb 12 11:02:18.275: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-5706
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 17:42:22.690: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d3453ee5-0c22-4ef3-b80c-0b8eaf920918" in namespace "downward-api-7290" to be "success or failure"
-Aug 30 17:42:22.698: INFO: Pod "downwardapi-volume-d3453ee5-0c22-4ef3-b80c-0b8eaf920918": Phase="Pending", Reason="", readiness=false. Elapsed: 7.857731ms
-Aug 30 17:42:24.704: INFO: Pod "downwardapi-volume-d3453ee5-0c22-4ef3-b80c-0b8eaf920918": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.013330565s
-STEP: Saw pod success
-Aug 30 17:42:24.704: INFO: Pod "downwardapi-volume-d3453ee5-0c22-4ef3-b80c-0b8eaf920918" satisfied condition "success or failure"
-Aug 30 17:42:24.709: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-d3453ee5-0c22-4ef3-b80c-0b8eaf920918 container client-container: 
-STEP: delete the pod
-Aug 30 17:42:24.777: INFO: Waiting for pod downwardapi-volume-d3453ee5-0c22-4ef3-b80c-0b8eaf920918 to disappear
-Aug 30 17:42:24.781: INFO: Pod downwardapi-volume-d3453ee5-0c22-4ef3-b80c-0b8eaf920918 no longer exists
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:42:24.781: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-7290" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":277,"skipped":4475,"failed":0}
-SSSSSSSSSSSSSSS
+[It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: create the rc
+STEP: delete the rc
+STEP: wait for the rc to be deleted
+STEP: Gathering metrics
+W0212 11:02:24.507731      21 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Feb 12 11:02:24.507: INFO: For apiserver_request_total:
+For apiserver_request_latency_seconds:
+For apiserver_init_events_total:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
+
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 11:02:24.508: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-5706" for this suite.
+
+• [SLOW TEST:6.249 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[sig-storage] Downward API volume 
-  should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]","total":280,"completed":276,"skipped":4477,"failed":0}
+SSSS
+------------------------------
+[sig-apps] Deployment 
+  deployment should delete old replica sets [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:42:24.797: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename downward-api
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in downward-api-1069
+Feb 12 11:02:24.524: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename deployment
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in deployment-8129
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:40
-[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test downward API volume plugin
-Aug 30 17:42:24.966: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e4ad67ea-7193-4bcc-84d3-04a3a3142c4b" in namespace "downward-api-1069" to be "success or failure"
-Aug 30 17:42:24.971: INFO: Pod "downwardapi-volume-e4ad67ea-7193-4bcc-84d3-04a3a3142c4b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.932212ms
-Aug 30 17:42:26.977: INFO: Pod "downwardapi-volume-e4ad67ea-7193-4bcc-84d3-04a3a3142c4b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.010977872s
-STEP: Saw pod success
-Aug 30 17:42:26.977: INFO: Pod "downwardapi-volume-e4ad67ea-7193-4bcc-84d3-04a3a3142c4b" satisfied condition "success or failure"
-Aug 30 17:42:26.981: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod downwardapi-volume-e4ad67ea-7193-4bcc-84d3-04a3a3142c4b container client-container: 
-STEP: delete the pod
-Aug 30 17:42:27.052: INFO: Waiting for pod downwardapi-volume-e4ad67ea-7193-4bcc-84d3-04a3a3142c4b to disappear
-Aug 30 17:42:27.056: INFO: Pod downwardapi-volume-e4ad67ea-7193-4bcc-84d3-04a3a3142c4b no longer exists
-[AfterEach] [sig-storage] Downward API volume
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:42:27.056: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "downward-api-1069" for this suite.
-•{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","total":280,"completed":278,"skipped":4490,"failed":0}
+[BeforeEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:69
+[It] deployment should delete old replica sets [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+Feb 12 11:02:24.707: INFO: Pod name cleanup-pod: Found 0 pods out of 1
+Feb 12 11:02:29.717: INFO: Pod name cleanup-pod: Found 1 pods out of 1
+STEP: ensuring each pod is running
+Feb 12 11:02:29.717: INFO: Creating deployment test-cleanup-deployment
+STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up
+[AfterEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/deployment.go:63
+Feb 12 11:02:31.845: INFO: Deployment "test-cleanup-deployment":
+&Deployment{ObjectMeta:{test-cleanup-deployment  deployment-8129 /apis/apps/v1/namespaces/deployment-8129/deployments/test-cleanup-deployment a4e1fa59-2ad1-44a4-918d-ad39753fcd96 39998 1 2021-02-12 11:02:29 +0000 UTC   map[name:cleanup-pod] map[deployment.kubernetes.io/revision:1] [] []  []},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:cleanup-pod] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc005e5b4e8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2021-02-12 11:02:29 +0000 UTC,LastTransitionTime:2021-02-12 11:02:29 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-cleanup-deployment-55ffc6b7b6" has successfully progressed.,LastUpdateTime:2021-02-12 11:02:31 +0000 UTC,LastTransitionTime:2021-02-12 11:02:29 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},}
+
+Feb 12 11:02:31.852: INFO: New ReplicaSet "test-cleanup-deployment-55ffc6b7b6" of Deployment "test-cleanup-deployment":
+&ReplicaSet{ObjectMeta:{test-cleanup-deployment-55ffc6b7b6  deployment-8129 /apis/apps/v1/namespaces/deployment-8129/replicasets/test-cleanup-deployment-55ffc6b7b6 2946d02c-eb3f-4d6d-9ad0-77d2c7577a48 39985 1 2021-02-12 11:02:29 +0000 UTC   map[name:cleanup-pod pod-template-hash:55ffc6b7b6] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-cleanup-deployment a4e1fa59-2ad1-44a4-918d-ad39753fcd96 0xc003197e87 0xc003197e88}] []  []},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,pod-template-hash: 55ffc6b7b6,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{      0 0001-01-01 00:00:00 +0000 UTC   map[name:cleanup-pod pod-template-hash:55ffc6b7b6] map[] [] []  []} {[] [] [{agnhost gcr.io/kubernetes-e2e-test-images/agnhost:2.8 [] []  [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,} false false false}] [] Always 0xc003197ef8  ClusterFirst map[]     false false false  &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,} []   nil default-scheduler [] []   nil []    map[] []}},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},}
+Feb 12 11:02:31.859: INFO: Pod "test-cleanup-deployment-55ffc6b7b6-nzbb2" is available:
+&Pod{ObjectMeta:{test-cleanup-deployment-55ffc6b7b6-nzbb2 test-cleanup-deployment-55ffc6b7b6- deployment-8129 /api/v1/namespaces/deployment-8129/pods/test-cleanup-deployment-55ffc6b7b6-nzbb2 08516683-530d-40dc-929e-786c572d780f 39984 0 2021-02-12 11:02:29 +0000 UTC   map[name:cleanup-pod pod-template-hash:55ffc6b7b6] map[cni.projectcalico.org/podIP:172.25.1.25/32] [{apps/v1 ReplicaSet test-cleanup-deployment-55ffc6b7b6 2946d02c-eb3f-4d6d-9ad0-77d2c7577a48 0xc0030182c7 0xc0030182c8}] []  []},Spec:PodSpec{Volumes:[]Volume{Volume{Name:default-token-9mnfm,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:&SecretVolumeSource{SecretName:default-token-9mnfm,Items:[]KeyToPath{},DefaultMode:*420,Optional:nil,},NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:nil,StorageOS:nil,CSI:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:default-token-9mnfm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:nostalgic-fermat-cdc5d8777-xg74t,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:nil,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 11:02:29 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 11:02:31 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 11:02:31 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2021-02-12 11:02:29 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:138.68.97.197,PodIP:172.25.1.25,StartTime:2021-02-12 11:02:29 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2021-02-12 11:02:31 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:gcr.io/kubernetes-e2e-test-images/agnhost:2.8,ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/agnhost@sha256:daf5332100521b1256d0e3c56d697a238eaec3af48897ed9167cbadd426773b5,ContainerID:docker://f36e265836255ac56b2013abcebe7e05453dfaf54c56347a8b878057a358b358,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:172.25.1.25,},},EphemeralContainerStatuses:[]ContainerStatus{},},}
+[AfterEach] [sig-apps] Deployment
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 11:02:31.859: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "deployment-8129" for this suite.
+
+• [SLOW TEST:7.357 seconds]
+[sig-apps] Deployment
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:23
+  deployment should delete old replica sets [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-apps] Deployment deployment should delete old replica sets [Conformance]","total":280,"completed":277,"skipped":4481,"failed":0}
 SS
 ------------------------------
-[k8s.io] Docker Containers 
-  should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Docker Containers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+[sig-scheduling] SchedulerPredicates [Serial] 
+  validates resource limits of pods that are allowed to run  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:42:27.071: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename containers
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in containers-4906
+Feb 12 11:02:31.882: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename sched-pred
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in sched-pred-5938
 STEP: Waiting for a default service account to be provisioned in namespace
-[It] should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-STEP: Creating a pod to test override command
-Aug 30 17:42:27.242: INFO: Waiting up to 5m0s for pod "client-containers-5989e026-75f5-4f8e-8d77-73348f32dd78" in namespace "containers-4906" to be "success or failure"
-Aug 30 17:42:27.248: INFO: Pod "client-containers-5989e026-75f5-4f8e-8d77-73348f32dd78": Phase="Pending", Reason="", readiness=false. Elapsed: 5.553687ms
-Aug 30 17:42:29.254: INFO: Pod "client-containers-5989e026-75f5-4f8e-8d77-73348f32dd78": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.011449889s
-STEP: Saw pod success
-Aug 30 17:42:29.254: INFO: Pod "client-containers-5989e026-75f5-4f8e-8d77-73348f32dd78" satisfied condition "success or failure"
-Aug 30 17:42:29.258: INFO: Trying to get logs from node adoring-wozniak-54dcfd79fc-948mf pod client-containers-5989e026-75f5-4f8e-8d77-73348f32dd78 container test-container: 
-STEP: delete the pod
-Aug 30 17:42:29.329: INFO: Waiting for pod client-containers-5989e026-75f5-4f8e-8d77-73348f32dd78 to disappear
-Aug 30 17:42:29.334: INFO: Pod client-containers-5989e026-75f5-4f8e-8d77-73348f32dd78 no longer exists
-[AfterEach] [k8s.io] Docker Containers
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:42:29.334: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "containers-4906" for this suite.
-•{"msg":"PASSED [k8s.io] Docker Containers should be able to override the image's default command (docker entrypoint) [NodeConformance] [Conformance]","total":280,"completed":279,"skipped":4492,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS
+[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:86
+Feb 12 11:02:32.051: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready
+Feb 12 11:02:32.073: INFO: Waiting for terminating namespaces to be deleted...
+Feb 12 11:02:32.080: INFO: 
+Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-dbhjj before test
+Feb 12 11:02:32.116: INFO: canal-k48gh from kube-system started at 2021-02-12 09:48:16 +0000 UTC (2 container statuses recorded)
+Feb 12 11:02:32.116: INFO: 	Container calico-node ready: true, restart count 0
+Feb 12 11:02:32.116: INFO: 	Container kube-flannel ready: true, restart count 0
+Feb 12 11:02:32.116: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-vkql5 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded)
+Feb 12 11:02:32.116: INFO: 	Container sonobuoy-worker ready: false, restart count 7
+Feb 12 11:02:32.116: INFO: 	Container systemd-logs ready: true, restart count 0
+Feb 12 11:02:32.116: INFO: openvpn-client-78d595f58b-2pzkd from kube-system started at 2021-02-12 10:38:49 +0000 UTC (2 container statuses recorded)
+Feb 12 11:02:32.116: INFO: 	Container dnat-controller ready: true, restart count 0
+Feb 12 11:02:32.116: INFO: 	Container openvpn-client ready: true, restart count 1
+Feb 12 11:02:32.116: INFO: coredns-6449d494cc-vrdxh from kube-system started at 2021-02-12 10:38:49 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.116: INFO: 	Container coredns ready: true, restart count 0
+Feb 12 11:02:32.116: INFO: user-ssh-keys-agent-8lrpd from kube-system started at 2021-02-12 09:48:16 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.116: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
+Feb 12 11:02:32.116: INFO: kube-proxy-c88c9 from kube-system started at 2021-02-12 09:48:16 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.116: INFO: 	Container kube-proxy ready: true, restart count 0
+Feb 12 11:02:32.116: INFO: sonobuoy-e2e-job-690c418daa0f4403 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded)
+Feb 12 11:02:32.116: INFO: 	Container e2e ready: true, restart count 0
+Feb 12 11:02:32.116: INFO: 	Container sonobuoy-worker ready: true, restart count 0
+Feb 12 11:02:32.116: INFO: node-local-dns-k2khb from kube-system started at 2021-02-12 09:48:36 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.116: INFO: 	Container node-cache ready: true, restart count 0
+Feb 12 11:02:32.116: INFO: logrotate-4f9dq from kube-system started at 2021-02-12 09:48:36 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.116: INFO: 	Container logrotate ready: true, restart count 0
+Feb 12 11:02:32.116: INFO: 
+Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-r6snm before test
+Feb 12 11:02:32.210: INFO: coredns-6449d494cc-psw4s from kube-system started at 2021-02-12 09:48:21 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.210: INFO: 	Container coredns ready: true, restart count 0
+Feb 12 11:02:32.210: INFO: logrotate-7vwrm from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.210: INFO: 	Container logrotate ready: true, restart count 0
+Feb 12 11:02:32.210: INFO: kube-proxy-gfdtg from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.210: INFO: 	Container kube-proxy ready: true, restart count 0
+Feb 12 11:02:32.210: INFO: user-ssh-keys-agent-vxhth from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.210: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
+Feb 12 11:02:32.210: INFO: dashboard-metrics-scraper-59bfc65dc9-rhjf7 from kubernetes-dashboard started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.210: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
+Feb 12 11:02:32.210: INFO: node-local-dns-fkrhg from kube-system started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.210: INFO: 	Container node-cache ready: true, restart count 0
+Feb 12 11:02:32.210: INFO: canal-b9q6f from kube-system started at 2021-02-12 09:48:02 +0000 UTC (2 container statuses recorded)
+Feb 12 11:02:32.210: INFO: 	Container calico-node ready: true, restart count 0
+Feb 12 11:02:32.210: INFO: 	Container kube-flannel ready: true, restart count 0
+Feb 12 11:02:32.210: INFO: sonobuoy from sonobuoy started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.210: INFO: 	Container kube-sonobuoy ready: true, restart count 0
+Feb 12 11:02:32.210: INFO: dashboard-metrics-scraper-59bfc65dc9-d8l4b from kubernetes-dashboard started at 2021-02-12 09:48:12 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.210: INFO: 	Container dashboard-metrics-scraper ready: true, restart count 0
+Feb 12 11:02:32.210: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-zp8g7 from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded)
+Feb 12 11:02:32.210: INFO: 	Container sonobuoy-worker ready: false, restart count 7
+Feb 12 11:02:32.210: INFO: 	Container systemd-logs ready: true, restart count 0
+Feb 12 11:02:32.210: INFO: 
+Logging pods the kubelet thinks is on node nostalgic-fermat-cdc5d8777-xg74t before test
+Feb 12 11:02:32.273: INFO: kube-proxy-4zfp6 from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.274: INFO: 	Container kube-proxy ready: true, restart count 0
+Feb 12 11:02:32.274: INFO: canal-ww2cz from kube-system started at 2021-02-12 09:48:02 +0000 UTC (2 container statuses recorded)
+Feb 12 11:02:32.274: INFO: 	Container calico-node ready: true, restart count 0
+Feb 12 11:02:32.274: INFO: 	Container kube-flannel ready: true, restart count 0
+Feb 12 11:02:32.274: INFO: node-local-dns-75w47 from kube-system started at 2021-02-12 10:39:16 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.274: INFO: 	Container node-cache ready: true, restart count 0
+Feb 12 11:02:32.274: INFO: sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-wbbpd from sonobuoy started at 2021-02-12 09:48:44 +0000 UTC (2 container statuses recorded)
+Feb 12 11:02:32.274: INFO: 	Container sonobuoy-worker ready: false, restart count 7
+Feb 12 11:02:32.274: INFO: 	Container systemd-logs ready: true, restart count 0
+Feb 12 11:02:32.274: INFO: logrotate-9f95x from kube-system started at 2021-02-12 10:39:16 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.275: INFO: 	Container logrotate ready: true, restart count 0
+Feb 12 11:02:32.275: INFO: test-cleanup-deployment-55ffc6b7b6-nzbb2 from deployment-8129 started at 2021-02-12 11:02:29 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.275: INFO: 	Container agnhost ready: true, restart count 0
+Feb 12 11:02:32.275: INFO: user-ssh-keys-agent-kc6zq from kube-system started at 2021-02-12 09:48:02 +0000 UTC (1 container statuses recorded)
+Feb 12 11:02:32.275: INFO: 	Container user-ssh-keys-agent ready: true, restart count 0
+[It] validates resource limits of pods that are allowed to run  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: verifying the node has the label node nostalgic-fermat-cdc5d8777-dbhjj
+STEP: verifying the node has the label node nostalgic-fermat-cdc5d8777-r6snm
+STEP: verifying the node has the label node nostalgic-fermat-cdc5d8777-xg74t
+Feb 12 11:02:32.371: INFO: Pod test-cleanup-deployment-55ffc6b7b6-nzbb2 requesting resource cpu=0m on Node nostalgic-fermat-cdc5d8777-xg74t
+Feb 12 11:02:32.372: INFO: Pod canal-b9q6f requesting resource cpu=250m on Node nostalgic-fermat-cdc5d8777-r6snm
+Feb 12 11:02:32.372: INFO: Pod canal-k48gh requesting resource cpu=250m on Node nostalgic-fermat-cdc5d8777-dbhjj
+Feb 12 11:02:32.372: INFO: Pod canal-ww2cz requesting resource cpu=250m on Node nostalgic-fermat-cdc5d8777-xg74t
+Feb 12 11:02:32.372: INFO: Pod coredns-6449d494cc-psw4s requesting resource cpu=50m on Node nostalgic-fermat-cdc5d8777-r6snm
+Feb 12 11:02:32.372: INFO: Pod coredns-6449d494cc-vrdxh requesting resource cpu=50m on Node nostalgic-fermat-cdc5d8777-dbhjj
+Feb 12 11:02:32.372: INFO: Pod kube-proxy-4zfp6 requesting resource cpu=75m on Node nostalgic-fermat-cdc5d8777-xg74t
+Feb 12 11:02:32.372: INFO: Pod kube-proxy-c88c9 requesting resource cpu=75m on Node nostalgic-fermat-cdc5d8777-dbhjj
+Feb 12 11:02:32.372: INFO: Pod kube-proxy-gfdtg requesting resource cpu=75m on Node nostalgic-fermat-cdc5d8777-r6snm
+Feb 12 11:02:32.372: INFO: Pod logrotate-4f9dq requesting resource cpu=75m on Node nostalgic-fermat-cdc5d8777-dbhjj
+Feb 12 11:02:32.372: INFO: Pod logrotate-7vwrm requesting resource cpu=75m on Node nostalgic-fermat-cdc5d8777-r6snm
+Feb 12 11:02:32.373: INFO: Pod logrotate-9f95x requesting resource cpu=75m on Node nostalgic-fermat-cdc5d8777-xg74t
+Feb 12 11:02:32.373: INFO: Pod node-local-dns-75w47 requesting resource cpu=0m on Node nostalgic-fermat-cdc5d8777-xg74t
+Feb 12 11:02:32.373: INFO: Pod node-local-dns-fkrhg requesting resource cpu=0m on Node nostalgic-fermat-cdc5d8777-r6snm
+Feb 12 11:02:32.373: INFO: Pod node-local-dns-k2khb requesting resource cpu=0m on Node nostalgic-fermat-cdc5d8777-dbhjj
+Feb 12 11:02:32.373: INFO: Pod openvpn-client-78d595f58b-2pzkd requesting resource cpu=30m on Node nostalgic-fermat-cdc5d8777-dbhjj
+Feb 12 11:02:32.373: INFO: Pod user-ssh-keys-agent-8lrpd requesting resource cpu=0m on Node nostalgic-fermat-cdc5d8777-dbhjj
+Feb 12 11:02:32.373: INFO: Pod user-ssh-keys-agent-kc6zq requesting resource cpu=0m on Node nostalgic-fermat-cdc5d8777-xg74t
+Feb 12 11:02:32.373: INFO: Pod user-ssh-keys-agent-vxhth requesting resource cpu=0m on Node nostalgic-fermat-cdc5d8777-r6snm
+Feb 12 11:02:32.373: INFO: Pod dashboard-metrics-scraper-59bfc65dc9-d8l4b requesting resource cpu=50m on Node nostalgic-fermat-cdc5d8777-r6snm
+Feb 12 11:02:32.373: INFO: Pod dashboard-metrics-scraper-59bfc65dc9-rhjf7 requesting resource cpu=50m on Node nostalgic-fermat-cdc5d8777-r6snm
+Feb 12 11:02:32.373: INFO: Pod sonobuoy requesting resource cpu=0m on Node nostalgic-fermat-cdc5d8777-r6snm
+Feb 12 11:02:32.373: INFO: Pod sonobuoy-e2e-job-690c418daa0f4403 requesting resource cpu=0m on Node nostalgic-fermat-cdc5d8777-dbhjj
+Feb 12 11:02:32.373: INFO: Pod sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-vkql5 requesting resource cpu=0m on Node nostalgic-fermat-cdc5d8777-dbhjj
+Feb 12 11:02:32.374: INFO: Pod sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-wbbpd requesting resource cpu=0m on Node nostalgic-fermat-cdc5d8777-xg74t
+Feb 12 11:02:32.374: INFO: Pod sonobuoy-systemd-logs-daemon-set-0d839e2b4c4d4a58-zp8g7 requesting resource cpu=0m on Node nostalgic-fermat-cdc5d8777-r6snm
+STEP: Starting Pods to consume most of the cluster CPU.
+Feb 12 11:02:32.374: INFO: Creating a pod which consumes cpu=784m on Node nostalgic-fermat-cdc5d8777-dbhjj
+Feb 12 11:02:32.390: INFO: Creating a pod which consumes cpu=735m on Node nostalgic-fermat-cdc5d8777-r6snm
+Feb 12 11:02:32.400: INFO: Creating a pod which consumes cpu=840m on Node nostalgic-fermat-cdc5d8777-xg74t
+STEP: Creating another pod that requires unavailable amount of CPU.
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-2ed5400c-0b0a-46c1-aa22-be870d7c2448.1662fb24e23436d2], Reason = [Scheduled], Message = [Successfully assigned sched-pred-5938/filler-pod-2ed5400c-0b0a-46c1-aa22-be870d7c2448 to nostalgic-fermat-cdc5d8777-r6snm]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-2ed5400c-0b0a-46c1-aa22-be870d7c2448.1662fb251bd5b053], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-2ed5400c-0b0a-46c1-aa22-be870d7c2448.1662fb2520990eed], Reason = [Created], Message = [Created container filler-pod-2ed5400c-0b0a-46c1-aa22-be870d7c2448]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-2ed5400c-0b0a-46c1-aa22-be870d7c2448.1662fb2529e91917], Reason = [Started], Message = [Started container filler-pod-2ed5400c-0b0a-46c1-aa22-be870d7c2448]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-75da8bdf-8f67-466a-a4c4-063153b0781c.1662fb24e1afae88], Reason = [Scheduled], Message = [Successfully assigned sched-pred-5938/filler-pod-75da8bdf-8f67-466a-a4c4-063153b0781c to nostalgic-fermat-cdc5d8777-dbhjj]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-75da8bdf-8f67-466a-a4c4-063153b0781c.1662fb251e06fe80], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-75da8bdf-8f67-466a-a4c4-063153b0781c.1662fb2521c8871f], Reason = [Created], Message = [Created container filler-pod-75da8bdf-8f67-466a-a4c4-063153b0781c]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-75da8bdf-8f67-466a-a4c4-063153b0781c.1662fb252bada652], Reason = [Started], Message = [Started container filler-pod-75da8bdf-8f67-466a-a4c4-063153b0781c]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-f404f1f0-5f17-4759-8093-b70ef82871dc.1662fb24e336aa64], Reason = [Scheduled], Message = [Successfully assigned sched-pred-5938/filler-pod-f404f1f0-5f17-4759-8093-b70ef82871dc to nostalgic-fermat-cdc5d8777-xg74t]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-f404f1f0-5f17-4759-8093-b70ef82871dc.1662fb251f4292d3], Reason = [Pulled], Message = [Container image "k8s.gcr.io/pause:3.1" already present on machine]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-f404f1f0-5f17-4759-8093-b70ef82871dc.1662fb2523b75533], Reason = [Created], Message = [Created container filler-pod-f404f1f0-5f17-4759-8093-b70ef82871dc]
+STEP: Considering event: 
+Type = [Normal], Name = [filler-pod-f404f1f0-5f17-4759-8093-b70ef82871dc.1662fb252e0df3c5], Reason = [Started], Message = [Started container filler-pod-f404f1f0-5f17-4759-8093-b70ef82871dc]
+STEP: Considering event: 
+Type = [Warning], Name = [additional-pod.1662fb25d3feec90], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 Insufficient cpu.]
+STEP: removing the label node off the node nostalgic-fermat-cdc5d8777-dbhjj
+STEP: verifying the node doesn't have the label node
+STEP: removing the label node off the node nostalgic-fermat-cdc5d8777-r6snm
+STEP: verifying the node doesn't have the label node
+STEP: removing the label node off the node nostalgic-fermat-cdc5d8777-xg74t
+STEP: verifying the node doesn't have the label node
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 11:02:37.602: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "sched-pred-5938" for this suite.
+[AfterEach] [sig-scheduling] SchedulerPredicates [Serial]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/predicates.go:77
+
+• [SLOW TEST:5.765 seconds]
+[sig-scheduling] SchedulerPredicates [Serial]
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/scheduling/framework.go:40
+  validates resource limits of pods that are allowed to run  [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
 ------------------------------
-[k8s.io] Security Context When creating a pod with privileged 
-  should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-[BeforeEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run  [Conformance]","total":280,"completed":278,"skipped":4483,"failed":0}
+SSSSSSSSSSSSSSSSSSSS
+------------------------------
+[sig-api-machinery] Garbage collector 
+  should orphan pods created by rc if delete options say so [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [sig-api-machinery] Garbage collector
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
 STEP: Creating a kubernetes client
-Aug 30 17:42:29.351: INFO: >>> kubeConfig: /tmp/kubeconfig-962893509
-STEP: Building a namespace api object, basename security-context-test
-STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in security-context-test-90
+Feb 12 11:02:37.650: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename gc
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in gc-8473
 STEP: Waiting for a default service account to be provisioned in namespace
-[BeforeEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/common/security_context.go:39
-[It] should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
-Aug 30 17:42:29.531: INFO: Waiting up to 5m0s for pod "busybox-privileged-false-2893db3e-2fcf-416f-888a-6284d1c29a21" in namespace "security-context-test-90" to be "success or failure"
-Aug 30 17:42:29.539: INFO: Pod "busybox-privileged-false-2893db3e-2fcf-416f-888a-6284d1c29a21": Phase="Pending", Reason="", readiness=false. Elapsed: 7.692098ms
-Aug 30 17:42:31.545: INFO: Pod "busybox-privileged-false-2893db3e-2fcf-416f-888a-6284d1c29a21": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.01369878s
-Aug 30 17:42:31.545: INFO: Pod "busybox-privileged-false-2893db3e-2fcf-416f-888a-6284d1c29a21" satisfied condition "success or failure"
-Aug 30 17:42:31.599: INFO: Got logs for pod "busybox-privileged-false-2893db3e-2fcf-416f-888a-6284d1c29a21": "ip: RTNETLINK answers: Operation not permitted\n"
-[AfterEach] [k8s.io] Security Context
-  /workspace/anago-v1.17.9-rc.0.37+d1c2f63bd4fc89/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
-Aug 30 17:42:31.599: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
-STEP: Destroying namespace "security-context-test-90" for this suite.
-•{"msg":"PASSED [k8s.io] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]","total":280,"completed":280,"skipped":4524,"failed":0}
-SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSAug 30 17:42:31.620: INFO: Running AfterSuite actions on all nodes
-Aug 30 17:42:31.620: INFO: Running AfterSuite actions on node 1
-Aug 30 17:42:31.620: INFO: Skipping dumping logs from cluster
-{"msg":"Test Suite completed","total":280,"completed":280,"skipped":4563,"failed":0}
-
-Ran 280 of 4843 Specs in 4370.080 seconds
-SUCCESS! -- 280 Passed | 0 Failed | 0 Pending | 4563 Skipped
+[It] should orphan pods created by rc if delete options say so [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: create the rc
+STEP: delete the rc
+STEP: wait for the rc to be deleted
+STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods
+STEP: Gathering metrics
+W0212 11:03:17.914479      21 metrics_grabber.go:79] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled.
+Feb 12 11:03:17.914: INFO: For apiserver_request_total:
+For apiserver_request_latency_seconds:
+For apiserver_init_events_total:
+For garbage_collector_attempt_to_delete_queue_latency:
+For garbage_collector_attempt_to_delete_work_duration:
+For garbage_collector_attempt_to_orphan_queue_latency:
+For garbage_collector_attempt_to_orphan_work_duration:
+For garbage_collector_dirty_processing_latency_microseconds:
+For garbage_collector_event_processing_latency_microseconds:
+For garbage_collector_graph_changes_queue_latency:
+For garbage_collector_graph_changes_work_duration:
+For garbage_collector_orphan_processing_latency_microseconds:
+For namespace_queue_latency:
+For namespace_queue_latency_sum:
+For namespace_queue_latency_count:
+For namespace_retries:
+For namespace_work_duration:
+For namespace_work_duration_sum:
+For namespace_work_duration_count:
+For function_duration_seconds:
+For errors_total:
+For evicted_pods_total:
+
+[AfterEach] [sig-api-machinery] Garbage collector
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 11:03:17.914: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "gc-8473" for this suite.
+
+• [SLOW TEST:40.280 seconds]
+[sig-api-machinery] Garbage collector
+/workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/apimachinery/framework.go:23
+  should orphan pods created by rc if delete options say so [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+------------------------------
+{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance]","total":280,"completed":279,"skipped":4503,"failed":0}
+SSSSSSSSSSSSSSSSSSS
+------------------------------
+[k8s.io] Variable Expansion 
+  should allow substituting values in a container's command [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+[BeforeEach] [k8s.io] Variable Expansion
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:151
+STEP: Creating a kubernetes client
+Feb 12 11:03:17.930: INFO: >>> kubeConfig: /tmp/kubeconfig-008079477
+STEP: Building a namespace api object, basename var-expansion
+STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in var-expansion-4257
+STEP: Waiting for a default service account to be provisioned in namespace
+[It] should allow substituting values in a container's command [NodeConformance] [Conformance]
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:721
+STEP: Creating a pod to test substitution in container's command
+Feb 12 11:03:18.094: INFO: Waiting up to 5m0s for pod "var-expansion-bc15d0c3-f6e4-4a85-b88a-39061fef0739" in namespace "var-expansion-4257" to be "success or failure"
+Feb 12 11:03:18.098: INFO: Pod "var-expansion-bc15d0c3-f6e4-4a85-b88a-39061fef0739": Phase="Pending", Reason="", readiness=false. Elapsed: 4.604909ms
+Feb 12 11:03:20.107: INFO: Pod "var-expansion-bc15d0c3-f6e4-4a85-b88a-39061fef0739": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013169123s
+Feb 12 11:03:22.113: INFO: Pod "var-expansion-bc15d0c3-f6e4-4a85-b88a-39061fef0739": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018996475s
+STEP: Saw pod success
+Feb 12 11:03:22.113: INFO: Pod "var-expansion-bc15d0c3-f6e4-4a85-b88a-39061fef0739" satisfied condition "success or failure"
+Feb 12 11:03:22.120: INFO: Trying to get logs from node nostalgic-fermat-cdc5d8777-xg74t pod var-expansion-bc15d0c3-f6e4-4a85-b88a-39061fef0739 container dapi-container: 
+STEP: delete the pod
+Feb 12 11:03:22.154: INFO: Waiting for pod var-expansion-bc15d0c3-f6e4-4a85-b88a-39061fef0739 to disappear
+Feb 12 11:03:22.159: INFO: Pod var-expansion-bc15d0c3-f6e4-4a85-b88a-39061fef0739 no longer exists
+[AfterEach] [k8s.io] Variable Expansion
+  /workspace/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:152
+Feb 12 11:03:22.159: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready
+STEP: Destroying namespace "var-expansion-4257" for this suite.
+•{"msg":"PASSED [k8s.io] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance]","total":280,"completed":280,"skipped":4522,"failed":0}
+SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSFeb 12 11:03:22.178: INFO: Running AfterSuite actions on all nodes
+Feb 12 11:03:22.178: INFO: Running AfterSuite actions on node 1
+Feb 12 11:03:22.178: INFO: Skipping dumping logs from cluster
+{"msg":"Test Suite completed","total":280,"completed":280,"skipped":4566,"failed":0}
+
+Ran 280 of 4846 Specs in 4453.797 seconds
+SUCCESS! -- 280 Passed | 0 Failed | 0 Pending | 4566 Skipped
 PASS
 
-Ginkgo ran 1 suite in 1h12m51.174094841s
+Ginkgo ran 1 suite in 1h14m14.89944785s
 Test Suite Passed
diff --git a/v1.17/kubermatic/junit_01.xml b/v1.17/kubermatic/junit_01.xml
index 5238b4b3cf..abf53de864 100644
--- a/v1.17/kubermatic/junit_01.xml
+++ b/v1.17/kubermatic/junit_01.xml
@@ -1,13972 +1,13981 @@
 
-  
-      
+  
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+          
+      
+      
+          
+      
+      
+          
+      
+      
           
       
       
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
       
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
-      
+      
           
       
   
\ No newline at end of file