Console Output

Skipping 3,678 KB.. Full Log
    		I1219 18:08:59.669951    1170 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.403383ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53856]
    		I1219 18:08:59.670173    1170 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.091763ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53856]
    		I1219 18:08:59.670375    1170 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (3.324604ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53856]
    		I1219 18:08:59.671457    1170 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.84259ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53856]
    		I1219 18:08:59.671505    1170 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.899998ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53856]
    		I1219 18:08:59.671617    1170 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.499423ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53856]
    		I1219 18:08:59.672053    1170 request.go:485] Throttling request took 68.328223ms, request: POST:https://127.0.0.1:20567/api/v1/namespaces/openshift-node/secrets
    		I1219 18:08:59.682974    1170 wrap.go:42] POST /api/v1/namespaces/openshift-node/secrets: (10.701829ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54092]
    		I1219 18:08:59.701363    1170 wrap.go:42] PUT /api/v1/namespaces/openshift-node/serviceaccounts/default: (3.572232ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54092]
    		I1219 18:08:59.732669    1170 wrap.go:42] POST /api/v1/namespaces/integration/secrets: (2.601187ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54092]
    		I1219 18:08:59.764467    1170 wrap.go:42] PUT /api/v1/namespaces/integration/serviceaccounts/builder: (2.22919ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54092]
    		I1219 18:08:59.794221    1170 request.go:485] Throttling request took 60.79452ms, request: POST:https://127.0.0.1:20567/api/v1/namespaces/integration/secrets
    		I1219 18:08:59.798983    1170 wrap.go:42] POST /api/v1/namespaces/integration/secrets: (4.497ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54092]
    		I1219 18:08:59.829254    1170 wrap.go:42] PUT /api/v1/namespaces/integration/serviceaccounts/default: (2.463131ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54092]
    		I1219 18:08:59.858784    1170 request.go:485] Throttling request took 59.310322ms, request: POST:https://127.0.0.1:20567/api/v1/namespaces/integration/secrets
    		I1219 18:08:59.860693    1170 wrap.go:42] POST /api/v1/namespaces/integration/secrets: (1.624295ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54092]
    		I1219 18:08:59.893802    1170 wrap.go:42] PUT /api/v1/namespaces/integration/serviceaccounts/deployer: (2.622563ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54092]
    		I1219 18:09:00.346604    1170 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.675094ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53856]
    		I1219 18:09:00.503078    1170 wrap.go:42] GET /api/v1/namespaces/integration/serviceaccounts/default: (1.736385ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54088]
    		I1219 18:09:00.505354    1170 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-token-p7tlz: (1.619385ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54088]
    		I1219 18:09:00.507383    1170 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-dockercfg-ng9qs: (1.253005ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54088]
    		I1219 18:09:00.510588    1170 wrap.go:42] GET /api/v1/namespaces/integration/limitranges: (1.294484ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53856]
    		I1219 18:09:00.511038    1170 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I1219 18:09:00.511112    1170 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I1219 18:09:00.526580    1170 wrap.go:42] GET /api/v1/namespaces/integration: (1.657875ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53856]
    		I1219 18:09:00.526975    1170 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:00.527031    1170 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I1219 18:09:00.527044    1170 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:00.527059    1170 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:00.527081    1170 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:00.527093    1170 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:00.527109    1170 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I1219 18:09:00.527118    1170 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:00.527127    1170 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:00.527142    1170 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:00.527169    1170 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I1219 18:09:00.527180    1170 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:00.527189    1170 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:00.527211    1170 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I1219 18:09:00.527304    1170 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I1219 18:09:00.530280    1170 wrap.go:42] POST /api/v1/namespaces/integration/pods: (22.046689ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54088]
    		I1219 18:09:00.532361    1170 factory.go:1147] About to try and schedule pod testkmb94
    		I1219 18:09:00.532379    1170 scheduler.go:439] Attempting to schedule pod: integration/testkmb94
    		I1219 18:09:00.532406    1170 scheduler.go:191] Failed to schedule pod: integration/testkmb94
    		I1219 18:09:00.532475    1170 factory.go:1262] Unable to schedule integration testkmb94: no nodes are registered to the cluster; waiting
    		I1219 18:09:00.532521    1170 factory.go:1375] Updating pod condition for integration/testkmb94 to (PodScheduled==False)
    		I1219 18:09:00.532824    1170 pvc_protection_controller.go:276] Got event on pod integration/testkmb94
    		I1219 18:09:00.532872    1170 taint_manager.go:350] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"testkmb94"}
    		I1219 18:09:00.532930    1170 disruption.go:328] addPod called on pod "testkmb94"
    		I1219 18:09:00.532946    1170 disruption.go:403] No PodDisruptionBudgets found for pod testkmb94, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:00.532956    1170 disruption.go:331] No matching pdb for pod "testkmb94"
    		I1219 18:09:00.533032    1170 backoff_utils.go:79] Backing off 1s
    		I1219 18:09:00.543960    1170 wrap.go:42] PUT /api/v1/namespaces/integration/pods/testkmb94/status: (10.498662ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:53856]
    		I1219 18:09:00.545989    1170 pvc_protection_controller.go:276] Got event on pod integration/testkmb94
    		I1219 18:09:00.546041    1170 disruption.go:340] updatePod called on pod "testkmb94"
    		I1219 18:09:00.546055    1170 disruption.go:403] No PodDisruptionBudgets found for pod testkmb94, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:00.546064    1170 disruption.go:343] No matching pdb for pod "testkmb94"
    		I1219 18:09:00.546979    1170 factory.go:1147] About to try and schedule pod testkmb94
    		I1219 18:09:00.546993    1170 scheduler.go:439] Attempting to schedule pod: integration/testkmb94
    		I1219 18:09:00.547016    1170 scheduler.go:191] Failed to schedule pod: integration/testkmb94
    		I1219 18:09:00.547042    1170 factory.go:1262] Unable to schedule integration testkmb94: no nodes are registered to the cluster; waiting
    		I1219 18:09:00.547077    1170 factory.go:1375] Updating pod condition for integration/testkmb94 to (PodScheduled==False)
    		W1219 18:09:00.547132    1170 factory.go:1304] Request for pod integration/testkmb94 already in flight, abandoning
    		I1219 18:09:00.555958    1170 wrap.go:42] POST /api/v1/namespaces/integration/events: (21.066212ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:53856]
    		I1219 18:09:00.556290    1170 store.go:370] GuaranteedUpdate of /kubernetes.io/pods/integration/testkmb94 failed because of a conflict, going to retry
    		I1219 18:09:00.559478    1170 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testkmb94, uid a1e644e0-228a-11ea-bedd-0242ac110002, event type update
    		I1219 18:09:00.559960    1170 factory.go:1147] About to try and schedule pod testkmb94
    		I1219 18:09:00.559996    1170 scheduler.go:435] Skip schedule deleting pod: integration/testkmb94
    		I1219 18:09:00.560019    1170 pvc_protection_controller.go:276] Got event on pod integration/testkmb94
    		I1219 18:09:00.560074    1170 disruption.go:340] updatePod called on pod "testkmb94"
    		I1219 18:09:00.560090    1170 disruption.go:403] No PodDisruptionBudgets found for pod testkmb94, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:00.560098    1170 disruption.go:343] No matching pdb for pod "testkmb94"
    		I1219 18:09:00.560126    1170 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testkmb94, uid a1e644e0-228a-11ea-bedd-0242ac110002, event type update
    		I1219 18:09:00.575960    1170 wrap.go:42] DELETE /api/v1/namespaces/integration/pods/testkmb94: (39.808856ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54088]
    		I1219 18:09:00.577085    1170 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testkmb94, uid a1e644e0-228a-11ea-bedd-0242ac110002, event type delete
    		I1219 18:09:00.577153    1170 pvc_protection_controller.go:276] Got event on pod integration/testkmb94
    		I1219 18:09:00.577182    1170 deployment_controller.go:357] Pod testkmb94 deleted.
    		I1219 18:09:00.577210    1170 taint_manager.go:343] Noticed pod deletion: types.NamespacedName{Namespace:"integration", Name:"testkmb94"}
    		I1219 18:09:00.577257    1170 disruption.go:369] deletePod called on pod "testkmb94"
    		I1219 18:09:00.577272    1170 disruption.go:403] No PodDisruptionBudgets found for pod testkmb94, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:00.577281    1170 disruption.go:372] No matching pdb for pod "testkmb94"
    		I1219 18:09:00.577315    1170 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testkmb94, uid a1e644e0-228a-11ea-bedd-0242ac110002, event type delete
    		I1219 18:09:00.578194    1170 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I1219 18:09:00.578241    1170 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I1219 18:09:00.582856    1170 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (9.319745ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53856]
    		I1219 18:09:00.590027    1170 wrap.go:42] GET /api/v1/namespaces/integration: (10.520423ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53856]
    		I1219 18:09:00.590303    1170 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:00.590337    1170 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I1219 18:09:00.590347    1170 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:00.590357    1170 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:00.590377    1170 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:00.590388    1170 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:00.590404    1170 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I1219 18:09:00.590413    1170 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:00.590422    1170 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:00.590435    1170 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:00.590455    1170 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I1219 18:09:00.590464    1170 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:00.590472    1170 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:00.590493    1170 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I1219 18:09:00.590555    1170 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I1219 18:09:00.611249    1170 wrap.go:42] POST /api/v1/namespaces/integration/pods: (33.744112ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54088]
    		I1219 18:09:00.612552    1170 wrap.go:42] PATCH /api/v1/namespaces/integration/events/testkmb94.15e1d8179fc1c109: (54.789974ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:53856]
    		INFO: 2019/12/19 18:09:00 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc43e192060
    		INFO: 2019/12/19 18:09:00 dialing to target with scheme: ""
    		INFO: 2019/12/19 18:09:00 could not get resolver for scheme: ""
    		INFO: 2019/12/19 18:09:00 balancerWrapper: is pickfirst: false
    		INFO: 2019/12/19 18:09:00 balancerWrapper: got update addr from Notify: [{127.0.0.1:20104 <nil>}]
    		INFO: 2019/12/19 18:09:00 ccBalancerWrapper: new subconn: [{127.0.0.1:20104 0  <nil>}]
    		INFO: 2019/12/19 18:09:00 balancerWrapper: handle subconn state change: 0xc42f1c2710, CONNECTING
    		INFO: 2019/12/19 18:09:00 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc43e192060
    		I1219 18:09:00.615409    1170 factory.go:1147] About to try and schedule pod testhbsxl
    		I1219 18:09:00.615424    1170 scheduler.go:439] Attempting to schedule pod: integration/testhbsxl
    		I1219 18:09:00.615445    1170 scheduler.go:191] Failed to schedule pod: integration/testhbsxl
    		I1219 18:09:00.615465    1170 factory.go:1262] Unable to schedule integration testhbsxl: no nodes are registered to the cluster; waiting
    		I1219 18:09:00.615499    1170 factory.go:1375] Updating pod condition for integration/testhbsxl to (PodScheduled==False)
    		I1219 18:09:00.615887    1170 pvc_protection_controller.go:276] Got event on pod integration/testhbsxl
    		I1219 18:09:00.615926    1170 taint_manager.go:350] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"testhbsxl"}
    		I1219 18:09:00.615960    1170 disruption.go:328] addPod called on pod "testhbsxl"
    		I1219 18:09:00.615972    1170 disruption.go:403] No PodDisruptionBudgets found for pod testhbsxl, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:00.615980    1170 disruption.go:331] No matching pdb for pod "testhbsxl"
    		I1219 18:09:00.616031    1170 backoff_utils.go:79] Backing off 1s
    		I1219 18:09:00.630672    1170 wrap.go:42] PUT /api/v1/namespaces/integration/pods/testhbsxl/status: (13.36654ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:53856]
    		INFO: 2019/12/19 18:09:00 balancerWrapper: handle subconn state change: 0xc42f1c2710, READY
    		INFO: 2019/12/19 18:09:00 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc43e192060
    		
    --- PASS: TestIntegration/TestAllowedSCCViaRBAC (42.11s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		level: s0:c8,c2 for selinux options in namespace project1
    		I1219 18:09:06.200124    1183 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I1219 18:09:06.200135    1183 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I1219 18:09:06.200140    1183 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I1219 18:09:06.200146    1183 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I1219 18:09:06.200154    1183 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I1219 18:09:06.200176    1183 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I1219 18:09:06.200183    1183 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I1219 18:09:06.200188    1183 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I1219 18:09:06.200221    1183 admission.go:217] validating pod test3 (generate: ) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I1219 18:09:06.200357    1183 admission.go:170] pod test3 (generate: ) validated against provider hostaccess
    		I1219 18:09:06.202335    1183 wrap.go:42] POST /api/v1/namespaces/project1/pods: (5.533245ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43546]
    		I1219 18:09:06.203123    1183 pvc_protection_controller.go:276] Got event on pod project1/test3
    		I1219 18:09:06.203335    1183 factory.go:1147] About to try and schedule pod test3
    		I1219 18:09:06.203352    1183 scheduler.go:439] Attempting to schedule pod: project1/test3
    		I1219 18:09:06.203376    1183 scheduler.go:191] Failed to schedule pod: project1/test3
    		I1219 18:09:06.203434    1183 factory.go:1262] Unable to schedule project1 test3: no nodes are registered to the cluster; waiting
    		I1219 18:09:06.203476    1183 factory.go:1375] Updating pod condition for project1/test3 to (PodScheduled==False)
    		I1219 18:09:06.203801    1183 admission.go:97] getting security context constraints for pod test4 (generate: ) in namespace project2 with user info &{user1 a08903c3-228a-11ea-9929-0242ac110002 [system:authenticated:oauth system:authenticated] map[scopes.authorization.openshift.io:[]]}
    		I1219 18:09:06.203927    1183 disruption.go:328] addPod called on pod "test3"
    		I1219 18:09:06.203950    1183 disruption.go:403] No PodDisruptionBudgets found for pod test3, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:06.203959    1183 disruption.go:331] No matching pdb for pod "test3"
    		I1219 18:09:06.204006    1183 taint_manager.go:350] Noticed pod update: types.NamespacedName{Namespace:"project1", Name:"test3"}
    		I1219 18:09:06.204122    1183 backoff_utils.go:79] Backing off 1s
    		I1219 18:09:06.204634    1183 admission.go:108] getting security context constraints for pod test4 (generate: ) with service account info &{system:serviceaccount:project2:default  [system:serviceaccounts system:serviceaccounts:project2] map[]}
    		I1219 18:09:06.207425    1183 wrap.go:42] GET /api/v1/namespaces/project2: (1.669074ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43096]
    		I1219 18:09:06.207723    1183 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I1219 18:09:06.207743    1183 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I1219 18:09:06.207776    1183 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I1219 18:09:06.207798    1183 admission.go:217] validating pod test4 (generate: ) against providers restricted
    		I1219 18:09:06.207873    1183 admission.go:179] unable to validate pod test4 (generate: ) against any security context constraint: [provider restricted: .spec.securityContext.hostPID: Invalid value: true: Host PID is not allowed to be used spec.containers[0].securityContext.hostPID: Invalid value: true: Host PID is not allowed to be used]
    		I1219 18:09:06.208010    1183 wrap.go:42] POST /api/v1/namespaces/project2/pods: (4.824188ms) 403 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43546]
    		I1219 18:09:06.208581    1183 wrap.go:42] PUT /api/v1/namespaces/project1/pods/test3/status: (4.026722ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:43096]
    		I1219 18:09:06.209223    1183 pvc_protection_controller.go:276] Got event on pod project1/test3
    		I1219 18:09:06.209305    1183 disruption.go:340] updatePod called on pod "test3"
    		I1219 18:09:06.209321    1183 disruption.go:403] No PodDisruptionBudgets found for pod test3, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:06.209331    1183 disruption.go:343] No matching pdb for pod "test3"
    		I1219 18:09:06.209792    1183 factory.go:1147] About to try and schedule pod test3
    		I1219 18:09:06.209809    1183 scheduler.go:439] Attempting to schedule pod: project1/test3
    		I1219 18:09:06.209829    1183 scheduler.go:191] Failed to schedule pod: project1/test3
    		I1219 18:09:06.209822    1183 admission.go:97] getting security context constraints for pod test5 (generate: ) in namespace project1 with user info &{user2 a0c0f0ad-228a-11ea-9929-0242ac110002 [system:authenticated:oauth system:authenticated] map[scopes.authorization.openshift.io:[]]}
    		I1219 18:09:06.209856    1183 factory.go:1262] Unable to schedule project1 test3: no nodes are registered to the cluster; waiting
    		I1219 18:09:06.209888    1183 factory.go:1375] Updating pod condition for project1/test3 to (PodScheduled==False)
    		W1219 18:09:06.209924    1183 factory.go:1304] Request for pod project1/test3 already in flight, abandoning
    		I1219 18:09:06.210150    1183 admission.go:108] getting security context constraints for pod test5 (generate: ) with service account info &{system:serviceaccount:project1:default  [system:serviceaccounts system:serviceaccounts:project1] map[]}
    		I1219 18:09:06.213563    1183 wrap.go:42] POST /api/v1/namespaces/project1/events: (7.835549ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:43096]
    		I1219 18:09:06.214069    1183 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (3.382304ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43096]
    		I1219 18:09:06.217058    1183 wrap.go:42] GET /api/v1/namespaces/project1: (5.703589ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43096]
    		I1219 18:09:06.217463    1183 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (6.201344ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43096]
    		I1219 18:09:06.217666    1183 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (3.733344ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43096]
    		I1219 18:09:06.218210    1183 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I1219 18:09:06.218240    1183 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I1219 18:09:06.218251    1183 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I1219 18:09:06.218260    1183 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I1219 18:09:06.218278    1183 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I1219 18:09:06.218290    1183 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I1219 18:09:06.218308    1183 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I1219 18:09:06.218318    1183 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I1219 18:09:06.218334    1183 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I1219 18:09:06.218350    1183 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I1219 18:09:06.218372    1183 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I1219 18:09:06.218382    1183 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I1219 18:09:06.218391    1183 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I1219 18:09:06.218420    1183 admission.go:217] validating pod test5 (generate: ) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I1219 18:09:06.218652    1183 admission.go:170] pod test5 (generate: ) validated against provider hostaccess
    		I1219 18:09:06.221132    1183 wrap.go:42] POST /api/v1/namespaces/project1/pods: (12.556011ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43546]
    		I1219 18:09:06.221690    1183 factory.go:1147] About to try and schedule pod test5
    		I1219 18:09:06.221706    1183 scheduler.go:439] Attempting to schedule pod: project1/test5
    		I1219 18:09:06.221726    1183 scheduler.go:191] Failed to schedule pod: project1/test5
    		I1219 18:09:06.221823    1183 factory.go:1262] Unable to schedule project1 test5: no nodes are registered to the cluster; waiting
    		I1219 18:09:06.221859    1183 factory.go:1375] Updating pod condition for project1/test5 to (PodScheduled==False)
    		I1219 18:09:06.222182    1183 pvc_protection_controller.go:276] Got event on pod project1/test5
    		I1219 18:09:06.222223    1183 backoff_utils.go:79] Backing off 1s
    		I1219 18:09:06.222280    1183 disruption.go:328] addPod called on pod "test5"
    		I1219 18:09:06.222293    1183 disruption.go:403] No PodDisruptionBudgets found for pod test5, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:06.222302    1183 disruption.go:331] No matching pdb for pod "test5"
    		I1219 18:09:06.222337    1183 taint_manager.go:350] Noticed pod update: types.NamespacedName{Namespace:"project1", Name:"test5"}
    		I1219 18:09:06.222481    1183 admission.go:97] getting security context constraints for pod test6 (generate: ) in namespace project2 with user info &{user2 a0c0f0ad-228a-11ea-9929-0242ac110002 [system:authenticated:oauth system:authenticated] map[scopes.authorization.openshift.io:[]]}
    		I1219 18:09:06.222789    1183 admission.go:108] getting security context constraints for pod test6 (generate: ) with service account info &{system:serviceaccount:project2:default  [system:serviceaccounts system:serviceaccounts:project2] map[]}
    		I1219 18:09:06.225404    1183 wrap.go:42] GET /api/v1/namespaces/project2: (1.475366ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43096]
    		I1219 18:09:06.225886    1183 wrap.go:42] PATCH /api/v1/namespaces/project1/events/test3.15e1d818f1c5b968: (10.358406ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:43096]
    		I1219 18:09:06.226127    1183 wrap.go:42] PUT /api/v1/namespaces/project1/pods/test5/status: (3.549861ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:43096]
    		I1219 18:09:06.226550    1183 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I1219 18:09:06.226583    1183 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I1219 18:09:06.226595    1183 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I1219 18:09:06.226604    1183 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I1219 18:09:06.226624    1183 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I1219 18:09:06.226636    1183 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I1219 18:09:06.226652    1183 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I1219 18:09:06.226662    1183 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I1219 18:09:06.226671    1183 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I1219 18:09:06.226685    1183 matcher.go:342] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I1219 18:09:06.226706    1183 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I1219 18:09:06.226716    1183 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I1219 18:09:06.226725    1183 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I1219 18:09:06.226744    1183 admission.go:217] validating pod test6 (generate: ) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I1219 18:09:06.227008    1183 admission.go:170] pod test6 (generate: ) validated against provider hostaccess
    		I1219 18:09:06.227313    1183 pvc_protection_controller.go:276] Got event on pod project1/test5
    		I1219 18:09:06.227404    1183 disruption.go:340] updatePod called on pod "test5"
    		I1219 18:09:06.227420    1183 disruption.go:403] No PodDisruptionBudgets found for pod test5, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:06.227430    1183 disruption.go:343] No matching pdb for pod "test5"
    		I1219 18:09:06.228052    1183 factory.go:1147] About to try and schedule pod test5
    		I1219 18:09:06.228068    1183 scheduler.go:439] Attempting to schedule pod: project1/test5
    		I1219 18:09:06.228086    1183 scheduler.go:191] Failed to schedule pod: project1/test5
    		I1219 18:09:06.228104    1183 factory.go:1262] Unable to schedule project1 test5: no nodes are registered to the cluster; waiting
    		I1219 18:09:06.228150    1183 factory.go:1375] Updating pod condition for project1/test5 to (PodScheduled==False)
    		W1219 18:09:06.228190    1183 factory.go:1304] Request for pod project1/test5 already in flight, abandoning
    		I1219 18:09:06.229815    1183 wrap.go:42] POST /api/v1/namespaces/project2/pods: (7.934891ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43546]
    		I1219 18:09:06.230220    1183 pvc_protection_controller.go:276] Got event on pod project2/test6
    		I1219 18:09:06.230261    1183 disruption.go:328] addPod called on pod "test6"
    		I1219 18:09:06.230274    1183 disruption.go:403] No PodDisruptionBudgets found for pod test6, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:06.230283    1183 disruption.go:331] No matching pdb for pod "test6"
    		I1219 18:09:06.230309    1183 taint_manager.go:350] Noticed pod update: types.NamespacedName{Namespace:"project2", Name:"test6"}
    		I1219 18:09:06.230461    1183 factory.go:1147] About to try and schedule pod test6
    		I1219 18:09:06.230474    1183 scheduler.go:439] Attempting to schedule pod: project2/test6
    		I1219 18:09:06.230492    1183 scheduler.go:191] Failed to schedule pod: project2/test6
    		I1219 18:09:06.230508    1183 factory.go:1262] Unable to schedule project2 test6: no nodes are registered to the cluster; waiting
    		I1219 18:09:06.230539    1183 factory.go:1375] Updating pod condition for project2/test6 to (PodScheduled==False)
    		I1219 18:09:06.230813    1183 backoff_utils.go:79] Backing off 1s
    		I1219 18:09:06.234401    1183 wrap.go:42] POST /api/v1/namespaces/project1/events: (6.482137ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:43096]
    		I1219 18:09:06.234790    1183 wrap.go:42] PUT /api/v1/namespaces/project2/pods/test6/status: (3.78066ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:43096]
    		I1219 18:09:06.235614    1183 pvc_protection_controller.go:276] Got event on pod project2/test6
    		I1219 18:09:06.235669    1183 disruption.go:340] updatePod called on pod "test6"
    		I1219 18:09:06.235684    1183 disruption.go:403] No PodDisruptionBudgets found for pod test6, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:06.235693    1183 disruption.go:343] No matching pdb for pod "test6"
    		I1219 18:09:06.237081    1183 wrap.go:42] GET /api/v1/namespaces/project1: (3.368397ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43096]
    		I1219 18:09:06.237451    1183 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I1219 18:09:06.237728    1183 wrap.go:42] POST /apis/security.openshift.io/v1/namespaces/project1/podsecuritypolicyselfsubjectreviews: (5.855853ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43546]
    		I1219 18:09:06.239780    1183 factory.go:1147] About to try and schedule pod test6
    		I1219 18:09:06.239798    1183 scheduler.go:439] Attempting to schedule pod: project2/test6
    		I1219 18:09:06.239819    1183 scheduler.go:191] Failed to schedule pod: project2/test6
    		I1219 18:09:06.239838    1183 factory.go:1262] Unable to schedule project2 test6: no nodes are registered to the cluster; waiting
    		I1219 18:09:06.239870    1183 factory.go:1375] Updating pod condition for project2/test6 to (PodScheduled==False)
    		W1219 18:09:06.239912    1183 factory.go:1304] Request for pod project2/test6 already in flight, abandoning
    		I1219 18:09:06.242624    1183 wrap.go:42] GET /api/v1/namespaces/project2: (3.192925ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43096]
    		I1219 18:09:06.242958    1183 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I1219 18:09:06.243186    1183 wrap.go:42] POST /apis/security.openshift.io/v1/namespaces/project2/podsecuritypolicyselfsubjectreviews: (4.834211ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:43546]
    		I1219 18:09:06.242961    1183 wrap.go:42] PATCH /api/v1/namespaces/project1/events/test5.15e1d818f2de3cce: (7.107256ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:43096]
    		INFO: 2019/12/19 18:09:06 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc43e749d40
    		INFO: 2019/12/19 18:09:06 dialing to target with scheme: ""
    		INFO: 2019/12/19 18:09:06 could not get resolver for scheme: ""
    		INFO: 2019/12/19 18:09:06 balancerWrapper: is pickfirst: false
    		INFO: 2019/12/19 18:09:06 balancerWrapper: got update addr from Notify: [{127.0.0.1:14424 <nil>}]
    		INFO: 2019/12/19 18:09:06 ccBalancerWrapper: new subconn: [{127.0.0.1:14424 0  <nil>}]
    		INFO: 2019/12/19 18:09:06 balancerWrapper: handle subconn state change: 0xc42de0bfa0, CONNECTING
    		INFO: 2019/12/19 18:09:06 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc43e749d40
    		I1219 18:09:06.248096    1183 wrap.go:42] POST /api/v1/namespaces/project2/events: (3.876954ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:43096]
    		I1219 18:09:06.253661    1183 wrap.go:42] PATCH /api/v1/namespaces/project2/events/test6.15e1d818f362ad7f: (4.617335ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:43096]
    		INFO: 2019/12/19 18:09:06 balancerWrapper: handle subconn state change: 0xc42de0bfa0, READY
    		INFO: 2019/12/19 18:09:06 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc43e749d40
    		
    --- PASS: TestIntegration/TestAlwaysPullImagesOff (38.57s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		.0.0.1:35648]
    		I1219 18:09:28.368628    1206 leaderelection.go:199] successfully renewed lease kube-system/kube-scheduler
    		I1219 18:09:28.405095    1206 wrap.go:42] GET /api/v1/namespaces/kube-system/configmaps/kube-controller-manager: (1.701836ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:35648]
    		I1219 18:09:28.408783    1206 wrap.go:42] PUT /api/v1/namespaces/kube-system/configmaps/kube-controller-manager: (2.910631ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:35648]
    		I1219 18:09:28.410111    1206 leaderelection.go:199] successfully renewed lease kube-system/kube-controller-manager
    		I1219 18:09:28.446746    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.156056ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.458392    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.909296ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.459173    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.514959ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.459264    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.330623ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.505464    1206 wrap.go:42] GET /api/v1/namespaces/integration/serviceaccounts/default: (1.414289ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35956]
    		I1219 18:09:28.507380    1206 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-token-r4xdr: (1.358455ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35956]
    		I1219 18:09:28.509116    1206 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-dockercfg-xgb4f: (1.197462ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35956]
    		I1219 18:09:28.512358    1206 wrap.go:42] GET /api/v1/namespaces/integration/limitranges: (1.609518ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.512804    1206 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I1219 18:09:28.512876    1206 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I1219 18:09:28.514356    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (957.387µs) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.515012    1206 wrap.go:42] GET /api/v1/namespaces/integration: (1.059668ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.515193    1206 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:28.515263    1206 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I1219 18:09:28.515284    1206 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:28.515299    1206 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:28.515330    1206 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:28.515351    1206 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:28.515376    1206 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I1219 18:09:28.515393    1206 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:28.515403    1206 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:28.515424    1206 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:28.515446    1206 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I1219 18:09:28.515464    1206 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:28.515474    1206 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:28.515502    1206 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I1219 18:09:28.515603    1206 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I1219 18:09:28.517994    1206 wrap.go:42] POST /api/v1/namespaces/integration/pods: (8.142981ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35956]
    		I1219 18:09:28.518242    1206 factory.go:1147] About to try and schedule pod test28xd2
    		I1219 18:09:28.518258    1206 scheduler.go:439] Attempting to schedule pod: integration/test28xd2
    		I1219 18:09:28.518284    1206 scheduler.go:191] Failed to schedule pod: integration/test28xd2
    		I1219 18:09:28.518340    1206 factory.go:1262] Unable to schedule integration test28xd2: no nodes are registered to the cluster; waiting
    		I1219 18:09:28.518387    1206 factory.go:1375] Updating pod condition for integration/test28xd2 to (PodScheduled==False)
    		I1219 18:09:28.518519    1206 pvc_protection_controller.go:276] Got event on pod integration/test28xd2
    		I1219 18:09:28.518547    1206 disruption.go:328] addPod called on pod "test28xd2"
    		I1219 18:09:28.518564    1206 disruption.go:403] No PodDisruptionBudgets found for pod test28xd2, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:28.518574    1206 disruption.go:331] No matching pdb for pod "test28xd2"
    		I1219 18:09:28.518664    1206 taint_manager.go:350] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"test28xd2"}
    		I1219 18:09:28.518733    1206 backoff_utils.go:79] Backing off 1s
    		I1219 18:09:28.523288    1206 store.go:370] GuaranteedUpdate of /kubernetes.io/pods/integration/test28xd2 failed because of a conflict, going to retry
    		I1219 18:09:28.523332    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.442781ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.523680    1206 wrap.go:42] PUT /api/v1/namespaces/integration/pods/test28xd2/status: (4.796032ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:35648]
    		I1219 18:09:28.524481    1206 factory.go:1147] About to try and schedule pod test28xd2
    		I1219 18:09:28.524509    1206 scheduler.go:439] Attempting to schedule pod: integration/test28xd2
    		I1219 18:09:28.524530    1206 scheduler.go:191] Failed to schedule pod: integration/test28xd2
    		I1219 18:09:28.524573    1206 factory.go:1262] Unable to schedule integration test28xd2: no nodes are registered to the cluster; waiting
    		I1219 18:09:28.524617    1206 factory.go:1375] Updating pod condition for integration/test28xd2 to (PodScheduled==False)
    		I1219 18:09:28.524714    1206 pvc_protection_controller.go:276] Got event on pod integration/test28xd2
    		W1219 18:09:28.524871    1206 factory.go:1304] Request for pod integration/test28xd2 already in flight, abandoning
    		I1219 18:09:28.524906    1206 disruption.go:340] updatePod called on pod "test28xd2"
    		I1219 18:09:28.524922    1206 disruption.go:403] No PodDisruptionBudgets found for pod test28xd2, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:28.524931    1206 disruption.go:343] No matching pdb for pod "test28xd2"
    		I1219 18:09:28.526081    1206 wrap.go:42] POST /api/v1/namespaces/integration/events: (6.920845ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:35648]
    		I1219 18:09:28.526947    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.634586ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.527376    1206 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name test28xd2, uid b294f185-228a-11ea-93cd-0242ac110002, event type update
    		I1219 18:09:28.527525    1206 pvc_protection_controller.go:276] Got event on pod integration/test28xd2
    		I1219 18:09:28.527590    1206 factory.go:1147] About to try and schedule pod test28xd2
    		I1219 18:09:28.527623    1206 scheduler.go:435] Skip schedule deleting pod: integration/test28xd2
    		I1219 18:09:28.527703    1206 disruption.go:340] updatePod called on pod "test28xd2"
    		I1219 18:09:28.527724    1206 disruption.go:403] No PodDisruptionBudgets found for pod test28xd2, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:28.527735    1206 disruption.go:343] No matching pdb for pod "test28xd2"
    		I1219 18:09:28.527743    1206 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name test28xd2, uid b294f185-228a-11ea-93cd-0242ac110002, event type update
    		I1219 18:09:28.528557    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (3.516255ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.531251    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.807423ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.531453    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.111074ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.531645    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.659858ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.531877    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (3.329272ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.531959    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (3.479529ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.532070    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (3.074182ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.533320    1206 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name test28xd2, uid b294f185-228a-11ea-93cd-0242ac110002, event type delete
    		I1219 18:09:28.533732    1206 wrap.go:42] DELETE /api/v1/namespaces/integration/pods/test28xd2: (15.056278ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35956]
    		I1219 18:09:28.533914    1206 pvc_protection_controller.go:276] Got event on pod integration/test28xd2
    		I1219 18:09:28.534003    1206 disruption.go:369] deletePod called on pod "test28xd2"
    		I1219 18:09:28.534019    1206 disruption.go:403] No PodDisruptionBudgets found for pod test28xd2, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:28.534028    1206 disruption.go:372] No matching pdb for pod "test28xd2"
    		I1219 18:09:28.534059    1206 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name test28xd2, uid b294f185-228a-11ea-93cd-0242ac110002, event type delete
    		I1219 18:09:28.534133    1206 taint_manager.go:343] Noticed pod deletion: types.NamespacedName{Namespace:"integration", Name:"test28xd2"}
    		I1219 18:09:28.534188    1206 deployment_controller.go:357] Pod test28xd2 deleted.
    		I1219 18:09:28.534781    1206 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I1219 18:09:28.534815    1206 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I1219 18:09:28.536134    1206 wrap.go:42] PATCH /api/v1/namespaces/integration/events/test28xd2.15e1d81e23d82d95: (8.752019ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:35648]
    		I1219 18:09:28.537651    1206 wrap.go:42] GET /api/v1/namespaces/integration: (1.171834ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.538068    1206 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:28.538109    1206 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I1219 18:09:28.538134    1206 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:28.538145    1206 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:28.538175    1206 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:28.538197    1206 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:28.538217    1206 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I1219 18:09:28.538234    1206 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:28.538244    1206 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:28.538268    1206 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:28.538299    1206 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I1219 18:09:28.538319    1206 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I1219 18:09:28.538329    1206 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I1219 18:09:28.538358    1206 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I1219 18:09:28.538414    1206 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I1219 18:09:28.541512    1206 wrap.go:42] POST /api/v1/namespaces/integration/events: (4.430532ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:35648]
    		I1219 18:09:28.543597    1206 wrap.go:42] POST /api/v1/namespaces/integration/pods: (9.289027ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35956]
    		I1219 18:09:28.543676    1206 pvc_protection_controller.go:276] Got event on pod integration/test7vmmq
    		I1219 18:09:28.543766    1206 factory.go:1147] About to try and schedule pod test7vmmq
    		I1219 18:09:28.543787    1206 scheduler.go:439] Attempting to schedule pod: integration/test7vmmq
    		I1219 18:09:28.543809    1206 scheduler.go:191] Failed to schedule pod: integration/test7vmmq
    		I1219 18:09:28.543838    1206 factory.go:1262] Unable to schedule integration test7vmmq: no nodes are registered to the cluster; waiting
    		I1219 18:09:28.543881    1206 factory.go:1375] Updating pod condition for integration/test7vmmq to (PodScheduled==False)
    		I1219 18:09:28.543943    1206 disruption.go:328] addPod called on pod "test7vmmq"
    		I1219 18:09:28.543973    1206 disruption.go:403] No PodDisruptionBudgets found for pod test7vmmq, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:28.543983    1206 disruption.go:331] No matching pdb for pod "test7vmmq"
    		I1219 18:09:28.544065    1206 taint_manager.go:350] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"test7vmmq"}
    		I1219 18:09:28.544116    1206 backoff_utils.go:79] Backing off 1s
    		INFO: 2019/12/19 18:09:28 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc42beb13e0
    		INFO: 2019/12/19 18:09:28 dialing to target with scheme: ""
    		INFO: 2019/12/19 18:09:28 could not get resolver for scheme: ""
    		INFO: 2019/12/19 18:09:28 balancerWrapper: is pickfirst: false
    		INFO: 2019/12/19 18:09:28 balancerWrapper: got update addr from Notify: [{127.0.0.1:10755 <nil>}]
    		INFO: 2019/12/19 18:09:28 ccBalancerWrapper: new subconn: [{127.0.0.1:10755 0  <nil>}]
    		INFO: 2019/12/19 18:09:28 balancerWrapper: handle subconn state change: 0xc42d262660, CONNECTING
    		INFO: 2019/12/19 18:09:28 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc42beb13e0
    		I1219 18:09:28.547527    1206 wrap.go:42] PUT /api/v1/namespaces/integration/pods/test7vmmq/status: (3.309712ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:35648]
    		I1219 18:09:28.548167    1206 pvc_protection_controller.go:276] Got event on pod integration/test7vmmq
    		I1219 18:09:28.548238    1206 factory.go:1147] About to try and schedule pod test7vmmq
    		I1219 18:09:28.548252    1206 scheduler.go:439] Attempting to schedule pod: integration/test7vmmq
    		I1219 18:09:28.548272    1206 scheduler.go:191] Failed to schedule pod: integration/test7vmmq
    		I1219 18:09:28.548283    1206 disruption.go:340] updatePod called on pod "test7vmmq"
    		I1219 18:09:28.548291    1206 factory.go:1262] Unable to schedule integration test7vmmq: no nodes are registered to the cluster; waiting
    		I1219 18:09:28.548296    1206 disruption.go:403] No PodDisruptionBudgets found for pod test7vmmq, PodDisruptionBudget controller will avoid syncing.
    		I1219 18:09:28.548303    1206 disruption.go:343] No matching pdb for pod "test7vmmq"
    		I1219 18:09:28.548322    1206 factory.go:1375] Updating pod condition for integration/test7vmmq to (PodScheduled==False)
    		W1219 18:09:28.548370    1206 factory.go:1304] Request for pod integration/test7vmmq already in flight, abandoning
    		I1219 18:09:28.550502    1206 wrap.go:42] POST /api/v1/namespaces/integration/events: (6.15268ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:35648]
    		I1219 18:09:28.553722    1206 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.051255ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35648]
    		I1219 18:09:28.555892    1206 wrap.go:42] PATCH /api/v1/namespaces/integration/events/test7vmmq.15e1d81e255d3c07: (4.545516ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:35648]
    		INFO: 2019/12/19 18:09:28 balancerWrapper: handle subconn state change: 0xc42d262660, READY
    		INFO: 2019/12/19 18:09:28 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc42beb13e0
    		INFO: 2019/12/19 18:09:28 balancerWrapper: got update addr from Notify: [{127.0.0.1:10755 <nil>}]
    		
    --- PASS: TestIntegration/TestApiGroupPreferredVersions (30.29s)
    	runner_test.go:187: 
    			master_routes_test.go:382: Looking for build api group in server group discovery
    		
    		=== OUTPUT
    		2] PUT /api/v1/namespaces/default/serviceaccounts/builder: (15.746586ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.254393    1220 wrap.go:42] PUT /apis/security.openshift.io/v1/rangeallocations/scc-uid: (9.247654ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:namespace-security-allocation-controller] 127.0.0.1:46620]
    		I1219 18:09:29.258850    1220 wrap.go:42] PUT /api/v1/namespaces/kube-public/serviceaccounts/builder: (12.468499ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.258975    1220 shared_informer.go:123] caches populated
    		I1219 18:09:29.258989    1220 build_controller.go:243] Starting build controller
    		I1219 18:09:29.259021    1220 metrics.go:62] build metrics registered with prometheus
    		I1219 18:09:29.259530    1220 wrap.go:42] PUT /api/v1/namespaces/kube-system/serviceaccounts/deployer: (13.551068ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.259790    1220 wrap.go:42] PUT /api/v1/namespaces/default/serviceaccounts/deployer: (13.147158ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.260040    1220 wrap.go:42] PUT /api/v1/namespaces/kube-system/serviceaccounts/builder: (15.235696ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.260473    1220 controller.go:537] quota admission added evaluator for: { namespaces}
    		I1219 18:09:29.261370    1220 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.471853ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:46444]
    		I1219 18:09:29.261723    1220 wrap.go:42] GET /api/v1/namespaces/kube-public/serviceaccounts/deployer: (2.38987ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.263342    1220 wrap.go:42] PUT /api/v1/namespaces/default: (8.4161ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:namespace-security-allocation-controller] 127.0.0.1:46620]
    		I1219 18:09:29.263628    1220 serviceaccounts_controller.go:188] Finished syncing namespace "default" (8.102µs)
    		I1219 18:09:29.266867    1220 shared_informer.go:123] caches populated
    		I1219 18:09:29.266886    1220 controller_utils.go:1026] Caches are synced for HPA controller
    		I1219 18:09:29.267059    1220 shared_informer.go:123] caches populated
    		I1219 18:09:29.267454    1220 wrap.go:42] PUT /apis/security.openshift.io/v1/rangeallocations/scc-uid: (3.359128ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:namespace-security-allocation-controller] 127.0.0.1:46620]
    		I1219 18:09:29.270863    1220 wrap.go:42] POST /api/v1/namespaces/kube-public/secrets: (3.385391ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.272438    1220 wrap.go:42] GET /api/v1/namespaces/openshift/serviceaccounts/builder: (5.131181ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.272674    1220 wrap.go:42] GET /api/v1/namespaces/openshift-node/serviceaccounts/builder: (4.37001ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.272926    1220 wrap.go:42] GET /api/v1/namespaces/openshift/serviceaccounts/deployer: (4.214866ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.273189    1220 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/builder: (5.358385ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.275116    1220 wrap.go:42] PUT /api/v1/namespaces/kube-public: (7.110366ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:namespace-security-allocation-controller] 127.0.0.1:46620]
    		I1219 18:09:29.276519    1220 wrap.go:42] PUT /api/v1/namespaces/kube-public/serviceaccounts/deployer: (5.212319ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.276627    1220 serviceaccounts_controller.go:188] Finished syncing namespace "kube-public" (8.432µs)
    		I1219 18:09:29.279597    1220 shared_informer.go:123] caches populated
    		I1219 18:09:29.279884    1220 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/deployer: (1.939134ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.281364    1220 wrap.go:42] PUT /apis/security.openshift.io/v1/rangeallocations/scc-uid: (4.099193ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:namespace-security-allocation-controller] 127.0.0.1:46620]
    		I1219 18:09:29.283116    1220 wrap.go:42] POST /api/v1/namespaces/openshift/secrets: (3.723146ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.291046    1220 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (6.304976ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:46444]
    		I1219 18:09:29.294027    1220 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (4.133687ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.295947    1220 wrap.go:42] PUT /api/v1/namespaces/kube-system: (13.884486ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:namespace-security-allocation-controller] 127.0.0.1:46620]
    		I1219 18:09:29.296071    1220 wrap.go:42] PUT /api/v1/namespaces/openshift/serviceaccounts/deployer: (12.402484ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.297053    1220 serviceaccounts_controller.go:188] Finished syncing namespace "kube-system" (7.869µs)
    		I1219 18:09:29.297141    1220 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (5.489093ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.297438    1220 wrap.go:42] POST /api/v1/namespaces/openshift/secrets: (6.619243ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.297500    1220 wrap.go:42] POST /api/v1/namespaces/openshift-node/secrets: (6.195118ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.301637    1220 wrap.go:42] GET /api/v1/namespaces/openshift-node/serviceaccounts/deployer: (1.400877ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.303437    1220 wrap.go:42] PUT /apis/security.openshift.io/v1/rangeallocations/scc-uid: (6.613314ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:namespace-security-allocation-controller] 127.0.0.1:46620]
    		I1219 18:09:29.308782    1220 wrap.go:42] PUT /api/v1/namespaces/openshift/serviceaccounts/builder: (10.351029ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.309079    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/deployer: (11.094626ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.309429    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-node/serviceaccounts/builder: (10.909505ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.309600    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/builder: (14.972866ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.308786    1220 shared_informer.go:123] caches populated
    		I1219 18:09:29.309640    1220 create_dockercfg_secrets.go:220] caches synced
    		I1219 18:09:29.316798    1220 wrap.go:42] POST /api/v1/namespaces/openshift-node/secrets: (7.648682ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.317948    1220 wrap.go:42] PUT /api/v1/namespaces/openshift: (13.973805ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:namespace-security-allocation-controller] 127.0.0.1:46620]
    		I1219 18:09:29.319221    1220 serviceaccounts_controller.go:188] Finished syncing namespace "openshift" (5.662µs)
    		I1219 18:09:29.321319    1220 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/admin: (6.794328ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:46634]
    		I1219 18:09:29.321473    1220 wrap.go:42] PUT /api/v1/namespaces/kube-system/serviceaccounts/clusterrole-aggregation-controller: (10.150631ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.322382    1220 create_dockercfg_secrets.go:444] Creating token secret "clusterrole-aggregation-controller-token-fv5t6" for service account kube-system/clusterrole-aggregation-controller
    		I1219 18:09:29.324368    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/cluster-quota-reconciliation-controller: (11.808661ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.324624    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/default-rolebindings-controller: (11.727277ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.324823    1220 wrap.go:42] PUT /apis/security.openshift.io/v1/rangeallocations/scc-uid: (5.915912ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:namespace-security-allocation-controller] 127.0.0.1:46620]
    		I1219 18:09:29.325300    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/build-config-change-controller: (13.407817ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.325535    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/build-controller: (13.346675ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.325686    1220 create_dockercfg_secrets.go:444] Creating token secret "cluster-quota-reconciliation-controller-token-hqxdv" for service account openshift-infra/cluster-quota-reconciliation-controller
    		I1219 18:09:29.325821    1220 create_dockercfg_secrets.go:444] Creating token secret "default-rolebindings-controller-token-6zgtl" for service account openshift-infra/default-rolebindings-controller
    		I1219 18:09:29.326090    1220 create_dockercfg_secrets.go:444] Creating token secret "build-controller-token-gn9j5" for service account openshift-infra/build-controller
    		I1219 18:09:29.326284    1220 create_dockercfg_secrets.go:444] Creating token secret "build-config-change-controller-token-ss8cx" for service account openshift-infra/build-config-change-controller
    		I1219 18:09:29.326790    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-node/serviceaccounts/deployer: (9.393724ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.329738    1220 wrap.go:42] POST /api/v1/namespaces/kube-system/secrets: (7.030106ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.330676    1220 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/edit: (8.164498ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:46634]
    		I1219 18:09:29.332104    1220 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (5.480073ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.332127    1220 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (5.144103ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.332329    1220 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (4.75538ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.332438    1220 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (5.315923ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.335357    1220 wrap.go:42] GET /api/v1/namespaces/kube-system/secrets/clusterrole-aggregation-controller-token-fv5t6: (5.358316ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.336573    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-infra: (10.539258ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:namespace-security-allocation-controller] 127.0.0.1:46620]
    		I1219 18:09:29.336797    1220 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/default-rolebindings-controller-token-6zgtl: (2.859052ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.337306    1220 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/view: (5.46237ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:46634]
    		I1219 18:09:29.337673    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/deployer-controller: (6.772833ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.338632    1220 serviceaccounts_controller.go:188] Finished syncing namespace "openshift-infra" (7.338µs)
    		I1219 18:09:29.338829    1220 create_dockercfg_secrets.go:444] Creating token secret "deployer-controller-token-b529j" for service account openshift-infra/deployer-controller
    		I1219 18:09:29.340465    1220 wrap.go:42] GET /api?timeout=32s: (575.735µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:46634]
    		I1219 18:09:29.341875    1220 wrap.go:42] GET /apis?timeout=32s: (864.257µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:46634]
    		INFO: 2019/12/19 18:09:29 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc42a2a7440
    		INFO: 2019/12/19 18:09:29 dialing to target with scheme: ""
    		INFO: 2019/12/19 18:09:29 could not get resolver for scheme: ""
    		INFO: 2019/12/19 18:09:29 balancerWrapper: is pickfirst: false
    		INFO: 2019/12/19 18:09:29 balancerWrapper: got update addr from Notify: [{127.0.0.1:13160 <nil>}]
    		INFO: 2019/12/19 18:09:29 ccBalancerWrapper: new subconn: [{127.0.0.1:13160 0  <nil>}]
    		INFO: 2019/12/19 18:09:29 balancerWrapper: handle subconn state change: 0xc42ad503c0, CONNECTING
    		INFO: 2019/12/19 18:09:29 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc42a2a7440
    		I1219 18:09:29.348144    1220 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (9.0486ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.348379    1220 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/cluster-quota-reconciliation-controller-token-hqxdv: (12.938147ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.348986    1220 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/build-config-change-controller-token-ss8cx: (12.373674ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.349229    1220 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/build-controller-token-gn9j5: (13.167848ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.354570    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/deploymentconfig-controller: (20.285661ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		I1219 18:09:29.355824    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/default-rolebindings-controller-token-6zgtl: (12.755872ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.355934    1220 wrap.go:42] PUT /api/v1/namespaces/kube-system/secrets/clusterrole-aggregation-controller-token-fv5t6: (10.4309ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:46444]
    		I1219 18:09:29.356977    1220 create_dockercfg_secrets.go:444] Creating token secret "deploymentconfig-controller-token-446b6" for service account openshift-infra/deploymentconfig-controller
    		I1219 18:09:29.357641    1220 wrap.go:42] PUT /apis/security.openshift.io/v1/rangeallocations/scc-uid: (19.936329ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:namespace-security-allocation-controller] 127.0.0.1:46620]
    		I1219 18:09:29.358075    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/horizontal-pod-autoscaler: (23.35948ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		INFO: 2019/12/19 18:09:29 balancerWrapper: handle subconn state change: 0xc42ad503c0, READY
    		INFO: 2019/12/19 18:09:29 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc42a2a7440
    		I1219 18:09:29.358498    1220 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/image-trigger-controller: (23.144164ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:46620]
    		
    --- PASS: TestIntegration/TestAggregator (31.60s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		o:444] Creating token secret "horizontal-pod-autoscaler-token-924c8" for service account openshift-infra/horizontal-pod-autoscaler
    		I1219 18:09:32.160084    1229 wrap.go:42] PUT /api/v1/namespaces/kube-system/secrets/statefulset-controller-token-5x5vt: (9.045527ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.160631    1229 wrap.go:42] POST /apis/rbac.authorization.k8s.io/v1/namespaces/openshift-node/rolebindings: (116.793646ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:default-rolebindings-controller] 127.0.0.1:45016]
    		I1219 18:09:32.160966    1229 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/image-import-controller: (6.44188ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45016]
    		I1219 18:09:32.161362    1229 create_dockercfg_secrets.go:444] Creating token secret "image-import-controller-token-28b48" for service account openshift-infra/image-import-controller
    		I1219 18:09:32.163227    1229 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/build-config-change-controller-token-j4h27: (6.203236ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.163925    1229 wrap.go:42] PUT /api/v1/namespaces/kube-system/secrets/resourcequota-controller-token-c252v: (7.82739ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.164007    1229 wrap.go:42] PUT /api/v1/namespaces/kube-system/secrets/service-controller-token-5kcwd: (12.216995ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.165589    1229 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/default-token-6ckmn: (3.961093ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.165675    1229 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/build-controller-token-dlr46: (7.249686ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.168521    1229 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/deployer-controller-token-h5l69: (2.842723ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.174658    1229 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/build-config-change-controller-token-j4h27: (5.291883ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.178434    1229 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/build-controller-token-dlr46: (5.867087ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.179832    1229 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/deployer-controller-token-h5l69: (4.8079ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.181529    1229 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.714979ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.181542    1229 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/default-rolebindings-controller-token-78qx7: (3.826566ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.181707    1229 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/cluster-quota-reconciliation-controller-token-p4cmk: (3.637098ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.181975    1229 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/deploymentconfig-controller-token-kvvln: (1.54981ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.189006    1229 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/default-token-6ckmn: (12.511043ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.192432    1229 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/image-trigger-controller: (6.022103ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45016]
    		I1219 18:09:32.192645    1229 create_dockercfg_secrets.go:444] Creating token secret "image-trigger-controller-token-69xfk" for service account openshift-infra/image-trigger-controller
    		I1219 18:09:32.195405    1229 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/cluster-quota-reconciliation-controller-token-p4cmk: (8.462763ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.195714    1229 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/deploymentconfig-controller-token-kvvln: (6.704532ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.195959    1229 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/default-rolebindings-controller-token-78qx7: (7.982171ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		I1219 18:09:32.200073    1229 wrap.go:42] POST /apis/rbac.authorization.k8s.io/v1/namespaces/openshift-node/rolebindings: (38.694229ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:default-rolebindings-controller] 127.0.0.1:45016]
    		I1219 18:09:32.204234    1229 wrap.go:42] GET /api?timeout=32s: (378.867µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.204949    1229 wrap.go:42] GET /apis?timeout=32s: (341.901µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.205932    1229 wrap.go:42] GET /api/v1?timeout=32s: (539.077µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.206819    1229 wrap.go:42] GET /apis/apiregistration.k8s.io/v1?timeout=32s: (332.009µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.207589    1229 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1?timeout=32s: (348.1µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.208321    1229 wrap.go:42] GET /apis/extensions/v1beta1?timeout=32s: (299.535µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.208997    1229 wrap.go:42] GET /apis/apps/v1?timeout=32s: (282.123µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.209699    1229 wrap.go:42] GET /apis/apps/v1beta2?timeout=32s: (287.553µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.210446    1229 wrap.go:42] GET /apis/apps/v1beta1?timeout=32s: (330.609µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.211138    1229 wrap.go:42] GET /apis/events.k8s.io/v1beta1?timeout=32s: (223.17µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.211730    1229 wrap.go:42] GET /apis/authentication.k8s.io/v1?timeout=32s: (222.867µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.212395    1229 wrap.go:42] GET /apis/authentication.k8s.io/v1beta1?timeout=32s: (291.509µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.213057    1229 wrap.go:42] GET /apis/authorization.k8s.io/v1?timeout=32s: (227.472µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.213650    1229 wrap.go:42] GET /apis/authorization.k8s.io/v1beta1?timeout=32s: (221.434µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.214388    1229 wrap.go:42] GET /apis/autoscaling/v1?timeout=32s: (336.834µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.215178    1229 wrap.go:42] GET /apis/autoscaling/v2beta1?timeout=32s: (350.064µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.215190    1229 wrap.go:42] POST /apis/rbac.authorization.k8s.io/v1/namespaces/openshift-node/rolebindings: (14.740259ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:default-rolebindings-controller] 127.0.0.1:45016]
    		I1219 18:09:32.215695    1229 request.go:485] Throttling request took 61.289604ms, request: PUT:https://127.0.0.1:10280/api/v1/namespaces/openshift-infra/serviceaccounts/ingress-to-route-controller
    		I1219 18:09:32.215927    1229 wrap.go:42] GET /apis/batch/v1?timeout=32s: (286.551µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.216668    1229 wrap.go:42] GET /apis/batch/v1beta1?timeout=32s: (324.011µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.217424    1229 wrap.go:42] GET /apis/certificates.k8s.io/v1beta1?timeout=32s: (290.432µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.218184    1229 wrap.go:42] GET /apis/networking.k8s.io/v1?timeout=32s: (336.239µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.218519    1229 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/ingress-to-route-controller: (2.568973ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45016]
    		I1219 18:09:32.218774    1229 create_dockercfg_secrets.go:444] Creating token secret "ingress-to-route-controller-token-6jcwp" for service account openshift-infra/ingress-to-route-controller
    		I1219 18:09:32.219208    1229 wrap.go:42] GET /apis/policy/v1beta1?timeout=32s: (279.711µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.219924    1229 wrap.go:42] GET /apis/authorization.openshift.io/v1?timeout=32s: (325.891µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.220594    1229 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1?timeout=32s: (273.573µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.221300    1229 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1beta1?timeout=32s: (267.033µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.221970    1229 wrap.go:42] GET /apis/storage.k8s.io/v1?timeout=32s: (272.454µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.222623    1229 wrap.go:42] GET /apis/storage.k8s.io/v1beta1?timeout=32s: (275.749µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.223271    1229 wrap.go:42] GET /apis/admissionregistration.k8s.io/v1beta1?timeout=32s: (232.769µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.223943    1229 wrap.go:42] GET /apis/apiextensions.k8s.io/v1beta1?timeout=32s: (307.164µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.224595    1229 wrap.go:42] GET /apis/apps.openshift.io/v1?timeout=32s: (290.122µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.225288    1229 wrap.go:42] GET /apis/build.openshift.io/v1?timeout=32s: (289.819µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.226073    1229 wrap.go:42] GET /apis/image.openshift.io/v1?timeout=32s: (333.032µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.226825    1229 wrap.go:42] GET /apis/network.openshift.io/v1?timeout=32s: (312.666µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.227504    1229 wrap.go:42] GET /apis/oauth.openshift.io/v1?timeout=32s: (259.4µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.228199    1229 wrap.go:42] GET /apis/project.openshift.io/v1?timeout=32s: (260.39µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.228851    1229 wrap.go:42] GET /apis/quota.openshift.io/v1?timeout=32s: (269.049µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.229492    1229 wrap.go:42] GET /apis/route.openshift.io/v1?timeout=32s: (248.042µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.230126    1229 wrap.go:42] GET /apis/security.openshift.io/v1?timeout=32s: (268.253µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.230817    1229 wrap.go:42] GET /apis/template.openshift.io/v1?timeout=32s: (289.584µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.231479    1229 wrap.go:42] GET /apis/user.openshift.io/v1?timeout=32s: (270.144µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.246793    1229 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.274099ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.248033    1229 request.go:485] Throttling request took 93.053005ms, request: PUT:https://127.0.0.1:10280/api/v1/namespaces/openshift-infra/serviceaccounts/namespace-security-allocation-controller
    		I1219 18:09:32.251025    1229 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/namespace-security-allocation-controller: (2.723303ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45016]
    		I1219 18:09:32.251258    1229 create_dockercfg_secrets.go:444] Creating token secret "namespace-security-allocation-controller-token-dbxd9" for service account openshift-infra/namespace-security-allocation-controller
    		I1219 18:09:32.257262    1229 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/admin: (1.324369ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45018]
    		I1219 18:09:32.259103    1229 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/edit: (1.208652ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45018]
    		I1219 18:09:32.260668    1229 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/view: (989.113µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45018]
    		I1219 18:09:32.267358    1229 wrap.go:42] GET /api/v1/namespaces/default: (4.163207ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45018]
    		I1219 18:09:32.268462    1229 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.014862ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.268998    1229 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.076241ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.269351    1229 wrap.go:42] GET /api/v1/namespaces/default: (978.79µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.269948    1229 wrap.go:42] GET /apis/project.openshift.io/v1/projects/default: (2.08115ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45018]
    		I1219 18:09:32.271023    1229 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (926.305µs) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.272018    1229 wrap.go:42] GET /api/v1/namespaces/default: (1.000318ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44748]
    		I1219 18:09:32.272317    1229 wrap.go:42] GET /apis/project.openshift.io/v1/projects/default: (1.893692ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45018]
    		I1219 18:09:32.273959    1229 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1/apiservices/v1.: (1.243171ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45018]
    		I1219 18:09:32.275455    1229 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1/apiservices/v1.project.openshift.io: (986.799µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45018]
    		I1219 18:09:32.276873    1229 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1/apiservices/v1beta1.rbac.authorization.k8s.io: (952.833µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45018]
    		INFO: 2019/12/19 18:09:32 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc42200af60
    		INFO: 2019/12/19 18:09:32 dialing to target with scheme: ""
    		INFO: 2019/12/19 18:09:32 could not get resolver for scheme: ""
    		INFO: 2019/12/19 18:09:32 balancerWrapper: is pickfirst: false
    		INFO: 2019/12/19 18:09:32 balancerWrapper: got update addr from Notify: [{127.0.0.1:28035 <nil>}]
    		INFO: 2019/12/19 18:09:32 ccBalancerWrapper: new subconn: [{127.0.0.1:28035 0  <nil>}]
    		INFO: 2019/12/19 18:09:32 balancerWrapper: handle subconn state change: 0xc42598cf40, CONNECTING
    		INFO: 2019/12/19 18:09:32 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc42200af60
    		I1219 18:09:32.280225    1229 request.go:485] Throttling request took 123.683738ms, request: POST:https://127.0.0.1:10280/api/v1/namespaces/openshift-infra/secrets
    		I1219 18:09:32.283068    1229 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (2.432947ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45016]
    		I1219 18:09:32.284529    1229 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/horizontal-pod-autoscaler-token-924c8: (1.084464ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44748]
    		INFO: 2019/12/19 18:09:32 balancerWrapper: handle subconn state change: 0xc42598cf40, READY
    		INFO: 2019/12/19 18:09:32 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc42200af60
    		INFO: 2019/12/19 18:09:32 balancerWrapper: got update addr from Notify: [{127.0.0.1:28035 <nil>}]
    		
PASS
ok  	github.com/openshift/origin/test/integration/runner	1935.175s
[INFO] [18:09:32+0000] jUnit XML report placed at _output/scripts/test-integration/artifacts/gotest_report_9Iwcs.xml
Of 202 tests executed in 1935.175s, 202 succeeded, 0 failed, and 0 were skipped.
[INFO] [18:09:32+0000] hack/test-go.sh exited with code 0 after 00h 32m 16s
[WARNING] [18:09:36+0000] Copying _output/local/releases from the container failed!
[WARNING] [18:09:36+0000] Error response from daemon: lstat /var/lib/docker/overlay2/9ab9ff0844cb501d394c8bf17dcae91403850ec0f59e02c39c83d8d55ffeadd8/merged/go/src/github.com/openshift/origin/_output/local/releases: no such file or directory
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: RUN INTEGRATION TESTS [00h 44m 58s] ##########
[PostBuildScript] - [INFO] Executing post build scripts.
[workspace] $ /bin/bash /tmp/jenkins6822355103847251795.sh
########## STARTING STAGE: DOWNLOAD ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f
++ export PATH=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/gathered
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/gathered
+ mkdir -p /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/gathered
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo stat /data/src/github.com/openshift/origin/_output/scripts
  File: ‘/data/src/github.com/openshift/origin/_output/scripts’
  Size: 61        	Blocks: 0          IO Block: 4096   directory
Device: ca02h/51714d	Inode: 163601381   Links: 5
Access: (2755/drwxr-sr-x)  Uid: ( 1001/  origin)   Gid: ( 1003/origin-git)
Context: unconfined_u:object_r:container_file_t:s0
Access: 1970-01-01 00:00:00.000000000 +0000
Modify: 2019-12-19 17:26:09.000000000 +0000
Change: 2019-12-19 18:09:36.143454859 +0000
 Birth: -
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod -R o+rX /data/src/github.com/openshift/origin/_output/scripts
+ scp -r -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel:/data/src/github.com/openshift/origin/_output/scripts /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/gathered
+ tree /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/gathered
/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/gathered
└── scripts
    ├── shell
    │   ├── artifacts
    │   ├── logs
    │   │   ├── 3462147127e5fe2335de2f38c11c7c33a01ceb7a3296282e8277eecbbef23f6d.json
    │   │   └── scripts.log
    │   └── openshift.local.home
    ├── test-integration
    │   ├── artifacts
    │   │   ├── gotest_report_9Iwcs
    │   │   └── gotest_report_9Iwcs.xml
    │   ├── logs
    │   │   ├── raw_test_output.log
    │   │   ├── scripts.log
    │   │   └── test-go-err.log
    │   └── openshift.local.home
    └── test-tools
        ├── artifacts
        ├── logs
        │   ├── raw_test_output.log
        │   └── scripts.log
        └── openshift.local.home

13 directories, 9 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins2565891987265593534.sh
########## STARTING STAGE: GENERATE ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f
++ export PATH=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/generated
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/generated
+ mkdir /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/generated
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a 2>&1'
  WARNING: You're not using the default seccomp profile
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo cat /etc/sysconfig/docker /etc/sysconfig/docker-network /etc/sysconfig/docker-storage /etc/sysconfig/docker-storage-setup /etc/systemd/system/docker.service 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'oc get --raw /metrics --server=https://$( uname --nodename ):10250 --config=/etc/origin/master/admin.kubeconfig 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'oc get --raw /metrics --config=/etc/origin/master/admin.kubeconfig 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo df -T -h && sudo pvs && sudo vgs && sudo lvs && sudo findmnt --all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo yum list installed 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl --dmesg --no-pager --all --lines=all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl _PID=1 --no-pager --all --lines=all 2>&1'
+ tree /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/generated
/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/generated
├── avc_denials.log
├── containers.log
├── dmesg.log
├── docker.config
├── docker.info
├── filesystem.info
├── installed_packages.log
├── master-metrics.log
├── node-metrics.log
└── pid1.journal

0 directories, 10 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins8617939199148953445.sh
########## STARTING STAGE: FETCH SYSTEMD JOURNALS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f
++ export PATH=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/journals
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/journals
+ mkdir /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/journals
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit docker.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
+ tree /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/journals
/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/artifacts/journals
├── dnsmasq.service
├── docker.service
└── systemd-journald.service

0 directories, 3 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins8511484286959191809.sh
########## STARTING STAGE: ASSEMBLE GCS OUTPUT ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f
++ export PATH=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config
+ trap 'exit 0' EXIT
+ mkdir -p gcs/artifacts gcs/artifacts/generated gcs/artifacts/journals gcs/artifacts/gathered
++ python -c 'import json; import urllib; print json.load(urllib.urlopen('\''https://ci.openshift.redhat.com/jenkins/job/test_pull_request_origin_integration_310/673/api/json'\''))['\''result'\'']'
+ result=SUCCESS
+ cat
++ date +%s
+ cat /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/builds/673/log
+ cp artifacts/generated/avc_denials.log artifacts/generated/containers.log artifacts/generated/dmesg.log artifacts/generated/docker.config artifacts/generated/docker.info artifacts/generated/filesystem.info artifacts/generated/installed_packages.log artifacts/generated/master-metrics.log artifacts/generated/node-metrics.log artifacts/generated/pid1.journal gcs/artifacts/generated/
+ cp artifacts/journals/dnsmasq.service artifacts/journals/docker.service artifacts/journals/systemd-journald.service gcs/artifacts/journals/
+ cp -r artifacts/gathered/scripts gcs/artifacts/
++ pwd
+ scp -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config -r /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/gcs openshiftdevel:/data
+ scp -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.config/gcloud/gcs-publisher-credentials.json openshiftdevel:/data/credentials.json
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins6094757902044033024.sh
########## STARTING STAGE: PUSH THE ARTIFACTS AND METADATA ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f
++ export PATH=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config
++ mktemp
+ script=/tmp/tmp.0ElTfJgS8r
+ cat
+ chmod +x /tmp/tmp.0ElTfJgS8r
+ scp -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.0ElTfJgS8r openshiftdevel:/tmp/tmp.0ElTfJgS8r
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 300 /tmp/tmp.0ElTfJgS8r"'
+ cd /home/origin
+ trap 'exit 0' EXIT
+ [[ -n {"type":"presubmit","job":"test_pull_request_origin_integration_310","buildid":"1207711961345691649","prowjobid":"9f15a470-2283-11ea-9805-0a58ac1090fb","refs":{"org":"openshift","repo":"origin","repo_link":"https://github.com/openshift/origin","base_ref":"release-3.10","base_sha":"a729edd23a86825fe5c85ac67bdf3ebc63c22bec","base_link":"https://github.com/openshift/origin/commit/a729edd23a86825fe5c85ac67bdf3ebc63c22bec","pulls":[{"number":24021,"author":"sallyom","sha":"aba40bf1ccfb896917d2219c3f5b7507aee790e1","link":"https://github.com/openshift/origin/pull/24021","commit_link":"https://github.com/openshift/origin/pull/24021/commits/aba40bf1ccfb896917d2219c3f5b7507aee790e1","author_link":"https://github.com/sallyom"}]}} ]]
++ jq --compact-output '.buildid |= "673"'
+ JOB_SPEC='{"type":"presubmit","job":"test_pull_request_origin_integration_310","buildid":"673","prowjobid":"9f15a470-2283-11ea-9805-0a58ac1090fb","refs":{"org":"openshift","repo":"origin","repo_link":"https://github.com/openshift/origin","base_ref":"release-3.10","base_sha":"a729edd23a86825fe5c85ac67bdf3ebc63c22bec","base_link":"https://github.com/openshift/origin/commit/a729edd23a86825fe5c85ac67bdf3ebc63c22bec","pulls":[{"number":24021,"author":"sallyom","sha":"aba40bf1ccfb896917d2219c3f5b7507aee790e1","link":"https://github.com/openshift/origin/pull/24021","commit_link":"https://github.com/openshift/origin/pull/24021/commits/aba40bf1ccfb896917d2219c3f5b7507aee790e1","author_link":"https://github.com/sallyom"}]}}'
+ docker run -e 'JOB_SPEC={"type":"presubmit","job":"test_pull_request_origin_integration_310","buildid":"673","prowjobid":"9f15a470-2283-11ea-9805-0a58ac1090fb","refs":{"org":"openshift","repo":"origin","repo_link":"https://github.com/openshift/origin","base_ref":"release-3.10","base_sha":"a729edd23a86825fe5c85ac67bdf3ebc63c22bec","base_link":"https://github.com/openshift/origin/commit/a729edd23a86825fe5c85ac67bdf3ebc63c22bec","pulls":[{"number":24021,"author":"sallyom","sha":"aba40bf1ccfb896917d2219c3f5b7507aee790e1","link":"https://github.com/openshift/origin/pull/24021","commit_link":"https://github.com/openshift/origin/pull/24021/commits/aba40bf1ccfb896917d2219c3f5b7507aee790e1","author_link":"https://github.com/sallyom"}]}}' -v /data:/data:z registry.svc.ci.openshift.org/ci/gcsupload:latest --dry-run=false --gcs-path=gs://origin-ci-test --gcs-credentials-file=/data/credentials.json --path-strategy=single --default-org=openshift --default-repo=origin /data/gcs/artifacts /data/gcs/build-log.txt /data/gcs/finished.json
Unable to find image 'registry.svc.ci.openshift.org/ci/gcsupload:latest' locally
Trying to pull repository registry.svc.ci.openshift.org/ci/gcsupload ... 
latest: Pulling from registry.svc.ci.openshift.org/ci/gcsupload
a073c86ecf9e: Already exists
cc3fc741b1a9: Already exists
e5005ac95b8c: Pulling fs layer
85cea451eec0: Pulling fs layer
85cea451eec0: Verifying Checksum
85cea451eec0: Download complete
e5005ac95b8c: Verifying Checksum
e5005ac95b8c: Download complete
e5005ac95b8c: Pull complete
85cea451eec0: Pull complete
Digest: sha256:fc5343b5e0c51e5a55d013fdc06a444d9cfb928fddd5fe684b342d235988d898
Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/gcsupload:latest
{"component":"gcsupload","file":"prow/gcsupload/run.go:168","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts","level":"info","msg":"Gathering artifacts from artifact directory: /data/gcs/artifacts","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/avc_denials.log in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/avc_denials.log\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/containers.log in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/containers.log\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/dmesg.log in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/dmesg.log\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.config in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/docker.config\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.info in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/docker.info\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/filesystem.info in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/filesystem.info\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/installed_packages.log in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/installed_packages.log\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/master-metrics.log in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/master-metrics.log\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/node-metrics.log in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/node-metrics.log\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/pid1.journal in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/pid1.journal\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/journals/dnsmasq.service in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/journals/dnsmasq.service\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/journals/docker.service in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/journals/docker.service\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/journals/systemd-journald.service in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/journals/systemd-journald.service\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/scripts/shell/logs/3462147127e5fe2335de2f38c11c7c33a01ceb7a3296282e8277eecbbef23f6d.json in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/shell/logs/3462147127e5fe2335de2f38c11c7c33a01ceb7a3296282e8277eecbbef23f6d.json\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/scripts/shell/logs/scripts.log in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/shell/logs/scripts.log\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/artifacts/gotest_report_9Iwcs in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/artifacts/gotest_report_9Iwcs\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/artifacts/gotest_report_9Iwcs.xml in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/artifacts/gotest_report_9Iwcs.xml\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/logs/raw_test_output.log in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/logs/raw_test_output.log\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/logs/scripts.log in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/logs/scripts.log\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/logs/test-go-err.log in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/logs/test-go-err.log\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-tools/logs/raw_test_output.log in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-tools/logs/raw_test_output.log\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:186","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-tools/logs/scripts.log in artifact directory. Uploading as pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-tools/logs/scripts.log\n","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/filesystem.info","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/dmesg.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/docker.config","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/journals/docker.service","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/artifacts/gotest_report_9Iwcs","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/avc_denials.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/docker.info","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/shell/logs/3462147127e5fe2335de2f38c11c7c33a01ceb7a3296282e8277eecbbef23f6d.json","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/logs/raw_test_output.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/build-log.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/artifacts/gotest_report_9Iwcs.xml","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-tools/logs/scripts.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/journals/systemd-journald.service","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/node-metrics.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/logs/scripts.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-tools/logs/raw_test_output.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/finished.json","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration_310/673.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/master-metrics.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/journals/dnsmasq.service","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/shell/logs/scripts.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/logs/test-go-err.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration_310/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/installed_packages.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/containers.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/pid1.journal","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/docker.info","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-tools/logs/scripts.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/logs/scripts.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/journals/dnsmasq.service","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/master-metrics.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/artifacts/gotest_report_9Iwcs.xml","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/shell/logs/3462147127e5fe2335de2f38c11c7c33a01ceb7a3296282e8277eecbbef23f6d.json","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/dmesg.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration_310/673.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration_310/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/pid1.journal","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/shell/logs/scripts.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/installed_packages.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/finished.json","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/filesystem.info","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/node-metrics.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-tools/logs/raw_test_output.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/docker.config","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/journals/systemd-journald.service","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/artifacts/gotest_report_9Iwcs","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/logs/test-go-err.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/avc_denials.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/generated/containers.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/journals/docker.service","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/artifacts/scripts/test-integration/logs/raw_test_output.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:05Z"}
{"component":"gcsupload","dest":"pr-logs/pull/24021/test_pull_request_origin_integration_310/673/build-log.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-12-19T18:10:06Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:67","func":"k8s.io/test-infra/prow/gcsupload.Options.Run","level":"info","msg":"Finished upload to GCS","time":"2019-12-19T18:10:06Z"}
+ exit 0
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: PUSH THE ARTIFACTS AND METADATA [00h 00m 07s] ##########
[workspace] $ /bin/bash /tmp/jenkins9139215619067616412.sh
########## STARTING STAGE: DEPROVISION CLOUD RESOURCES ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f
++ export PATH=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config
+ oct deprovision

PLAYBOOK: main.yml *************************************************************
4 plays in /var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml

PLAY [ensure we have the parameters necessary to deprovision virtual hosts] ****

TASK [ensure all required variables are set] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:9
skipping: [localhost] => (item=origin_ci_inventory_dir)  => {
    "changed": false, 
    "generated_timestamp": "2019-12-19 13:10:07.542588", 
    "item": "origin_ci_inventory_dir", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_region)  => {
    "changed": false, 
    "generated_timestamp": "2019-12-19 13:10:07.545862", 
    "item": "origin_ci_aws_region", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

PLAY [deprovision virtual hosts in EC2] ****************************************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

TASK [deprovision a virtual EC2 host] ******************************************
task path: /var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:28
included: /var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml for localhost

TASK [update the SSH configuration to remove AWS EC2 specifics] ****************
task path: /var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:2
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-12-19 13:10:08.499694", 
    "msg": ""
}

TASK [rename EC2 instance for termination reaper] ******************************
task path: /var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:8
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-12-19 13:10:09.239050", 
    "msg": "Tags {'Name': 'oct-terminate'} created for resource i-073664f2ce6c0d1a3."
}

TASK [tear down the EC2 instance] **********************************************
task path: /var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:15
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-12-19 13:10:10.219241", 
    "instance_ids": [
        "i-073664f2ce6c0d1a3"
    ], 
    "instances": [
        {
            "ami_launch_index": "0", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-084ddbd146646982b"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-0b14f56fb05fd21a6"
                }
            }, 
            "dns_name": "ec2-54-211-163-226.compute-1.amazonaws.com", 
            "ebs_optimized": false, 
            "groups": {
                "sg-7e73221a": "default"
            }, 
            "hypervisor": "xen", 
            "id": "i-073664f2ce6c0d1a3", 
            "image_id": "ami-01a9f71561742665f", 
            "instance_type": "m4.xlarge", 
            "kernel": null, 
            "key_name": "openshift-dev", 
            "launch_time": "2019-12-19T17:19:33.000Z", 
            "placement": "us-east-1d", 
            "private_dns_name": "ip-172-18-11-188.ec2.internal", 
            "private_ip": "172.18.11.188", 
            "public_dns_name": "ec2-54-211-163-226.compute-1.amazonaws.com", 
            "public_ip": "54.211.163.226", 
            "ramdisk": null, 
            "region": "us-east-1", 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "running", 
            "state_code": 16, 
            "tags": {
                "Name": "oct-terminate", 
                "openshift_etcd": "", 
                "openshift_master": "", 
                "openshift_node": ""
            }, 
            "tenancy": "default", 
            "virtualization_type": "hvm"
        }
    ], 
    "tagged_instances": []
}

TASK [remove the serialized host variables] ************************************
task path: /var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:22
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-12-19 13:10:10.472591", 
    "path": "/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory/host_vars/172.18.11.188.yml", 
    "state": "absent"
}

PLAY [deprovision virtual hosts locally manged by Vagrant] *********************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

PLAY [clean up local configuration for deprovisioned instances] ****************

TASK [remove inventory configuration directory] ********************************
task path: /var/lib/jenkins/origin-ci-tool/6f8991b9d2cf322fadd103555399c635aac6ce5f/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:61
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-12-19 13:10:11.093678", 
    "path": "/var/lib/jenkins/jobs/test_pull_request_origin_integration_310/workspace/.config/origin-ci-tool/inventory", 
    "state": "absent"
}

PLAY RECAP *********************************************************************
localhost                  : ok=8    changed=4    unreachable=0    failed=0   

+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPROVISION CLOUD RESOURCES [00h 00m 05s] ##########
Archiving artifacts
[WS-CLEANUP] Deleting project workspace...
[WS-CLEANUP] Deferred wipeout is used...
[WS-CLEANUP] done
Finished: SUCCESS