FailedConsole Output

Skipping 13,598 KB.. Full Log
    		I0522 14:56:59.436570    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (5.609642ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.437014    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-controller: (6.590819ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:33328]
    		I0522 14:56:59.448425    1203 get.go:238] Starting watch for /api/v1/namespaces/openshift-infra/secrets, rv=346 labels= fields=type=kubernetes.io/service-account-token timeout=1h16m52.758959682s
    		I0522 14:56:59.448967    1203 wrap.go:42] GET /api/v1/namespaces/kube-system/configmaps/kube-scheduler: (11.161119ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:33328]
    		I0522 14:56:59.452736    1203 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (5.830173ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:33328]
    		I0522 14:56:59.457460    1203 wrap.go:42] PUT /api/v1/namespaces/kube-system/configmaps/kube-scheduler: (6.619099ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:33328]
    		I0522 14:56:59.457731    1203 leaderelection.go:199] successfully renewed lease kube-system/kube-scheduler
    		I0522 14:56:59.459309    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/template-instance-controller-token-p666c: (3.380837ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.461719    1203 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-controller: (8.36443ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:33328]
    		I0522 14:56:59.463433    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-controller: (3.096852ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.463875    1203 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (9.377387ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.464226    1203 client_builder.go:233] Verified credential for template-instance-controller/openshift-infra
    		I0522 14:56:59.464484    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&resourceVersion=346&watch=true: (16.521653ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.477826    1203 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/admin?timeout=1m0s: (13.572431ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33900]
    		I0522 14:56:59.478082    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-controller: (13.550489ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.482770    1203 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/edit?timeout=1m0s: (3.106074ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33900]
    		I0522 14:56:59.484242    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (4.008128ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.487297    1203 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (667.657µs) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.487554    1203 client_builder.go:233] Verified credential for template-instance-controller/openshift-infra
    		I0522 14:56:59.488675    1203 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/view?timeout=1m0s: (2.634232ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33900]
    		I0522 14:56:59.489756    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-controller: (1.512303ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.493059    1203 wrap.go:42] GET /api?timeout=1m0s: (675.312µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33900]
    		I0522 14:56:59.493256    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (3.048844ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.494585    1203 wrap.go:42] GET /apis?timeout=1m0s: (1.008447ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33900]
    		I0522 14:56:59.495096    1203 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (659.788µs) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.495373    1203 client_builder.go:233] Verified credential for template-instance-controller/openshift-infra
    		I0522 14:56:59.496309    1203 wrap.go:42] GET /apis/build.openshift.io/v1?timeout=1m0s: (792.876µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33900]
    		I0522 14:56:59.496870    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-controller: (1.254093ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.500683    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (3.27899ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.502503    1203 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (609.691µs) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.502944    1203 client_builder.go:233] Verified credential for template-instance-controller/openshift-infra
    		I0522 14:56:59.503157    1203 controller_manager.go:201] Started "openshift.io/templateinstance"
    		I0522 14:56:59.503173    1203 controller_manager.go:191] Starting "openshift.io/namespace-security-allocation"
    		I0522 14:56:59.504578    1203 wrap.go:42] POST /api/v1/namespaces?timeout=1m0s: (5.484262ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33900]
    		I0522 14:56:59.504605    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/namespace-security-allocation-controller: (1.202606ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.505859    1203 wrap.go:42] GET /apis/build.openshift.io/v1: (494.036µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33900]
    		I0522 14:56:59.506604    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra: (1.268549ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.510343    1203 wrap.go:42] POST /api/v1/namespaces/openshift-infra/serviceaccounts: (3.03013ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.517046    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.267144ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.517239    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.026924ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.517823    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.50302ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.528056    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/namespace-security-allocation-controller: (17.244825ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:33328]
    		I0522 14:56:59.530347    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (18.03549ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.534900    1203 get.go:238] Starting watch for /api/v1/namespaces/openshift-infra/secrets, rv=351 labels= fields=type=kubernetes.io/service-account-token timeout=1h48m53.573895376s
    		I0522 14:56:59.541970    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (6.891258ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.542225    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.534039ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.543160    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (4.006525ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.543360    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (4.698874ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.543476    1203 wrap.go:42] PUT /api/v1/namespaces/testapigroup700101949/finalize: (31.721807ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.543526    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (5.407264ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.543649    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (6.030362ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.544121    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (8.470003ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.544178    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (4.562944ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.544320    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (3.118056ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.544354    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (7.256272ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.544469    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (7.88376ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.544538    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (3.858185ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.544661    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (4.48902ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.544703    1203 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (8.584095ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.545514    1203 wrap.go:42] POST /apis/authorization.k8s.io/v1/subjectaccessreviews: (756.632µs) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.547127    1203 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (9.005086ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:33328]
    		I0522 14:56:59.547958    1203 controller.go:537] quota admission added evaluator for: {build.openshift.io builds}
    		I0522 14:56:59.548304    1203 controller.go:537] quota admission added evaluator for: {build.openshift.io builds}
    		I0522 14:56:59.555732    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/namespace-security-allocation-controller-token-mb78n: (3.652015ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.556039    1203 wrap.go:42] GET /api/v1/namespaces/testapigroup700101949/resourcequotas: (5.780694ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.558655    1203 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/namespace-security-allocation-controller: (7.827364ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:33328]
    		I0522 14:56:59.579180    1203 wrap.go:42] POST /apis/build.openshift.io/v1/namespaces/testapigroup700101949/builds?timeout=1m0s: (68.617543ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33900]
    		I0522 14:56:59.579564    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/namespace-security-allocation-controller: (22.601903ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.580102    1203 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (30.936501ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.580408    1203 client_builder.go:233] Verified credential for namespace-security-allocation-controller/openshift-infra
    		I0522 14:56:59.581116    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&resourceVersion=351&watch=true: (46.791752ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.581653    1203 wrap.go:42] GET /apis/build.openshift.io/v1/namespaces/testapigroup700101949/builds/foo: (1.439895ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33900]
    		I0522 14:56:59.582613    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/namespace-security-allocation-controller: (1.444962ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.586387    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (3.244045ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.588264    1203 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (694.995µs) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.588541    1203 client_builder.go:233] Verified credential for namespace-security-allocation-controller/openshift-infra
    		I0522 14:56:59.588801    1203 controller_manager.go:201] Started "openshift.io/namespace-security-allocation"
    		I0522 14:56:59.588818    1203 controller_manager.go:191] Starting "openshift.io/image-import"
    		I0522 14:56:59.588958    1203 controller_utils.go:1019] Waiting for caches to sync for namespace-security-allocation-controller controller
    		I0522 14:56:59.596641    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/image-import-controller: (7.557904ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		I0522 14:56:59.596929    1203 wrap.go:42] DELETE /api/v1/namespaces/testapigroup700101949?timeout=1m0s: (14.393065ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33900]
    		INFO: 2018/05/22 14:56:59 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc43c3f0b40
    		INFO: 2018/05/22 14:56:59 dialing to target with scheme: ""
    		INFO: 2018/05/22 14:56:59 could not get resolver for scheme: ""
    		INFO: 2018/05/22 14:56:59 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/22 14:56:59 balancerWrapper: got update addr from Notify: [{127.0.0.1:17773 <nil>}]
    		INFO: 2018/05/22 14:56:59 ccBalancerWrapper: new subconn: [{127.0.0.1:17773 0  <nil>}]
    		INFO: 2018/05/22 14:56:59 balancerWrapper: handle subconn state change: 0xc42bc74730, CONNECTING
    		INFO: 2018/05/22 14:56:59 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc43c3f0b40
    		I0522 14:56:59.657052    1203 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (59.661893ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:33328]
    		INFO: 2018/05/22 14:56:59 balancerWrapper: handle subconn state change: 0xc42bc74730, READY
    		INFO: 2018/05/22 14:56:59 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc43c3f0b40
    		INFO: 2018/05/22 14:56:59 balancerWrapper: got update addr from Notify: [{127.0.0.1:17773 <nil>}]
    		
    --- PASS: TestIntegration/TestAlwaysPullImagesOff (41.47s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		.186ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:52614]
    		I0522 14:57:18.286430    1216 wrap.go:42] PUT /api/v1/namespaces/kube-system/configmaps/kube-scheduler: (8.628998ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:52614]
    		I0522 14:57:18.286954    1216 leaderelection.go:199] successfully renewed lease kube-system/kube-scheduler
    		I0522 14:57:18.289106    1216 wrap.go:42] PUT /api/v1/namespaces/integration/serviceaccounts/deployer: (10.061638ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:53064]
    		I0522 14:57:18.333065    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.458078ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.352623    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (9.0147ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.352642    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (13.133239ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.352754    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (10.85497ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.352860    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (11.271968ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.353054    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (9.657575ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.353221    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (10.690185ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.353335    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (11.912928ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.353389    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (12.411923ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.353486    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (12.9791ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.353530    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (10.340328ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.353643    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (13.586946ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.353689    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (10.707963ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.353871    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (11.531956ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.354031    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (11.24676ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:18.354185    1216 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (12.094594ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:19.121135    1216 wrap.go:42] GET /api/v1/namespaces/integration/serviceaccounts/default?timeout=1m0s: (1.425145ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53082]
    		I0522 14:57:19.123077    1216 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-token-j92rz?timeout=1m0s: (1.326605ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53082]
    		I0522 14:57:19.124784    1216 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-dockercfg-tcrwk?timeout=1m0s: (1.110632ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53082]
    		I0522 14:57:19.127441    1216 wrap.go:42] GET /api/v1/namespaces/integration/limitranges: (1.22209ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:19.128718    1216 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I0522 14:57:19.128796    1216 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I0522 14:57:19.131690    1216 wrap.go:42] GET /api/v1/namespaces/integration: (1.842202ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:19.131966    1216 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 14:57:19.132007    1216 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 14:57:19.132019    1216 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 14:57:19.132028    1216 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 14:57:19.132049    1216 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 14:57:19.132060    1216 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 14:57:19.132077    1216 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 14:57:19.132086    1216 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 14:57:19.132094    1216 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 14:57:19.132107    1216 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 14:57:19.132128    1216 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 14:57:19.132138    1216 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 14:57:19.132146    1216 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 14:57:19.132166    1216 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 14:57:19.132256    1216 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I0522 14:57:19.134320    1216 wrap.go:42] POST /api/v1/namespaces/integration/pods?timeout=1m0s: (8.908718ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53082]
    		I0522 14:57:19.135449    1216 factory.go:1147] About to try and schedule pod testkmqhw
    		I0522 14:57:19.135465    1216 scheduler.go:439] Attempting to schedule pod: integration/testkmqhw
    		I0522 14:57:19.135484    1216 scheduler.go:191] Failed to schedule pod: integration/testkmqhw
    		I0522 14:57:19.135541    1216 factory.go:1262] Unable to schedule integration testkmqhw: no nodes are registered to the cluster; waiting
    		I0522 14:57:19.135579    1216 factory.go:1375] Updating pod condition for integration/testkmqhw to (PodScheduled==False)
    		I0522 14:57:19.135834    1216 pvc_protection_controller.go:276] Got event on pod integration/testkmqhw
    		I0522 14:57:19.135907    1216 disruption.go:328] addPod called on pod "testkmqhw"
    		I0522 14:57:19.135922    1216 disruption.go:403] No PodDisruptionBudgets found for pod testkmqhw, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:19.135930    1216 disruption.go:331] No matching pdb for pod "testkmqhw"
    		I0522 14:57:19.135966    1216 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"testkmqhw"}
    		I0522 14:57:19.136038    1216 backoff_utils.go:79] Backing off 1s
    		I0522 14:57:19.138570    1216 wrap.go:42] PUT /api/v1/namespaces/integration/pods/testkmqhw/status: (2.38523ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:52614]
    		I0522 14:57:19.138815    1216 pvc_protection_controller.go:276] Got event on pod integration/testkmqhw
    		I0522 14:57:19.138894    1216 disruption.go:340] updatePod called on pod "testkmqhw"
    		I0522 14:57:19.138911    1216 disruption.go:403] No PodDisruptionBudgets found for pod testkmqhw, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:19.138919    1216 disruption.go:343] No matching pdb for pod "testkmqhw"
    		I0522 14:57:19.139098    1216 factory.go:1147] About to try and schedule pod testkmqhw
    		I0522 14:57:19.139115    1216 scheduler.go:439] Attempting to schedule pod: integration/testkmqhw
    		I0522 14:57:19.139132    1216 scheduler.go:191] Failed to schedule pod: integration/testkmqhw
    		I0522 14:57:19.139148    1216 factory.go:1262] Unable to schedule integration testkmqhw: no nodes are registered to the cluster; waiting
    		I0522 14:57:19.139176    1216 factory.go:1375] Updating pod condition for integration/testkmqhw to (PodScheduled==False)
    		W0522 14:57:19.139238    1216 factory.go:1304] Request for pod integration/testkmqhw already in flight, abandoning
    		I0522 14:57:19.141059    1216 store.go:370] GuaranteedUpdate of /kubernetes.io/pods/integration/testkmqhw failed because of a conflict, going to retry
    		I0522 14:57:19.145310    1216 wrap.go:42] POST /api/v1/namespaces/integration/events: (8.974762ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:52614]
    		I0522 14:57:19.145571    1216 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testkmqhw, uid 6c98a800-5dd0-11e8-9c33-0242ac110002, event type update
    		I0522 14:57:19.145745    1216 factory.go:1147] About to try and schedule pod testkmqhw
    		I0522 14:57:19.145783    1216 scheduler.go:435] Skip schedule deleting pod: integration/testkmqhw
    		I0522 14:57:19.145982    1216 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testkmqhw, uid 6c98a800-5dd0-11e8-9c33-0242ac110002, event type update
    		I0522 14:57:19.146026    1216 pvc_protection_controller.go:276] Got event on pod integration/testkmqhw
    		I0522 14:57:19.146061    1216 disruption.go:340] updatePod called on pod "testkmqhw"
    		I0522 14:57:19.146081    1216 disruption.go:403] No PodDisruptionBudgets found for pod testkmqhw, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:19.146089    1216 disruption.go:343] No matching pdb for pod "testkmqhw"
    		I0522 14:57:19.147969    1216 wrap.go:42] DELETE /api/v1/namespaces/integration/pods/testkmqhw?timeout=1m0s: (12.122948ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53082]
    		I0522 14:57:19.148085    1216 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testkmqhw, uid 6c98a800-5dd0-11e8-9c33-0242ac110002, event type delete
    		I0522 14:57:19.148263    1216 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testkmqhw, uid 6c98a800-5dd0-11e8-9c33-0242ac110002, event type delete
    		I0522 14:57:19.148292    1216 pvc_protection_controller.go:276] Got event on pod integration/testkmqhw
    		I0522 14:57:19.148314    1216 disruption.go:369] deletePod called on pod "testkmqhw"
    		I0522 14:57:19.148324    1216 disruption.go:403] No PodDisruptionBudgets found for pod testkmqhw, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:19.148332    1216 disruption.go:372] No matching pdb for pod "testkmqhw"
    		I0522 14:57:19.148350    1216 taint_manager.go:338] Noticed pod deletion: types.NamespacedName{Namespace:"integration", Name:"testkmqhw"}
    		I0522 14:57:19.148374    1216 deployment_controller.go:357] Pod testkmqhw deleted.
    		I0522 14:57:19.148989    1216 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I0522 14:57:19.149023    1216 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I0522 14:57:19.150979    1216 wrap.go:42] GET /api/v1/namespaces/integration: (1.160391ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52614]
    		I0522 14:57:19.151274    1216 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 14:57:19.151304    1216 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 14:57:19.151314    1216 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 14:57:19.151323    1216 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 14:57:19.151342    1216 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 14:57:19.151353    1216 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 14:57:19.151378    1216 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 14:57:19.151388    1216 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 14:57:19.151399    1216 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 14:57:19.151413    1216 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 14:57:19.151437    1216 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 14:57:19.151457    1216 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 14:57:19.151466    1216 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 14:57:19.151486    1216 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 14:57:19.151542    1216 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I0522 14:57:19.152544    1216 wrap.go:42] PATCH /api/v1/namespaces/integration/events/testkmqhw.1530ff5acd53e9b3: (6.079068ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:52614]
    		I0522 14:57:19.154566    1216 wrap.go:42] POST /api/v1/namespaces/integration/pods?timeout=1m0s: (6.038376ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:53082]
    		I0522 14:57:19.155507    1216 factory.go:1147] About to try and schedule pod test4csfc
    		I0522 14:57:19.155522    1216 scheduler.go:439] Attempting to schedule pod: integration/test4csfc
    		I0522 14:57:19.155538    1216 scheduler.go:191] Failed to schedule pod: integration/test4csfc
    		I0522 14:57:19.155555    1216 factory.go:1262] Unable to schedule integration test4csfc: no nodes are registered to the cluster; waiting
    		I0522 14:57:19.155582    1216 factory.go:1375] Updating pod condition for integration/test4csfc to (PodScheduled==False)
    		I0522 14:57:19.155781    1216 pvc_protection_controller.go:276] Got event on pod integration/test4csfc
    		I0522 14:57:19.155817    1216 disruption.go:328] addPod called on pod "test4csfc"
    		I0522 14:57:19.155830    1216 disruption.go:403] No PodDisruptionBudgets found for pod test4csfc, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:19.155838    1216 disruption.go:331] No matching pdb for pod "test4csfc"
    		I0522 14:57:19.155859    1216 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"test4csfc"}
    		I0522 14:57:19.155917    1216 backoff_utils.go:79] Backing off 1s
    		INFO: 2018/05/22 14:57:19 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc433651200
    		INFO: 2018/05/22 14:57:19 dialing to target with scheme: ""
    		INFO: 2018/05/22 14:57:19 could not get resolver for scheme: ""
    		INFO: 2018/05/22 14:57:19 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/22 14:57:19 balancerWrapper: got update addr from Notify: [{127.0.0.1:14785 <nil>}]
    		INFO: 2018/05/22 14:57:19 ccBalancerWrapper: new subconn: [{127.0.0.1:14785 0  <nil>}]
    		INFO: 2018/05/22 14:57:19 balancerWrapper: handle subconn state change: 0xc42e50b350, CONNECTING
    		INFO: 2018/05/22 14:57:19 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc433651200
    		I0522 14:57:19.160107    1216 wrap.go:42] PUT /api/v1/namespaces/integration/pods/test4csfc/status: (3.159684ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:52614]
    		I0522 14:57:19.160983    1216 pvc_protection_controller.go:276] Got event on pod integration/test4csfc
    		I0522 14:57:19.161021    1216 disruption.go:340] updatePod called on pod "test4csfc"
    		I0522 14:57:19.161033    1216 disruption.go:403] No PodDisruptionBudgets found for pod test4csfc, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:19.161041    1216 disruption.go:343] No matching pdb for pod "test4csfc"
    		I0522 14:57:19.161529    1216 factory.go:1147] About to try and schedule pod test4csfc
    		I0522 14:57:19.161543    1216 scheduler.go:439] Attempting to schedule pod: integration/test4csfc
    		I0522 14:57:19.161559    1216 scheduler.go:191] Failed to schedule pod: integration/test4csfc
    		I0522 14:57:19.161600    1216 factory.go:1262] Unable to schedule integration test4csfc: no nodes are registered to the cluster; waiting
    		I0522 14:57:19.161632    1216 factory.go:1375] Updating pod condition for integration/test4csfc to (PodScheduled==False)
    		W0522 14:57:19.161672    1216 factory.go:1304] Request for pod integration/test4csfc already in flight, abandoning
    		I0522 14:57:19.162996    1216 wrap.go:42] POST /api/v1/namespaces/integration/events: (6.873526ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:52614]
    		I0522 14:57:19.170982    1216 wrap.go:42] POST /api/v1/namespaces/integration/events: (7.170198ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:52614]
    		INFO: 2018/05/22 14:57:19 balancerWrapper: handle subconn state change: 0xc42e50b350, READY
    		INFO: 2018/05/22 14:57:19 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc433651200
    		
    --- PASS: TestIntegration/TestAlwaysPullImagesOn (36.04s)
    	runner_test.go:187: 
    		I0522 14:57:22.263827    1228 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (920.928µs) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54122]
    		INFO: 2018/05/22 14:57:22 balancerWrapper: got update addr from Notify: [{127.0.0.1:16742 <nil>}]
    		
    		=== OUTPUT
    		roller: (2.094171ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54744]
    		I0522 14:57:21.983398    1228 create_dockercfg_secrets.go:478] Creating dockercfg secret "deployment-controller-dockercfg-jjw9z" for service account kube-system/deployment-controller
    		I0522 14:57:22.013061    1228 request.go:485] Throttling request took 62.331943ms, request: POST:https://127.0.0.1:13895/api/v1/namespaces/integration/secrets
    		I0522 14:57:22.015149    1228 wrap.go:42] POST /api/v1/namespaces/integration/secrets: (1.818935ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54744]
    		I0522 14:57:22.047748    1228 wrap.go:42] PUT /api/v1/namespaces/integration/serviceaccounts/default: (2.169315ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54744]
    		I0522 14:57:22.047988    1228 create_dockercfg_secrets.go:478] Creating dockercfg secret "daemon-set-controller-dockercfg-2w4n4" for service account kube-system/daemon-set-controller
    		I0522 14:57:22.077792    1228 request.go:485] Throttling request took 62.350536ms, request: POST:https://127.0.0.1:13895/api/v1/namespaces/kube-system/secrets
    		I0522 14:57:22.081103    1228 wrap.go:42] POST /api/v1/namespaces/kube-system/secrets: (2.4142ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54744]
    		I0522 14:57:22.112730    1228 wrap.go:42] PUT /api/v1/namespaces/kube-system/serviceaccounts/attachdetach-controller: (2.602922ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54744]
    		I0522 14:57:22.112995    1228 create_dockercfg_secrets.go:478] Creating dockercfg secret "default-dockercfg-nkmwl" for service account kube-system/default
    		I0522 14:57:22.142069    1228 request.go:485] Throttling request took 60.65634ms, request: POST:https://127.0.0.1:13895/api/v1/namespaces/kube-public/secrets
    		I0522 14:57:22.144702    1228 wrap.go:42] POST /api/v1/namespaces/kube-public/secrets: (2.39914ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54744]
    		I0522 14:57:22.177274    1228 wrap.go:42] PUT /api/v1/namespaces/kube-public/serviceaccounts/default: (2.604674ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54744]
    		I0522 14:57:22.177492    1228 create_dockercfg_secrets.go:478] Creating dockercfg secret "cronjob-controller-dockercfg-s96lv" for service account kube-system/cronjob-controller
    		I0522 14:57:22.206644    1228 request.go:485] Throttling request took 61.615988ms, request: POST:https://127.0.0.1:13895/api/v1/namespaces/kube-system/secrets
    		I0522 14:57:22.208788    1228 wrap.go:42] POST /api/v1/namespaces/kube-system/secrets: (1.848416ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54744]
    		I0522 14:57:22.223287    1228 wrap.go:42] GET /api/v1/namespaces/integration/serviceaccounts/default?timeout=1m0s: (1.180455ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54776]
    		I0522 14:57:22.225130    1228 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-token-5sf2w?timeout=1m0s: (1.256921ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54776]
    		I0522 14:57:22.226884    1228 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-dockercfg-tbx9m?timeout=1m0s: (1.120278ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54776]
    		I0522 14:57:22.229527    1228 wrap.go:42] GET /api/v1/namespaces/integration/limitranges: (1.272769ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54122]
    		I0522 14:57:22.229865    1228 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I0522 14:57:22.229931    1228 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I0522 14:57:22.231726    1228 wrap.go:42] GET /api/v1/namespaces/integration: (988.265µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54122]
    		I0522 14:57:22.231931    1228 matcher.go:292] got preallocated value for level: s0:c3,c2 for selinux options in namespace integration
    		I0522 14:57:22.231968    1228 matcher.go:279] got preallocated values for min: 1000010000, max: 1000019999 for uid range in namespace integration
    		I0522 14:57:22.231975    1228 matcher.go:292] got preallocated value for level: s0:c3,c2 for selinux options in namespace integration
    		I0522 14:57:22.231981    1228 matcher.go:322] got preallocated value for groups: 1000010000/10000 in namespace integration
    		I0522 14:57:22.231993    1228 matcher.go:292] got preallocated value for level: s0:c3,c2 for selinux options in namespace integration
    		I0522 14:57:22.232010    1228 matcher.go:292] got preallocated value for level: s0:c3,c2 for selinux options in namespace integration
    		I0522 14:57:22.232020    1228 matcher.go:279] got preallocated values for min: 1000010000, max: 1000019999 for uid range in namespace integration
    		I0522 14:57:22.232027    1228 matcher.go:292] got preallocated value for level: s0:c3,c2 for selinux options in namespace integration
    		I0522 14:57:22.232042    1228 matcher.go:322] got preallocated value for groups: 1000010000/10000 in namespace integration
    		I0522 14:57:22.232056    1228 matcher.go:342] got preallocated value for groups: 1000010000/10000 in namespace integration
    		I0522 14:57:22.232079    1228 matcher.go:279] got preallocated values for min: 1000010000, max: 1000019999 for uid range in namespace integration
    		I0522 14:57:22.232088    1228 matcher.go:292] got preallocated value for level: s0:c3,c2 for selinux options in namespace integration
    		I0522 14:57:22.232094    1228 matcher.go:322] got preallocated value for groups: 1000010000/10000 in namespace integration
    		I0522 14:57:22.232116    1228 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 14:57:22.232192    1228 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I0522 14:57:22.234067    1228 wrap.go:42] POST /api/v1/namespaces/integration/pods?timeout=1m0s: (6.630689ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54776]
    		I0522 14:57:22.234378    1228 factory.go:1147] About to try and schedule pod test68zrt
    		I0522 14:57:22.234398    1228 scheduler.go:439] Attempting to schedule pod: integration/test68zrt
    		I0522 14:57:22.234420    1228 scheduler.go:191] Failed to schedule pod: integration/test68zrt
    		I0522 14:57:22.234478    1228 factory.go:1262] Unable to schedule integration test68zrt: no nodes are registered to the cluster; waiting
    		I0522 14:57:22.234522    1228 factory.go:1375] Updating pod condition for integration/test68zrt to (PodScheduled==False)
    		I0522 14:57:22.234573    1228 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"test68zrt"}
    		I0522 14:57:22.234589    1228 pvc_protection_controller.go:276] Got event on pod integration/test68zrt
    		I0522 14:57:22.234663    1228 disruption.go:328] addPod called on pod "test68zrt"
    		I0522 14:57:22.234677    1228 disruption.go:403] No PodDisruptionBudgets found for pod test68zrt, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:22.234685    1228 disruption.go:331] No matching pdb for pod "test68zrt"
    		I0522 14:57:22.234687    1228 backoff_utils.go:79] Backing off 1s
    		I0522 14:57:22.238028    1228 wrap.go:42] PUT /api/v1/namespaces/integration/pods/test68zrt/status: (3.202763ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:54122]
    		I0522 14:57:22.238155    1228 store.go:370] GuaranteedUpdate of /kubernetes.io/pods/integration/test68zrt failed because of a conflict, going to retry
    		I0522 14:57:22.238637    1228 pvc_protection_controller.go:276] Got event on pod integration/test68zrt
    		I0522 14:57:22.238770    1228 disruption.go:340] updatePod called on pod "test68zrt"
    		I0522 14:57:22.238785    1228 disruption.go:403] No PodDisruptionBudgets found for pod test68zrt, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:22.238794    1228 disruption.go:343] No matching pdb for pod "test68zrt"
    		I0522 14:57:22.239070    1228 wrap.go:42] POST /api/v1/namespaces/integration/events: (4.121391ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:54122]
    		I0522 14:57:22.239282    1228 factory.go:1147] About to try and schedule pod test68zrt
    		I0522 14:57:22.239296    1228 scheduler.go:439] Attempting to schedule pod: integration/test68zrt
    		I0522 14:57:22.239311    1228 scheduler.go:191] Failed to schedule pod: integration/test68zrt
    		I0522 14:57:22.239327    1228 factory.go:1262] Unable to schedule integration test68zrt: no nodes are registered to the cluster; waiting
    		I0522 14:57:22.239353    1228 factory.go:1375] Updating pod condition for integration/test68zrt to (PodScheduled==False)
    		W0522 14:57:22.239422    1228 factory.go:1304] Request for pod integration/test68zrt already in flight, abandoning
    		I0522 14:57:22.241489    1228 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name test68zrt, uid 6e71aafb-5dd0-11e8-9dea-0242ac110002, event type update
    		I0522 14:57:22.241564    1228 factory.go:1147] About to try and schedule pod test68zrt
    		I0522 14:57:22.241571    1228 disruption.go:340] updatePod called on pod "test68zrt"
    		I0522 14:57:22.241583    1228 pvc_protection_controller.go:276] Got event on pod integration/test68zrt
    		I0522 14:57:22.241587    1228 disruption.go:403] No PodDisruptionBudgets found for pod test68zrt, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:22.241597    1228 disruption.go:343] No matching pdb for pod "test68zrt"
    		I0522 14:57:22.241586    1228 scheduler.go:435] Skip schedule deleting pod: integration/test68zrt
    		I0522 14:57:22.241618    1228 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name test68zrt, uid 6e71aafb-5dd0-11e8-9dea-0242ac110002, event type update
    		I0522 14:57:22.244298    1228 wrap.go:42] DELETE /api/v1/namespaces/integration/pods/test68zrt?timeout=1m0s: (9.344769ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54776]
    		I0522 14:57:22.244768    1228 taint_manager.go:338] Noticed pod deletion: types.NamespacedName{Namespace:"integration", Name:"test68zrt"}
    		I0522 14:57:22.244847    1228 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name test68zrt, uid 6e71aafb-5dd0-11e8-9dea-0242ac110002, event type delete
    		I0522 14:57:22.244895    1228 deployment_controller.go:357] Pod test68zrt deleted.
    		I0522 14:57:22.244910    1228 pvc_protection_controller.go:276] Got event on pod integration/test68zrt
    		I0522 14:57:22.244930    1228 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name test68zrt, uid 6e71aafb-5dd0-11e8-9dea-0242ac110002, event type delete
    		I0522 14:57:22.244983    1228 disruption.go:369] deletePod called on pod "test68zrt"
    		I0522 14:57:22.244994    1228 disruption.go:403] No PodDisruptionBudgets found for pod test68zrt, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:22.245001    1228 disruption.go:372] No matching pdb for pod "test68zrt"
    		I0522 14:57:22.245380    1228 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I0522 14:57:22.245418    1228 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I0522 14:57:22.246110    1228 wrap.go:42] PUT /api/v1/namespaces/kube-system/serviceaccounts/clusterrole-aggregation-controller: (6.02692ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:54744]
    		I0522 14:57:22.246385    1228 create_dockercfg_secrets.go:478] Creating dockercfg secret "disruption-controller-dockercfg-ngz2z" for service account kube-system/disruption-controller
    		I0522 14:57:22.247691    1228 wrap.go:42] PATCH /api/v1/namespaces/integration/events/test68zrt.1530ff5b860a067b: (7.610918ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:54122]
    		I0522 14:57:22.247858    1228 wrap.go:42] GET /api/v1/namespaces/integration: (1.369866ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54122]
    		I0522 14:57:22.248140    1228 matcher.go:292] got preallocated value for level: s0:c3,c2 for selinux options in namespace integration
    		I0522 14:57:22.248172    1228 matcher.go:279] got preallocated values for min: 1000010000, max: 1000019999 for uid range in namespace integration
    		I0522 14:57:22.248182    1228 matcher.go:292] got preallocated value for level: s0:c3,c2 for selinux options in namespace integration
    		I0522 14:57:22.248190    1228 matcher.go:322] got preallocated value for groups: 1000010000/10000 in namespace integration
    		I0522 14:57:22.248208    1228 matcher.go:292] got preallocated value for level: s0:c3,c2 for selinux options in namespace integration
    		I0522 14:57:22.248218    1228 matcher.go:292] got preallocated value for level: s0:c3,c2 for selinux options in namespace integration
    		I0522 14:57:22.248234    1228 matcher.go:279] got preallocated values for min: 1000010000, max: 1000019999 for uid range in namespace integration
    		I0522 14:57:22.248243    1228 matcher.go:292] got preallocated value for level: s0:c3,c2 for selinux options in namespace integration
    		I0522 14:57:22.248250    1228 matcher.go:322] got preallocated value for groups: 1000010000/10000 in namespace integration
    		I0522 14:57:22.248263    1228 matcher.go:342] got preallocated value for groups: 1000010000/10000 in namespace integration
    		I0522 14:57:22.248281    1228 matcher.go:279] got preallocated values for min: 1000010000, max: 1000019999 for uid range in namespace integration
    		I0522 14:57:22.248289    1228 matcher.go:292] got preallocated value for level: s0:c3,c2 for selinux options in namespace integration
    		I0522 14:57:22.248297    1228 matcher.go:322] got preallocated value for groups: 1000010000/10000 in namespace integration
    		I0522 14:57:22.248315    1228 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 14:57:22.248360    1228 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I0522 14:57:22.250052    1228 wrap.go:42] POST /api/v1/namespaces/integration/pods?timeout=1m0s: (5.22067ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:54776]
    		I0522 14:57:22.251512    1228 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"testp8mss"}
    		I0522 14:57:22.251642    1228 factory.go:1147] About to try and schedule pod testp8mss
    		I0522 14:57:22.251663    1228 scheduler.go:439] Attempting to schedule pod: integration/testp8mss
    		I0522 14:57:22.251681    1228 scheduler.go:191] Failed to schedule pod: integration/testp8mss
    		I0522 14:57:22.251699    1228 factory.go:1262] Unable to schedule integration testp8mss: no nodes are registered to the cluster; waiting
    		I0522 14:57:22.251735    1228 factory.go:1375] Updating pod condition for integration/testp8mss to (PodScheduled==False)
    		I0522 14:57:22.251794    1228 pvc_protection_controller.go:276] Got event on pod integration/testp8mss
    		I0522 14:57:22.251945    1228 disruption.go:328] addPod called on pod "testp8mss"
    		I0522 14:57:22.251968    1228 disruption.go:403] No PodDisruptionBudgets found for pod testp8mss, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:22.251976    1228 disruption.go:331] No matching pdb for pod "testp8mss"
    		I0522 14:57:22.252004    1228 backoff_utils.go:79] Backing off 1s
    		INFO: 2018/05/22 14:57:22 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc43a68cf60
    		INFO: 2018/05/22 14:57:22 dialing to target with scheme: ""
    		INFO: 2018/05/22 14:57:22 could not get resolver for scheme: ""
    		INFO: 2018/05/22 14:57:22 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/22 14:57:22 balancerWrapper: got update addr from Notify: [{127.0.0.1:16742 <nil>}]
    		INFO: 2018/05/22 14:57:22 ccBalancerWrapper: new subconn: [{127.0.0.1:16742 0  <nil>}]
    		INFO: 2018/05/22 14:57:22 balancerWrapper: handle subconn state change: 0xc42c0c1180, CONNECTING
    		INFO: 2018/05/22 14:57:22 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc43a68cf60
    		I0522 14:57:22.253039    1228 wrap.go:42] POST /api/v1/namespaces/integration/events: (4.908589ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:54122]
    		I0522 14:57:22.258965    1228 pvc_protection_controller.go:276] Got event on pod integration/testp8mss
    		I0522 14:57:22.259042    1228 disruption.go:340] updatePod called on pod "testp8mss"
    		I0522 14:57:22.259055    1228 disruption.go:403] No PodDisruptionBudgets found for pod testp8mss, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:22.259058    1228 wrap.go:42] PUT /api/v1/namespaces/integration/pods/testp8mss/status: (7.09194ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:54122]
    		I0522 14:57:22.259062    1228 disruption.go:343] No matching pdb for pod "testp8mss"
    		I0522 14:57:22.260591    1228 wrap.go:42] POST /api/v1/namespaces/integration/events: (7.046573ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:54122]
    		I0522 14:57:22.262604    1228 factory.go:1147] About to try and schedule pod testp8mss
    		I0522 14:57:22.262619    1228 scheduler.go:439] Attempting to schedule pod: integration/testp8mss
    		I0522 14:57:22.262635    1228 scheduler.go:191] Failed to schedule pod: integration/testp8mss
    		I0522 14:57:22.262650    1228 factory.go:1262] Unable to schedule integration testp8mss: no nodes are registered to the cluster; waiting
    		I0522 14:57:22.262690    1228 factory.go:1375] Updating pod condition for integration/testp8mss to (PodScheduled==False)
    		W0522 14:57:22.262828    1228 factory.go:1304] Request for pod integration/testp8mss already in flight, abandoning
    		INFO: 2018/05/22 14:57:22 balancerWrapper: handle subconn state change: 0xc42c0c1180, READY
    		INFO: 2018/05/22 14:57:22 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc43a68cf60
    		
    --- PASS: TestIntegration/TestAggregator (31.89s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		g request took 158.261361ms, request: POST:https://127.0.0.1:16061/api/v1/namespaces/openshift-infra/secrets
    		I0522 14:57:31.279969    1260 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (1.826025ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:56708]
    		I0522 14:57:31.281154    1260 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/template-instance-controller-token-jxpkg: (978.838µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:56646]
    		I0522 14:57:31.287731    1260 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/template-instance-controller-token-jxpkg: (2.431832ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:56646]
    		I0522 14:57:31.310061    1260 request.go:485] Throttling request took 159.027074ms, request: PUT:https://127.0.0.1:16061/api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-finalizer-controller
    		I0522 14:57:31.312726    1260 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-finalizer-controller: (2.389383ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:56708]
    		I0522 14:57:31.312957    1260 create_dockercfg_secrets.go:441] Creating token secret "template-instance-finalizer-controller-token-4vwnx" for service account openshift-infra/template-instance-finalizer-controller
    		I0522 14:57:31.330429    1260 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.094185ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.342321    1260 request.go:485] Throttling request took 159.204169ms, request: PUT:https://127.0.0.1:16061/api/v1/namespaces/openshift-infra/serviceaccounts/unidling-controller
    		I0522 14:57:31.345072    1260 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/unidling-controller: (2.483176ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:56708]
    		I0522 14:57:31.345302    1260 create_dockercfg_secrets.go:441] Creating token secret "unidling-controller-token-8m9lp" for service account openshift-infra/unidling-controller
    		I0522 14:57:31.374609    1260 request.go:485] Throttling request took 159.078677ms, request: PUT:https://127.0.0.1:16061/api/v1/namespaces/kube-system/serviceaccounts/daemon-set-controller
    		I0522 14:57:31.377250    1260 wrap.go:42] PUT /api/v1/namespaces/kube-system/serviceaccounts/daemon-set-controller: (2.358983ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:56708]
    		I0522 14:57:31.377485    1260 create_dockercfg_secrets.go:441] Creating token secret "daemon-set-controller-token-8vrfz" for service account kube-system/daemon-set-controller
    		I0522 14:57:31.406858    1260 request.go:485] Throttling request took 159.008285ms, request: PUT:https://127.0.0.1:16061/api/v1/namespaces/default/serviceaccounts/builder
    		I0522 14:57:31.409566    1260 wrap.go:42] PUT /api/v1/namespaces/default/serviceaccounts/builder: (2.424613ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:56708]
    		I0522 14:57:31.409801    1260 create_dockercfg_secrets.go:441] Creating token secret "builder-token-8brwt" for service account default/builder
    		I0522 14:57:31.439100    1260 request.go:485] Throttling request took 158.812242ms, request: PUT:https://127.0.0.1:16061/api/v1/namespaces/kube-system/serviceaccounts/builder
    		I0522 14:57:31.442074    1260 wrap.go:42] PUT /api/v1/namespaces/kube-system/serviceaccounts/builder: (2.653496ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:56708]
    		I0522 14:57:31.442317    1260 create_dockercfg_secrets.go:441] Creating token secret "builder-token-9c8hj" for service account kube-system/builder
    		I0522 14:57:31.471343    1260 request.go:485] Throttling request took 158.328075ms, request: POST:https://127.0.0.1:16061/api/v1/namespaces/openshift-infra/secrets
    		I0522 14:57:31.473491    1260 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (1.86929ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:56708]
    		I0522 14:57:31.474761    1260 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/template-instance-finalizer-controller-token-4vwnx: (978.356µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:56646]
    		I0522 14:57:31.481643    1260 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/template-instance-finalizer-controller-token-4vwnx: (2.587447ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:56646]
    		I0522 14:57:31.503536    1260 request.go:485] Throttling request took 158.172655ms, request: POST:https://127.0.0.1:16061/api/v1/namespaces/openshift-infra/secrets
    		I0522 14:57:31.505382    1260 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (1.646648ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:56708]
    		I0522 14:57:31.506713    1260 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/unidling-controller-token-8m9lp: (1.005014ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:56646]
    		I0522 14:57:31.513243    1260 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/unidling-controller-token-8m9lp: (2.236182ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:56646]
    		I0522 14:57:31.535873    1260 request.go:485] Throttling request took 158.337645ms, request: POST:https://127.0.0.1:16061/api/v1/namespaces/kube-system/secrets
    		I0522 14:57:31.537565    1260 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.037803ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.538211    1260 wrap.go:42] POST /api/v1/namespaces/kube-system/secrets: (2.006185ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:56708]
    		I0522 14:57:31.539452    1260 wrap.go:42] GET /api/v1/namespaces/kube-system/secrets/daemon-set-controller-token-8vrfz: (940.279µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:56646]
    		I0522 14:57:31.546336    1260 wrap.go:42] PUT /api/v1/namespaces/kube-system/secrets/daemon-set-controller-token-8vrfz: (2.715175ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:56646]
    		I0522 14:57:31.553534    1260 wrap.go:42] GET /api?timeout=32s: (258.015µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.554282    1260 wrap.go:42] GET /apis?timeout=32s: (329.644µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.555148    1260 wrap.go:42] GET /api/v1?timeout=32s: (391.98µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.555948    1260 wrap.go:42] GET /apis/apiregistration.k8s.io/v1?timeout=32s: (283.654µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.556702    1260 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1?timeout=32s: (280.013µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.557452    1260 wrap.go:42] GET /apis/extensions/v1beta1?timeout=32s: (361.46µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.558301    1260 wrap.go:42] GET /apis/apps/v1?timeout=32s: (349.864µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.559080    1260 wrap.go:42] GET /apis/apps/v1beta2?timeout=32s: (303.07µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.559358    1260 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/admin?timeout=1m0s: (1.664821ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56718]
    		I0522 14:57:31.559838    1260 wrap.go:42] GET /apis/apps/v1beta1?timeout=32s: (307.773µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.560653    1260 wrap.go:42] GET /apis/events.k8s.io/v1beta1?timeout=32s: (312.882µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.561461    1260 wrap.go:42] GET /apis/authentication.k8s.io/v1?timeout=32s: (372.278µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.561473    1260 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/edit?timeout=1m0s: (1.414442ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56718]
    		I0522 14:57:31.562311    1260 wrap.go:42] GET /apis/authentication.k8s.io/v1beta1?timeout=32s: (350.976µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.563021    1260 wrap.go:42] GET /apis/authorization.k8s.io/v1?timeout=32s: (260.685µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.563263    1260 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/view?timeout=1m0s: (1.165314ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56718]
    		I0522 14:57:31.563872    1260 wrap.go:42] GET /apis/authorization.k8s.io/v1beta1?timeout=32s: (405.662µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.564711    1260 wrap.go:42] GET /apis/autoscaling/v1?timeout=32s: (356.79µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.565501    1260 wrap.go:42] GET /apis/autoscaling/v2beta1?timeout=32s: (291.7µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.566289    1260 wrap.go:42] GET /apis/batch/v1?timeout=32s: (307.487µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.567142    1260 wrap.go:42] GET /apis/batch/v1beta1?timeout=32s: (342.153µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.567792    1260 wrap.go:42] GET /api/v1/namespaces/default?timeout=1m0s: (1.640111ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56718]
    		I0522 14:57:31.567966    1260 wrap.go:42] GET /apis/batch/v2alpha1?timeout=32s: (289.37µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.568148    1260 request.go:485] Throttling request took 158.292981ms, request: POST:https://127.0.0.1:16061/api/v1/namespaces/default/secrets
    		I0522 14:57:31.568952    1260 wrap.go:42] GET /apis/certificates.k8s.io/v1beta1?timeout=32s: (378.939µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.569719    1260 wrap.go:42] GET /api/v1/namespaces/default: (878.021µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.569813    1260 wrap.go:42] GET /apis/networking.k8s.io/v1?timeout=32s: (402.211µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.570581    1260 wrap.go:42] GET /apis/project.openshift.io/v1/projects/default?timeout=1m0s: (2.222096ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56718]
    		I0522 14:57:31.570753    1260 wrap.go:42] GET /apis/policy/v1beta1?timeout=32s: (444.38µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.571083    1260 wrap.go:42] POST /api/v1/namespaces/default/secrets: (2.594935ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:56708]
    		I0522 14:57:31.572138    1260 wrap.go:42] GET /apis/authorization.openshift.io/v1?timeout=32s: (386.867µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.572934    1260 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1?timeout=32s: (391.911µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.573138    1260 wrap.go:42] GET /api/v1/namespaces/default/secrets/builder-token-8brwt: (1.101817ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:56646]
    		I0522 14:57:31.573144    1260 wrap.go:42] GET /api/v1/namespaces/default: (1.183822ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.573501    1260 wrap.go:42] GET /apis/project.openshift.io/v1/projects/default?timeout=1m0s: (2.309829ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56718]
    		I0522 14:57:31.573765    1260 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1beta1?timeout=32s: (388.247µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.574680    1260 wrap.go:42] GET /apis/storage.k8s.io/v1?timeout=32s: (375.569µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.575613    1260 wrap.go:42] GET /apis/storage.k8s.io/v1beta1?timeout=32s: (426.706µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.576346    1260 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1/apiservices/v1.?timeout=1m0s: (2.372644ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56718]
    		I0522 14:57:31.576450    1260 wrap.go:42] GET /apis/admissionregistration.k8s.io/v1beta1?timeout=32s: (357.566µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.578909    1260 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1/apiservices/v1.project.openshift.io?timeout=1m0s: (1.150763ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56718]
    		I0522 14:57:31.579892    1260 wrap.go:42] GET /apis/apiextensions.k8s.io/v1beta1?timeout=32s: (2.622684ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.580456    1260 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1/apiservices/v1beta1.rbac.authorization.k8s.io?timeout=1m0s: (1.062528ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56718]
    		I0522 14:57:31.580855    1260 wrap.go:42] GET /apis/apps.openshift.io/v1?timeout=32s: (436.477µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.581786    1260 wrap.go:42] GET /apis/build.openshift.io/v1?timeout=32s: (438.184µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		INFO: 2018/05/22 14:57:31 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc434fbcf00
    		INFO: 2018/05/22 14:57:31 dialing to target with scheme: ""
    		INFO: 2018/05/22 14:57:31 could not get resolver for scheme: ""
    		INFO: 2018/05/22 14:57:31 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/22 14:57:31 balancerWrapper: got update addr from Notify: [{127.0.0.1:14771 <nil>}]
    		INFO: 2018/05/22 14:57:31 ccBalancerWrapper: new subconn: [{127.0.0.1:14771 0  <nil>}]
    		INFO: 2018/05/22 14:57:31 balancerWrapper: handle subconn state change: 0xc428aa1ea0, CONNECTING
    		INFO: 2018/05/22 14:57:31 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc434fbcf00
    		I0522 14:57:31.582665    1260 wrap.go:42] GET /apis/image.openshift.io/v1?timeout=32s: (434.344µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.583564    1260 wrap.go:42] GET /apis/network.openshift.io/v1?timeout=32s: (423.04µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.583816    1260 wrap.go:42] PUT /api/v1/namespaces/default/secrets/builder-token-8brwt: (4.546414ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:56646]
    		I0522 14:57:31.584713    1260 wrap.go:42] GET /apis/oauth.openshift.io/v1?timeout=32s: (386.706µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.585405    1260 wrap.go:42] GET /apis/project.openshift.io/v1?timeout=32s: (316.799µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.586196    1260 wrap.go:42] GET /apis/quota.openshift.io/v1?timeout=32s: (359.557µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.586876    1260 wrap.go:42] GET /apis/route.openshift.io/v1?timeout=32s: (298.957µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.587646    1260 wrap.go:42] GET /apis/security.openshift.io/v1?timeout=32s: (361.463µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.588226    1260 wrap.go:42] GET /apis/template.openshift.io/v1?timeout=32s: (231.321µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.589208    1260 wrap.go:42] GET /apis/user.openshift.io/v1?timeout=32s: (329.584µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.591276    1260 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.97591ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.591316    1260 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.534559ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.591409    1260 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.07637ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		I0522 14:57:31.591444    1260 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.715168ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:56646]
    		INFO: 2018/05/22 14:57:31 balancerWrapper: handle subconn state change: 0xc428aa1ea0, READY
    		INFO: 2018/05/22 14:57:31 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc434fbcf00
    		INFO: 2018/05/22 14:57:31 balancerWrapper: got update addr from Notify: [{127.0.0.1:14771 <nil>}]
    		
    --- PASS: TestIntegration/TestAllowedSCCViaRBAC (42.39s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		in namespace project1
    		I0522 14:57:31.682702    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 14:57:31.682708    1243 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 14:57:31.682721    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 14:57:31.682729    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 14:57:31.682738    1243 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0522 14:57:31.682743    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 14:57:31.682748    1243 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 14:57:31.682755    1243 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 14:57:31.682767    1243 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0522 14:57:31.682772    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 14:57:31.682777    1243 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 14:57:31.682790    1243 admission.go:217] validating pod test3 (generate: ) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 14:57:31.682918    1243 admission.go:170] pod test3 (generate: ) validated against provider hostaccess
    		I0522 14:57:31.684746    1243 wrap.go:42] POST /api/v1/namespaces/project1/pods?timeout=1m0s: (5.046491ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52630]
    		I0522 14:57:31.684986    1243 factory.go:1147] About to try and schedule pod test3
    		I0522 14:57:31.685001    1243 scheduler.go:439] Attempting to schedule pod: project1/test3
    		I0522 14:57:31.685019    1243 scheduler.go:191] Failed to schedule pod: project1/test3
    		I0522 14:57:31.685037    1243 pvc_protection_controller.go:276] Got event on pod project1/test3
    		I0522 14:57:31.685066    1243 disruption.go:328] addPod called on pod "test3"
    		I0522 14:57:31.685074    1243 factory.go:1262] Unable to schedule project1 test3: no nodes are registered to the cluster; waiting
    		I0522 14:57:31.685077    1243 disruption.go:403] No PodDisruptionBudgets found for pod test3, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:31.685084    1243 disruption.go:331] No matching pdb for pod "test3"
    		I0522 14:57:31.685087    1243 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"project1", Name:"test3"}
    		I0522 14:57:31.685113    1243 factory.go:1375] Updating pod condition for project1/test3 to (PodScheduled==False)
    		I0522 14:57:31.685145    1243 backoff_utils.go:79] Backing off 1s
    		I0522 14:57:31.685968    1243 admission.go:97] getting security context constraints for pod test4 (generate: ) in namespace project2 with user info &{user1 6f97dbff-5dd0-11e8-a9f3-0242ac110002 [system:authenticated:oauth system:authenticated] map[scopes.authorization.openshift.io:[]]}
    		I0522 14:57:31.686528    1243 admission.go:108] getting security context constraints for pod test4 (generate: ) with service account info &{system:serviceaccount:project2:default  [system:serviceaccounts system:serviceaccounts:project2] map[]}
    		I0522 14:57:31.688308    1243 wrap.go:42] PUT /api/v1/namespaces/project1/pods/test3/status: (2.863437ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:52364]
    		I0522 14:57:31.688362    1243 wrap.go:42] GET /api/v1/namespaces/project2: (1.181146ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52364]
    		I0522 14:57:31.688659    1243 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I0522 14:57:31.688677    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 14:57:31.688687    1243 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0522 14:57:31.688706    1243 admission.go:217] validating pod test4 (generate: ) against providers restricted
    		I0522 14:57:31.688760    1243 admission.go:179] unable to validate pod test4 (generate: ) against any security context constraint: [provider restricted: .spec.securityContext.hostPID: Invalid value: true: Host PID is not allowed to be used spec.containers[0].securityContext.hostPID: Invalid value: true: Host PID is not allowed to be used]
    		I0522 14:57:31.688892    1243 wrap.go:42] POST /api/v1/namespaces/project2/pods?timeout=1m0s: (3.49282ms) 403 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52630]
    		I0522 14:57:31.689142    1243 factory.go:1147] About to try and schedule pod test3
    		I0522 14:57:31.689156    1243 scheduler.go:439] Attempting to schedule pod: project1/test3
    		I0522 14:57:31.689173    1243 scheduler.go:191] Failed to schedule pod: project1/test3
    		I0522 14:57:31.689191    1243 factory.go:1262] Unable to schedule project1 test3: no nodes are registered to the cluster; waiting
    		I0522 14:57:31.689220    1243 factory.go:1375] Updating pod condition for project1/test3 to (PodScheduled==False)
    		I0522 14:57:31.689293    1243 pvc_protection_controller.go:276] Got event on pod project1/test3
    		I0522 14:57:31.689326    1243 disruption.go:340] updatePod called on pod "test3"
    		I0522 14:57:31.689338    1243 disruption.go:403] No PodDisruptionBudgets found for pod test3, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:31.689345    1243 disruption.go:343] No matching pdb for pod "test3"
    		W0522 14:57:31.689400    1243 factory.go:1304] Request for pod project1/test3 already in flight, abandoning
    		I0522 14:57:31.689470    1243 wrap.go:42] POST /api/v1/namespaces/project1/events: (3.96982ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:52364]
    		I0522 14:57:31.689930    1243 admission.go:97] getting security context constraints for pod test5 (generate: ) in namespace project1 with user info &{user2 6fa2a15f-5dd0-11e8-a9f3-0242ac110002 [system:authenticated:oauth system:authenticated] map[scopes.authorization.openshift.io:[]]}
    		I0522 14:57:31.690132    1243 admission.go:108] getting security context constraints for pod test5 (generate: ) with service account info &{system:serviceaccount:project1:default  [system:serviceaccounts system:serviceaccounts:project1] map[]}
    		I0522 14:57:31.691950    1243 wrap.go:42] GET /api/v1/namespaces/project1: (986.403µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52364]
    		I0522 14:57:31.692145    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 14:57:31.692198    1243 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0522 14:57:31.692209    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 14:57:31.692218    1243 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 14:57:31.692235    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 14:57:31.692248    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 14:57:31.692265    1243 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0522 14:57:31.692274    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 14:57:31.692282    1243 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 14:57:31.692295    1243 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 14:57:31.692316    1243 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0522 14:57:31.692324    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 14:57:31.692332    1243 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 14:57:31.692349    1243 admission.go:217] validating pod test5 (generate: ) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 14:57:31.692506    1243 admission.go:170] pod test5 (generate: ) validated against provider hostaccess
    		I0522 14:57:31.693723    1243 factory.go:1147] About to try and schedule pod test5
    		I0522 14:57:31.693743    1243 scheduler.go:439] Attempting to schedule pod: project1/test5
    		I0522 14:57:31.693759    1243 scheduler.go:191] Failed to schedule pod: project1/test5
    		I0522 14:57:31.693784    1243 factory.go:1262] Unable to schedule project1 test5: no nodes are registered to the cluster; waiting
    		I0522 14:57:31.693786    1243 wrap.go:42] POST /api/v1/namespaces/project1/pods?timeout=1m0s: (4.371579ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52630]
    		I0522 14:57:31.693812    1243 factory.go:1375] Updating pod condition for project1/test5 to (PodScheduled==False)
    		I0522 14:57:31.693969    1243 backoff_utils.go:79] Backing off 1s
    		I0522 14:57:31.694330    1243 pvc_protection_controller.go:276] Got event on pod project1/test5
    		I0522 14:57:31.694363    1243 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"project1", Name:"test5"}
    		I0522 14:57:31.694484    1243 disruption.go:328] addPod called on pod "test5"
    		I0522 14:57:31.694510    1243 disruption.go:403] No PodDisruptionBudgets found for pod test5, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:31.694525    1243 disruption.go:331] No matching pdb for pod "test5"
    		I0522 14:57:31.694841    1243 admission.go:97] getting security context constraints for pod test6 (generate: ) in namespace project2 with user info &{user2 6fa2a15f-5dd0-11e8-a9f3-0242ac110002 [system:authenticated:oauth system:authenticated] map[scopes.authorization.openshift.io:[]]}
    		I0522 14:57:31.695136    1243 admission.go:108] getting security context constraints for pod test6 (generate: ) with service account info &{system:serviceaccount:project2:default  [system:serviceaccounts system:serviceaccounts:project2] map[]}
    		I0522 14:57:31.695809    1243 wrap.go:42] PATCH /api/v1/namespaces/project1/events/test3.1530ff5db9569bd6: (5.456418ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:52364]
    		I0522 14:57:31.699140    1243 wrap.go:42] GET /api/v1/namespaces/project2: (2.410701ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52364]
    		I0522 14:57:31.699328    1243 wrap.go:42] PUT /api/v1/namespaces/project1/pods/test5/status: (5.24373ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:52364]
    		I0522 14:57:31.699419    1243 pvc_protection_controller.go:276] Got event on pod project1/test5
    		I0522 14:57:31.699451    1243 disruption.go:340] updatePod called on pod "test5"
    		I0522 14:57:31.699464    1243 disruption.go:403] No PodDisruptionBudgets found for pod test5, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:31.699472    1243 disruption.go:343] No matching pdb for pod "test5"
    		I0522 14:57:31.699329    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 14:57:31.699774    1243 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I0522 14:57:31.699785    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 14:57:31.699794    1243 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0522 14:57:31.699813    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 14:57:31.699825    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 14:57:31.699841    1243 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I0522 14:57:31.699849    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 14:57:31.699857    1243 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0522 14:57:31.699870    1243 matcher.go:342] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0522 14:57:31.699893    1243 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I0522 14:57:31.699904    1243 factory.go:1147] About to try and schedule pod test5
    		I0522 14:57:31.699910    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 14:57:31.699914    1243 scheduler.go:439] Attempting to schedule pod: project1/test5
    		I0522 14:57:31.699919    1243 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0522 14:57:31.699948    1243 admission.go:217] validating pod test6 (generate: ) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 14:57:31.700092    1243 admission.go:170] pod test6 (generate: ) validated against provider hostaccess
    		I0522 14:57:31.699927    1243 scheduler.go:191] Failed to schedule pod: project1/test5
    		I0522 14:57:31.700223    1243 factory.go:1262] Unable to schedule project1 test5: no nodes are registered to the cluster; waiting
    		I0522 14:57:31.700262    1243 factory.go:1375] Updating pod condition for project1/test5 to (PodScheduled==False)
    		W0522 14:57:31.700304    1243 factory.go:1304] Request for pod project1/test5 already in flight, abandoning
    		I0522 14:57:31.701034    1243 wrap.go:42] POST /api/v1/namespaces/project1/events: (4.039998ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:52364]
    		I0522 14:57:31.702033    1243 wrap.go:42] POST /api/v1/namespaces/project2/pods?timeout=1m0s: (7.725477ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52630]
    		I0522 14:57:31.702369    1243 factory.go:1147] About to try and schedule pod test6
    		I0522 14:57:31.702383    1243 scheduler.go:439] Attempting to schedule pod: project2/test6
    		I0522 14:57:31.702399    1243 scheduler.go:191] Failed to schedule pod: project2/test6
    		I0522 14:57:31.702419    1243 factory.go:1262] Unable to schedule project2 test6: no nodes are registered to the cluster; waiting
    		I0522 14:57:31.702448    1243 factory.go:1375] Updating pod condition for project2/test6 to (PodScheduled==False)
    		I0522 14:57:31.702665    1243 pvc_protection_controller.go:276] Got event on pod project2/test6
    		I0522 14:57:31.702693    1243 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"project2", Name:"test6"}
    		I0522 14:57:31.702722    1243 disruption.go:328] addPod called on pod "test6"
    		I0522 14:57:31.702733    1243 disruption.go:403] No PodDisruptionBudgets found for pod test6, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:31.702741    1243 disruption.go:331] No matching pdb for pod "test6"
    		I0522 14:57:31.702790    1243 backoff_utils.go:79] Backing off 1s
    		I0522 14:57:31.703401    1243 controller.go:537] quota admission added evaluator for: {security.openshift.io podsecuritypolicyselfsubjectreviews}
    		I0522 14:57:31.704660    1243 wrap.go:42] PUT /api/v1/namespaces/project2/pods/test6/status: (1.743252ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:52364]
    		I0522 14:57:31.705063    1243 pvc_protection_controller.go:276] Got event on pod project2/test6
    		I0522 14:57:31.705091    1243 disruption.go:340] updatePod called on pod "test6"
    		I0522 14:57:31.705102    1243 disruption.go:403] No PodDisruptionBudgets found for pod test6, PodDisruptionBudget controller will avoid syncing.
    		I0522 14:57:31.705119    1243 disruption.go:343] No matching pdb for pod "test6"
    		I0522 14:57:31.705372    1243 factory.go:1147] About to try and schedule pod test6
    		I0522 14:57:31.705390    1243 scheduler.go:439] Attempting to schedule pod: project2/test6
    		I0522 14:57:31.705406    1243 scheduler.go:191] Failed to schedule pod: project2/test6
    		I0522 14:57:31.705424    1243 factory.go:1262] Unable to schedule project2 test6: no nodes are registered to the cluster; waiting
    		I0522 14:57:31.705456    1243 factory.go:1375] Updating pod condition for project2/test6 to (PodScheduled==False)
    		W0522 14:57:31.705500    1243 factory.go:1304] Request for pod project2/test6 already in flight, abandoning
    		I0522 14:57:31.705674    1243 wrap.go:42] GET /api/v1/namespaces/project1: (1.654022ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52364]
    		I0522 14:57:31.705823    1243 wrap.go:42] PATCH /api/v1/namespaces/project1/events/test5.1530ff5db9db7040: (4.1863ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:52364]
    		I0522 14:57:31.706012    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 14:57:31.706284    1243 wrap.go:42] POST /apis/security.openshift.io/v1/namespaces/project1/podsecuritypolicyselfsubjectreviews?timeout=1m0s: (3.40142ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52630]
    		I0522 14:57:31.709049    1243 wrap.go:42] GET /api/v1/namespaces/project2: (1.011182ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52364]
    		I0522 14:57:31.709275    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 14:57:31.709340    1243 wrap.go:42] POST /api/v1/namespaces/project2/events: (2.965427ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:52364]
    		I0522 14:57:31.709452    1243 wrap.go:42] POST /apis/security.openshift.io/v1/namespaces/project2/podsecuritypolicyselfsubjectreviews?timeout=1m0s: (2.44046ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:52630]
    		INFO: 2018/05/22 14:57:31 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc43b1800c0
    		INFO: 2018/05/22 14:57:31 dialing to target with scheme: ""
    		INFO: 2018/05/22 14:57:31 could not get resolver for scheme: ""
    		INFO: 2018/05/22 14:57:31 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/22 14:57:31 balancerWrapper: got update addr from Notify: [{127.0.0.1:23479 <nil>}]
    		INFO: 2018/05/22 14:57:31 ccBalancerWrapper: new subconn: [{127.0.0.1:23479 0  <nil>}]
    		INFO: 2018/05/22 14:57:31 balancerWrapper: handle subconn state change: 0xc430b7c980, CONNECTING
    		INFO: 2018/05/22 14:57:31 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc43b1800c0
    		I0522 14:57:31.713231    1243 wrap.go:42] PATCH /api/v1/namespaces/project2/events/test6.1530ff5dba5f31e6: (3.255955ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:52364]
    		INFO: 2018/05/22 14:57:31 balancerWrapper: handle subconn state change: 0xc430b7c980, READY
    		INFO: 2018/05/22 14:57:31 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc43b1800c0
    		INFO: 2018/05/22 14:57:31 balancerWrapper: got update addr from Notify: [{127.0.0.1:23479 <nil>}]
    		
FAIL
exit status 1
FAIL	github.com/openshift/origin/test/integration/runner	1963.773s
[INFO] [14:57:32+0000] jUnit XML report placed at _output/scripts/test-integration/artifacts/gotest_report_niAgv.xml
Of 203 tests executed in 1963.773s, 199 succeeded, 4 failed, and 0 were skipped.

In suite "github.com/openshift/origin/test/integration/runner", test case "TestIntegration" failed:
runner_test.go:84: using existing binary

In suite "github.com/openshift/origin/test/integration/runner", test case "TestIntegration/TestConcurrentBuildImageChangeTriggerControllers" failed:
runner_test.go:129: FAILED TestConcurrentBuildImageChangeTriggerControllers, retrying:

controllers.go:532: expected watch event type ADDED, got 


In suite "github.com/openshift/origin/test/integration/runner", test case "TestIntegration/TestImageStreamImportScheduled" failed:
runner_test.go:129: FAILED TestImageStreamImportScheduled, retrying:

imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 1
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 2
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 3
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 4
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 5
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 6
imageimporter_test.go:771: unexpected object: <nil>


In suite "github.com/openshift/origin/test/integration/runner", test case "TestIntegration/TestOAuthServiceAccountClientEvent" failed:
runner_test.go:129: FAILED TestOAuthServiceAccountClientEvent, retrying:

oauth_serviceaccount_client_events_test.go:141: test-good-url: annotationPrefix serviceaccounts.openshift.io/oauth-redirecturi.one, annotation /oauthcallback
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:22823/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=http%!A(MISSING)%!F(MISSING)%!F(MISSING)127.0.0.1%!A(MISSING)46165%!F(MISSING)oauthcallback&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:22823/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=http%!A(MISSING)%!F(MISSING)%!F(MISSING)127.0.0.1%!A(MISSING)46165%!F(MISSING)oauthcallback&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:362: 302 Location: https://127.0.0.1:22823/oauth/authorize/approve?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=http%!A(MISSING)%!F(MISSING)%!F(MISSING)127.0.0.1%!A(MISSING)46165%!F(MISSING)oauthcallback&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project&then=..%!F(MISSING)authorize%!F(MISSING)client_id%!D(MISSING)system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default%!r(MISSING)edirect_uri%!D(MISSING)http%!A(MISSING)%!F(MISSING)%!F(MISSING)127.0.0.1%!A(MISSING)46165%!F(MISSING)oauthcallback%!r(MISSING)esponse_type%!D(MISSING)code%!s(MISSING)cope%!D(MISSING)user%!A(MISSING)info%!B(MISSING)role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:371: POST https://127.0.0.1:22823/oauth/authorize/approve
oauth_serviceaccount_client_events_test.go:362: 302 Location: https://127.0.0.1:22823/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=http%!A(MISSING)%!F(MISSING)%!F(MISSING)127.0.0.1%!A(MISSING)46165%!F(MISSING)oauthcallback&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:362: 302 Location: http://127.0.0.1:46165/oauthcallback?code=rqEympzhDEovO-KK8uyCWGjafVa6dOjacTI7Ed8jyAY&state=
oauth_serviceaccount_client_events_test.go:141: test-bad-url-parse: annotationPrefix serviceaccounts.openshift.io/oauth-redirecturi.one, annotation ::
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:22823/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=%!A(MISSING)%!A(MISSING)&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:396: Bad Request: HTTP/2.0 400 Bad Request
Content-Length: 155
Cache-Control: no-cache, no-store, max-age=0, must-revalidate
Content-Type: application/json
Date: Tue, 22 May 2018 14:37:03 GMT
Expires: Fri, 01 Jan 1990 00:00:00 GMT
Pragma: no-cache

{"error":"server_error","error_description":"The authorization server encountered an unexpected condition that prevented it from fulfilling the request."}
oauth_serviceaccount_client_events_test.go:141: test-bad-redirect-type-parse: annotationPrefix serviceaccounts.openshift.io/oauth-redirectreference.1, annotation {asdf":"adsf"}
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:22823/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=%!B(MISSING)asdf%3A%!a(MISSING)dsf%7D&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:396: Bad Request: HTTP/2.0 400 Bad Request
Content-Length: 155
Cache-Control: no-cache, no-store, max-age=0, must-revalidate
Content-Type: application/json
Date: Tue, 22 May 2018 14:37:03 GMT
Expires: Fri, 01 Jan 1990 00:00:00 GMT
Pragma: no-cache

{"error":"server_error","error_description":"The authorization server encountered an unexpected condition that prevented it from fulfilling the request."}
oauth_serviceaccount_client_events_test.go:141: test-bad-redirect-route-not-found: annotationPrefix serviceaccounts.openshift.io/oauth-redirectreference.1, annotation {"kind":"OAuthRedirectReference","apiVersion":"oauth.openshift.io/v1","metadata":{"creationTimestamp":null},"reference":{"group":"route.openshift.io","kind":"Route","name":"route1"}}
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:22823/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=%!B(MISSING)%!k(MISSING)ind%3A%!O(MISSING)AuthRedirectReference%2C%!a(MISSING)piVersion%3A%!o(MISSING)auth.openshift.io%!F(MISSING)v1%2C%!m(MISSING)etadata%3A%!B(MISSING)%!c(MISSING)reationTimestamp%3Anull%!D(MISSING)%!C(MISSING)%!r(MISSING)eference%3A%!B(MISSING)%!g(MISSING)roup%3A%!r(MISSING)oute.openshift.io%2C%!k(MISSING)ind%3A%!R(MISSING)oute%2C%!n(MISSING)ame%3A%!r(MISSING)oute1%7D%!D(MISSING)%!A(MISSING)&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:396: Bad Request: HTTP/2.0 400 Bad Request
Content-Length: 155
Cache-Control: no-cache, no-store, max-age=0, must-revalidate
Content-Type: application/json
Date: Tue, 22 May 2018 14:37:03 GMT
Expires: Fri, 01 Jan 1990 00:00:00 GMT
Pragma: no-cache

{"error":"server_error","error_description":"The authorization server encountered an unexpected condition that prevented it from fulfilling the request."}
oauth_serviceaccount_client_events_test.go:141: test-bad-redirect-route-wrong-group: annotationPrefix serviceaccounts.openshift.io/oauth-redirectreference.1, annotation {"kind":"OAuthRedirectReference","apiVersion":"oauth.openshift.io/v1","metadata":{"creationTimestamp":null},"reference":{"group":"foo","kind":"Route","name":"route1"}}
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:22823/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=%!B(MISSING)%!k(MISSING)ind%3A%!O(MISSING)AuthRedirectReference%2C%!a(MISSING)piVersion%3A%!o(MISSING)auth.openshift.io%!F(MISSING)v1%2C%!m(MISSING)etadata%3A%!B(MISSING)%!c(MISSING)reationTimestamp%3Anull%!D(MISSING)%!C(MISSING)%!r(MISSING)eference%3A%!B(MISSING)%!g(MISSING)roup%3A%!f(MISSING)oo%2C%!k(MISSING)ind%3A%!R(MISSING)oute%2C%!n(MISSING)ame%3A%!r(MISSING)oute1%7D%!D(MISSING)%!A(MISSING)&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:396: Bad Request: HTTP/2.0 400 Bad Request
Content-Length: 155
Cache-Control: no-cache, no-store, max-age=0, must-revalidate
Content-Type: application/json
Date: Tue, 22 May 2018 14:37:03 GMT
Expires: Fri, 01 Jan 1990 00:00:00 GMT
Pragma: no-cache

{"error":"server_error","error_description":"The authorization server encountered an unexpected condition that prevented it from fulfilling the request."}
oauth_serviceaccount_client_events_test.go:141: test-bad-url: annotationPrefix serviceaccounts.openshift.io/oauth-redirecturi.one, annotation foo:foo
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:22823/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=foo%!A(MISSING)foo&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:396: Bad Request: HTTP/2.0 400 Bad Request
Content-Length: 155
Cache-Control: no-cache, no-store, max-age=0, must-revalidate
Content-Type: application/json
Date: Tue, 22 May 2018 14:37:03 GMT
Expires: Fri, 01 Jan 1990 00:00:00 GMT
Pragma: no-cache

{"error":"server_error","error_description":"The authorization server encountered an unexpected condition that prevented it from fulfilling the request."}
oauth_serviceaccount_client_events_test.go:163: test-bad-url: expected 1 events, found 0
[ERROR] [14:57:33+0000] hack/test-go.sh exited with code 1 after 00h 32m 46s
[ERROR] [14:57:33+0000] PID 1249: hack/test-integration.sh:18: `COVERAGE_SPEC=" " DETECT_RACES=false TMPDIR="${BASETMPDIR}" TIMEOUT=45m GOTEST_FLAGS="${gotest_flags}" "${OS_ROOT}/hack/test-go.sh" "test/integration/runner"` exited with status 1.
[INFO] [14:57:33+0000] 		Stack Trace: 
[INFO] [14:57:33+0000] 		  1: hack/test-integration.sh:18: `COVERAGE_SPEC=" " DETECT_RACES=false TMPDIR="${BASETMPDIR}" TIMEOUT=45m GOTEST_FLAGS="${gotest_flags}" "${OS_ROOT}/hack/test-go.sh" "test/integration/runner"`
[INFO] [14:57:33+0000]   Exiting with code 1.
make: *** [test-integration] Error 1
[WARNING] [14:57:34+0000] Copying _output/local/releases from the container failed!
[WARNING] [14:57:34+0000] Error response from daemon: lstat /var/lib/docker/overlay2/2ff7be8f41a87da6ab8e26276117adaa2e99601cbc09d759c0cd1dafacc70840/merged/go/src/github.com/openshift/origin/_output/local/releases: no such file or directory
[ERROR] [14:57:34+0000] PID 10820: hack/lib/build/environment.sh:172: `return "${exitcode}"` exited with status 2.
[INFO] [14:57:34+0000] 		Stack Trace: 
[INFO] [14:57:34+0000] 		  1: hack/lib/build/environment.sh:172: `return "${exitcode}"`
[INFO] [14:57:34+0000] 		  2: hack/lib/build/environment.sh:281: os::build::environment::withsource
[INFO] [14:57:34+0000] 		  3: hack/env:42: os::build::environment::run
[INFO] [14:57:34+0000]   Exiting with code 2.
++ export status=FAILURE
++ status=FAILURE
+ set +o xtrace
########## FINISHED STAGE: FAILURE: RUN INTEGRATION TESTS [00h 44m 28s] ##########
Build step 'Execute shell' marked build as failure
[PostBuildScript] - Executing post build scripts.
[workspace@5] $ /bin/bash /tmp/jenkins8979509408606200961.sh
########## STARTING STAGE: DOWNLOAD ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/gathered
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/gathered
+ mkdir -p /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/gathered
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo stat /data/src/github.com/openshift/origin/_output/scripts
  File: ‘/data/src/github.com/openshift/origin/_output/scripts’
  Size: 61        	Blocks: 0          IO Block: 4096   directory
Device: ca02h/51714d	Inode: 105085521   Links: 5
Access: (2755/drwxr-sr-x)  Uid: ( 1001/  origin)   Gid: ( 1003/origin-git)
Context: unconfined_u:object_r:container_file_t:s0
Access: 1970-01-01 00:00:00.000000000 +0000
Modify: 2018-05-22 14:14:13.000000000 +0000
Change: 2018-05-22 14:57:34.485372631 +0000
 Birth: -
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod -R o+rX /data/src/github.com/openshift/origin/_output/scripts
+ scp -r -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel:/data/src/github.com/openshift/origin/_output/scripts /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/gathered
+ tree /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/gathered
/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/gathered
└── scripts
    ├── shell
    │   ├── artifacts
    │   ├── logs
    │   │   ├── 3c90aaea435f9c0fa32804533bdaa35e42350b4cc7c2b60d095ed2cd9517d62a.json
    │   │   └── scripts.log
    │   └── openshift.local.home
    ├── test-integration
    │   ├── artifacts
    │   │   ├── gotest_report_niAgv
    │   │   └── gotest_report_niAgv.xml
    │   ├── logs
    │   │   ├── raw_test_output.log
    │   │   ├── scripts.log
    │   │   └── test-go-err.log
    │   └── openshift.local.home
    └── test-tools
        ├── artifacts
        ├── logs
        │   ├── raw_test_output.log
        │   └── scripts.log
        └── openshift.local.home

13 directories, 9 files
+ exit 0
[workspace@5] $ /bin/bash /tmp/jenkins2133628339626679279.sh
########## STARTING STAGE: GENERATE ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/generated
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/generated
+ mkdir /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/generated
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a 2>&1'
  WARNING: You're not using the default seccomp profile
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo cat /etc/sysconfig/docker /etc/sysconfig/docker-network /etc/sysconfig/docker-storage /etc/sysconfig/docker-storage-setup /etc/systemd/system/docker.service 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'oc get --raw /metrics --server=https://$( uname --nodename ):10250 --config=/etc/origin/master/admin.kubeconfig 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'oc get --raw /metrics --config=/etc/origin/master/admin.kubeconfig 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo df -T -h && sudo pvs && sudo vgs && sudo lvs && sudo findmnt --all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo yum list installed 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl --dmesg --no-pager --all --lines=all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl _PID=1 --no-pager --all --lines=all 2>&1'
+ tree /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/generated
/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/generated
├── avc_denials.log
├── containers.log
├── dmesg.log
├── docker.config
├── docker.info
├── filesystem.info
├── installed_packages.log
├── master-metrics.log
├── node-metrics.log
└── pid1.journal

0 directories, 10 files
+ exit 0
[workspace@5] $ /bin/bash /tmp/jenkins1664019974491882616.sh
########## STARTING STAGE: FETCH SYSTEMD JOURNALS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/journals
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/journals
+ mkdir /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/journals
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit docker.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
+ tree /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/journals
/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/artifacts/journals
├── dnsmasq.service
├── docker.service
└── systemd-journald.service

0 directories, 3 files
+ exit 0
[workspace@5] $ /bin/bash /tmp/jenkins3255425121048336142.sh
########## STARTING STAGE: ASSEMBLE GCS OUTPUT ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config
+ trap 'exit 0' EXIT
+ mkdir -p gcs/artifacts gcs/artifacts/generated gcs/artifacts/journals gcs/artifacts/gathered
++ python -c 'import json; import urllib; print json.load(urllib.urlopen('\''https://ci.openshift.redhat.com/jenkins/job/test_pull_request_origin_integration/18027/api/json'\''))['\''result'\'']'
+ result=FAILURE
+ cat
++ date +%s
+ cat /var/lib/jenkins/jobs/test_pull_request_origin_integration/builds/18027/log
+ cp artifacts/generated/avc_denials.log artifacts/generated/containers.log artifacts/generated/dmesg.log artifacts/generated/docker.config artifacts/generated/docker.info artifacts/generated/filesystem.info artifacts/generated/installed_packages.log artifacts/generated/master-metrics.log artifacts/generated/node-metrics.log artifacts/generated/pid1.journal gcs/artifacts/generated/
+ cp artifacts/journals/dnsmasq.service artifacts/journals/docker.service artifacts/journals/systemd-journald.service gcs/artifacts/journals/
+ cp -r artifacts/gathered/scripts gcs/artifacts/
++ pwd
+ scp -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config -r /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/gcs openshiftdevel:/data
+ scp -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.config/gcloud/gcs-publisher-credentials.json openshiftdevel:/data/credentials.json
+ exit 0
[workspace@5] $ /bin/bash /tmp/jenkins3282670375356456706.sh
########## STARTING STAGE: PUSH THE ARTIFACTS AND METADATA ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config
++ mktemp
+ script=/tmp/tmp.N8jkC0QeHb
+ cat
+ chmod +x /tmp/tmp.N8jkC0QeHb
+ scp -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.N8jkC0QeHb openshiftdevel:/tmp/tmp.N8jkC0QeHb
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 300 /tmp/tmp.N8jkC0QeHb"'
+ cd /home/origin
+ trap 'exit 0' EXIT
+ [[ -n {"type":"presubmit","job":"test_pull_request_origin_integration","buildid":"c87b1d8d-5dc9-11e8-ae4a-0a58ac100556","refs":{"org":"openshift","repo":"origin","base_ref":"master","base_sha":"b89edde21a383ff0ab88317b8d74aedf892239cf","pulls":[{"number":19759,"author":"wozniakjan","sha":"dd19bd96ab7d2816baa1396104877b3b7ec36276"}]}} ]]
++ jq --compact-output .buildid
+ [[ "c87b1d8d-5dc9-11e8-ae4a-0a58ac100556" =~ ^"[0-9]+"$ ]]
+ echo 'Using BUILD_NUMBER'
Using BUILD_NUMBER
++ jq --compact-output '.buildid |= "18027"'
+ JOB_SPEC='{"type":"presubmit","job":"test_pull_request_origin_integration","buildid":"18027","refs":{"org":"openshift","repo":"origin","base_ref":"master","base_sha":"b89edde21a383ff0ab88317b8d74aedf892239cf","pulls":[{"number":19759,"author":"wozniakjan","sha":"dd19bd96ab7d2816baa1396104877b3b7ec36276"}]}}'
+ docker run -e 'JOB_SPEC={"type":"presubmit","job":"test_pull_request_origin_integration","buildid":"18027","refs":{"org":"openshift","repo":"origin","base_ref":"master","base_sha":"b89edde21a383ff0ab88317b8d74aedf892239cf","pulls":[{"number":19759,"author":"wozniakjan","sha":"dd19bd96ab7d2816baa1396104877b3b7ec36276"}]}}' -v /data:/data:z registry.svc.ci.openshift.org/ci/gcsupload:latest --dry-run=false --gcs-path=gs://origin-ci-test --gcs-credentials-file=/data/credentials.json --path-strategy=single --default-org=openshift --default-repo=origin /data/gcs/artifacts /data/gcs/build-log.txt /data/gcs/finished.json
Unable to find image 'registry.svc.ci.openshift.org/ci/gcsupload:latest' locally
Trying to pull repository registry.svc.ci.openshift.org/ci/gcsupload ... 
latest: Pulling from registry.svc.ci.openshift.org/ci/gcsupload
6d987f6f4279: Already exists
4cccebe844ee: Already exists
cafcba51f636: Pulling fs layer
cafcba51f636: Verifying Checksum
cafcba51f636: Download complete
cafcba51f636: Pull complete
Digest: sha256:c452798b56e3f4649c557c3ff7273126042065e3be152689a1ffb880638e617d
Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/gcsupload:latest
{"component":"gcsupload","level":"info","msg":"Gathering artifacts from artifact directory: /data/gcs/artifacts","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/avc_denials.log in artifact directory. Uploading as artifacts/generated/avc_denials.log\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/containers.log in artifact directory. Uploading as artifacts/generated/containers.log\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/dmesg.log in artifact directory. Uploading as artifacts/generated/dmesg.log\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.config in artifact directory. Uploading as artifacts/generated/docker.config\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.info in artifact directory. Uploading as artifacts/generated/docker.info\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/filesystem.info in artifact directory. Uploading as artifacts/generated/filesystem.info\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/installed_packages.log in artifact directory. Uploading as artifacts/generated/installed_packages.log\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/master-metrics.log in artifact directory. Uploading as artifacts/generated/master-metrics.log\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/node-metrics.log in artifact directory. Uploading as artifacts/generated/node-metrics.log\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/pid1.journal in artifact directory. Uploading as artifacts/generated/pid1.journal\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/dnsmasq.service in artifact directory. Uploading as artifacts/journals/dnsmasq.service\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/docker.service in artifact directory. Uploading as artifacts/journals/docker.service\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/systemd-journald.service in artifact directory. Uploading as artifacts/journals/systemd-journald.service\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/shell/logs/3c90aaea435f9c0fa32804533bdaa35e42350b4cc7c2b60d095ed2cd9517d62a.json in artifact directory. Uploading as artifacts/scripts/shell/logs/3c90aaea435f9c0fa32804533bdaa35e42350b4cc7c2b60d095ed2cd9517d62a.json\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/shell/logs/scripts.log in artifact directory. Uploading as artifacts/scripts/shell/logs/scripts.log\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/artifacts/gotest_report_niAgv in artifact directory. Uploading as artifacts/scripts/test-integration/artifacts/gotest_report_niAgv\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/artifacts/gotest_report_niAgv.xml in artifact directory. Uploading as artifacts/scripts/test-integration/artifacts/gotest_report_niAgv.xml\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/logs/raw_test_output.log in artifact directory. Uploading as artifacts/scripts/test-integration/logs/raw_test_output.log\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/logs/scripts.log in artifact directory. Uploading as artifacts/scripts/test-integration/logs/scripts.log\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/logs/test-go-err.log in artifact directory. Uploading as artifacts/scripts/test-integration/logs/test-go-err.log\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-tools/logs/raw_test_output.log in artifact directory. Uploading as artifacts/scripts/test-tools/logs/raw_test_output.log\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-tools/logs/scripts.log in artifact directory. Uploading as artifacts/scripts/test-tools/logs/scripts.log\n","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-integration/artifacts/gotest_report_niAgv.xml","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-integration/logs/scripts.log","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/avc_denials.log","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/dmesg.log","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/journals/docker.service","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration/latest-build.txt","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/docker.config","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/containers.log","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/installed_packages.log","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-integration/logs/test-go-err.log","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/docker.info","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/journals/systemd-journald.service","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-integration/artifacts/gotest_report_niAgv","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration/18027.txt","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/filesystem.info","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/shell/logs/3c90aaea435f9c0fa32804533bdaa35e42350b4cc7c2b60d095ed2cd9517d62a.json","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/build-log.txt","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/latest-build.txt","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/journals/dnsmasq.service","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/shell/logs/scripts.log","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-integration/logs/raw_test_output.log","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-tools/logs/scripts.log","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/finished.json","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/master-metrics.log","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/node-metrics.log","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/pid1.journal","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-tools/logs/raw_test_output.log","level":"info","msg":"Queued for upload","time":"2018-05-22T14:57:55Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-tools/logs/scripts.log","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/finished.json","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/docker.info","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/filesystem.info","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/docker.config","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/installed_packages.log","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/journals/systemd-journald.service","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration/18027.txt","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/avc_denials.log","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration/latest-build.txt","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/node-metrics.log","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/master-metrics.log","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/journals/docker.service","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/journals/dnsmasq.service","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/latest-build.txt","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-integration/logs/scripts.log","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/dmesg.log","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/pid1.journal","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-tools/logs/raw_test_output.log","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/shell/logs/3c90aaea435f9c0fa32804533bdaa35e42350b4cc7c2b60d095ed2cd9517d62a.json","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/generated/containers.log","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/shell/logs/scripts.log","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-integration/logs/test-go-err.log","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:56Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-integration/logs/raw_test_output.log","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:57Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-integration/artifacts/gotest_report_niAgv.xml","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:57Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/build-log.txt","level":"info","msg":"Finished upload","time":"2018-05-22T14:57:57Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18027/artifacts/scripts/test-integration/artifacts/gotest_report_niAgv","level":"info","msg":"Finished upload","time":"2018-05-22T14:58:00Z"}
{"component":"gcsupload","level":"info","msg":"Finished upload to GCS","time":"2018-05-22T14:58:00Z"}
+ exit 0
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: PUSH THE ARTIFACTS AND METADATA [00h 00m 09s] ##########
[workspace@5] $ /bin/bash /tmp/jenkins3196492222163848806.sh
########## STARTING STAGE: DEPROVISION CLOUD RESOURCES ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config
+ oct deprovision

PLAYBOOK: main.yml *************************************************************
4 plays in /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml

PLAY [ensure we have the parameters necessary to deprovision virtual hosts] ****

TASK [ensure all required variables are set] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:9
skipping: [localhost] => (item=origin_ci_inventory_dir)  => {
    "changed": false, 
    "generated_timestamp": "2018-05-22 10:58:01.441692", 
    "item": "origin_ci_inventory_dir", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_region)  => {
    "changed": false, 
    "generated_timestamp": "2018-05-22 10:58:01.445129", 
    "item": "origin_ci_aws_region", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

PLAY [deprovision virtual hosts in EC2] ****************************************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

TASK [deprovision a virtual EC2 host] ******************************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:28
included: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml for localhost

TASK [update the SSH configuration to remove AWS EC2 specifics] ****************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:2
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-05-22 10:58:02.224694", 
    "msg": ""
}

TASK [rename EC2 instance for termination reaper] ******************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:8
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-05-22 10:58:02.794611", 
    "msg": "Tags {'Name': 'oct-terminate'} created for resource i-08bd240c0a8a7fd37."
}

TASK [tear down the EC2 instance] **********************************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:15
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-05-22 10:58:03.552182", 
    "instance_ids": [
        "i-08bd240c0a8a7fd37"
    ], 
    "instances": [
        {
            "ami_launch_index": "0", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-0222e2a4a338689a6"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-0639b629fb049edbe"
                }
            }, 
            "dns_name": "ec2-54-172-21-143.compute-1.amazonaws.com", 
            "ebs_optimized": false, 
            "groups": {
                "sg-7e73221a": "default"
            }, 
            "hypervisor": "xen", 
            "id": "i-08bd240c0a8a7fd37", 
            "image_id": "ami-0f07d2d9a03af96ec", 
            "instance_type": "m4.xlarge", 
            "kernel": null, 
            "key_name": "libra", 
            "launch_time": "2018-05-22T14:10:21.000Z", 
            "placement": "us-east-1d", 
            "private_dns_name": "ip-172-18-15-48.ec2.internal", 
            "private_ip": "172.18.15.48", 
            "public_dns_name": "ec2-54-172-21-143.compute-1.amazonaws.com", 
            "public_ip": "54.172.21.143", 
            "ramdisk": null, 
            "region": "us-east-1", 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "running", 
            "state_code": 16, 
            "tags": {
                "Name": "oct-terminate", 
                "openshift_etcd": "", 
                "openshift_master": "", 
                "openshift_node": ""
            }, 
            "tenancy": "default", 
            "virtualization_type": "hvm"
        }
    ], 
    "tagged_instances": []
}

TASK [remove the serialized host variables] ************************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:22
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-05-22 10:58:03.799355", 
    "path": "/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory/host_vars/172.18.15.48.yml", 
    "state": "absent"
}

PLAY [deprovision virtual hosts locally manged by Vagrant] *********************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

PLAY [clean up local configuration for deprovisioned instances] ****************

TASK [remove inventory configuration directory] ********************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:61
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-05-22 10:58:04.246097", 
    "path": "/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@5/.config/origin-ci-tool/inventory", 
    "state": "absent"
}

PLAY RECAP *********************************************************************
localhost                  : ok=8    changed=4    unreachable=0    failed=0   

+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPROVISION CLOUD RESOURCES [00h 00m 04s] ##########
Archiving artifacts
[BFA] Scanning build for known causes...
[BFA] Found failure cause(s):
[BFA] Unit or Integration Test Failed
[BFA] Job Stage Failed
[BFA] Done. 23s
[WS-CLEANUP] Deleting project workspace...[WS-CLEANUP] done
Finished: FAILURE