FailedConsole Output

Skipping 22,201 KB.. Full Log
    		W0518 11:04:14.277314     586 factory.go:1304] Request for pod integration/test5q8s4 already in flight, abandoning
    		INFO: 2018/05/18 11:04:14 balancerWrapper: handle subconn state change: 0xc42303c650, READY
    		INFO: 2018/05/18 11:04:14 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc42e0ec0c0
    		
    --- PASS: TestIntegration/TestApiGroupPreferredVersions (31.75s)
    	runner_test.go:187: 
    			master_routes_test.go:382: Looking for build api group in server group discovery
    		
    		=== OUTPUT
    		security.openshift.io/v1?timeout=32s: (227.751µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:36112]
    		I0518 11:04:14.861937     605 wrap.go:42] GET /apis/template.openshift.io/v1?timeout=32s: (217.246µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:36112]
    		I0518 11:04:14.862561     605 wrap.go:42] GET /apis/user.openshift.io/v1?timeout=32s: (246.267µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:36112]
    		I0518 11:04:14.863089     605 controller_manager.go:201] Started "openshift.io/horizontalpodautoscaling"
    		I0518 11:04:14.863111     605 controller_manager.go:191] Starting "openshift.io/default-rolebindings"
    		I0518 11:04:14.863250     605 horizontal.go:128] Starting HPA controller
    		I0518 11:04:14.863264     605 controller_utils.go:1019] Waiting for caches to sync for HPA controller
    		I0518 11:04:14.864702     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/default-rolebindings-controller: (1.34943ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.866456     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra: (1.262237ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.868858     605 wrap.go:42] POST /api/v1/namespaces/openshift-infra/serviceaccounts: (1.951009ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.879154     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/default-rolebindings-controller: (10.096493ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:35530]
    		I0518 11:04:14.880660     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (11.435066ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.883061     605 get.go:238] Starting watch for /api/v1/namespaces/openshift-infra/secrets, rv=361 labels= fields=type=kubernetes.io/service-account-token timeout=1h14m42.028870614s
    		I0518 11:04:14.888944     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.241274ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.894468     605 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (3.034073ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:35530]
    		I0518 11:04:14.905390     605 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/default-rolebindings-controller: (10.257218ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:35530]
    		I0518 11:04:14.911974     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (10.469484ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.912222     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/default-rolebindings-controller-token-lj88r: (11.910419ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.913662     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/default-rolebindings-controller: (1.094006ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.913998     605 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (18.882345ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.914270     605 client_builder.go:233] Verified credential for default-rolebindings-controller/openshift-infra
    		I0518 11:04:14.914427     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&resourceVersion=361&watch=true: (31.792563ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.915272     605 controller_manager.go:201] Started "openshift.io/default-rolebindings"
    		I0518 11:04:14.915301     605 controller_manager.go:191] Starting "openshift.io/image-trigger"
    		I0518 11:04:14.915455     605 defaultrolebindings.go:150] Starting DefaultRoleBindingController
    		I0518 11:04:14.915469     605 controller_utils.go:1019] Waiting for caches to sync for DefaultRoleBindingController controller
    		I0518 11:04:14.916719     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/image-trigger-controller: (1.169103ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.918265     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra: (1.122196ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.920653     605 wrap.go:42] POST /api/v1/namespaces/openshift-infra/serviceaccounts: (1.847084ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.923258     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/image-trigger-controller: (2.284789ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:35530]
    		I0518 11:04:14.924076     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (2.990868ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.925457     605 get.go:238] Starting watch for /api/v1/namespaces/openshift-infra/secrets, rv=364 labels= fields=type=kubernetes.io/service-account-token timeout=1h56m27.992461193s
    		I0518 11:04:14.932330     605 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (2.731802ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:35530]
    		I0518 11:04:14.958515     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/image-trigger-controller-token-gz7cm: (24.729534ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.961534     605 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/image-trigger-controller: (28.330102ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:35530]
    		I0518 11:04:14.963639     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/image-trigger-controller: (3.853271ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.964011     605 wrap.go:42] GET /api/v1/namespaces/kube-system/configmaps/kube-scheduler: (6.795228ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:35530]
    		I0518 11:04:14.964187     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (7.574643ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.964704     605 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (32.043583ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.965909     605 client_builder.go:233] Verified credential for image-trigger-controller/openshift-infra
    		I0518 11:04:14.966205     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&resourceVersion=364&watch=true: (41.178586ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.968189     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/image-trigger-controller: (1.930853ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.971154     605 wrap.go:42] PUT /api/v1/namespaces/kube-system/configmaps/kube-scheduler: (5.938548ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:35530]
    		I0518 11:04:14.971393     605 leaderelection.go:199] successfully renewed lease kube-system/kube-scheduler
    		I0518 11:04:14.973043     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (2.639366ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.974587     605 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (588.591µs) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.974830     605 client_builder.go:233] Verified credential for image-trigger-controller/openshift-infra
    		I0518 11:04:14.976109     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/image-trigger-controller: (1.076676ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.979096     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (2.533779ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.980727     605 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (675.836µs) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.981052     605 client_builder.go:233] Verified credential for image-trigger-controller/openshift-infra
    		I0518 11:04:14.981879     605 controller_manager.go:201] Started "openshift.io/image-trigger"
    		I0518 11:04:14.981895     605 controller_manager.go:191] Starting "openshift.io/ingress-ip"
    		I0518 11:04:14.982030     605 image_trigger_controller.go:215] Starting trigger controller
    		I0518 11:04:14.983178     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/service-ingress-ip-controller: (1.040662ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.985889     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra: (2.31684ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.988461     605 wrap.go:42] POST /api/v1/namespaces/openshift-infra/serviceaccounts: (2.043258ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.994838     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (5.099347ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:14.995091     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/service-ingress-ip-controller: (5.849078ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:35530]
    		I0518 11:04:15.005892     605 get.go:238] Starting watch for /api/v1/namespaces/openshift-infra/secrets, rv=368 labels= fields=type=kubernetes.io/service-account-token timeout=1h35m39.369327062s
    		I0518 11:04:15.008779     605 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/admin?timeout=1m0s: (5.684769ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36138]
    		I0518 11:04:15.016853     605 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (7.108517ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:35530]
    		I0518 11:04:15.030883     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/service-ingress-ip-controller-token-6fk45: (13.606963ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.032015     605 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/edit?timeout=1m0s: (21.127543ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36138]
    		I0518 11:04:15.032440     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (20.740774ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.032614     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (14.175941ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.032779     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (14.652819ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.032941     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (16.198257ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.033083     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (19.071273ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.033221     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (19.644746ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.033381     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (20.155161ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.033526     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (20.438198ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.033679     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (19.984498ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.033822     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (21.057477ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.033957     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (21.675465ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.034098     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (22.333871ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.047923     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/service-ingress-ip-controller: (11.637577ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.048157     605 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/service-ingress-ip-controller: (30.729327ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:35530]
    		I0518 11:04:15.049012     605 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/view?timeout=1m0s: (13.215461ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36138]
    		I0518 11:04:15.049142     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.512796ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.049291     605 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.245578ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.049560     605 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (33.901017ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.050018     605 client_builder.go:233] Verified credential for service-ingress-ip-controller/openshift-infra
    		I0518 11:04:15.050489     605 controller_manager.go:201] Started "openshift.io/ingress-ip"
    		I0518 11:04:15.050508     605 controller_manager.go:204] Started Origin Controllers
    		I0518 11:04:15.050969     605 resource_quota_monitor.go:275] QuotaMonitor started 36 new monitors, 36 currently running
    		I0518 11:04:15.051452     605 reflector.go:202] Starting reflector *v1beta1.ReplicaSet (10m0s) from github.com/openshift/origin/vendor/k8s.io/client-go/informers/factory.go:87
    		I0518 11:04:15.051469     605 reflector.go:240] Listing and watching *v1beta1.ReplicaSet from github.com/openshift/origin/vendor/k8s.io/client-go/informers/factory.go:87
    		I0518 11:04:15.051532     605 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&resourceVersion=368&watch=true: (46.100445ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.051952     605 reflector.go:202] Starting reflector *template.Template (10m0s) from github.com/openshift/origin/pkg/template/generated/informers/internalversion/factory.go:58
    		I0518 11:04:15.051971     605 reflector.go:240] Listing and watching *template.Template from github.com/openshift/origin/pkg/template/generated/informers/internalversion/factory.go:58
    		I0518 11:04:15.052409     605 wrap.go:42] GET /api?timeout=1m0s: (589.998µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36138]
    		I0518 11:04:15.055702     605 wrap.go:42] GET /apis/template.openshift.io/v1/templates?limit=500&resourceVersion=0: (2.521162ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.055789     605 wrap.go:42] GET /apis?timeout=1m0s: (819.786µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36138]
    		I0518 11:04:15.056738     605 get.go:238] Starting watch for /apis/template.openshift.io/v1/templates, rv=370 labels= fields= timeout=5m37s
    		INFO: 2018/05/18 11:04:15 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc43b464660
    		INFO: 2018/05/18 11:04:15 dialing to target with scheme: ""
    		INFO: 2018/05/18 11:04:15 could not get resolver for scheme: ""
    		INFO: 2018/05/18 11:04:15 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/18 11:04:15 balancerWrapper: got update addr from Notify: [{127.0.0.1:25274 <nil>}]
    		INFO: 2018/05/18 11:04:15 ccBalancerWrapper: new subconn: [{127.0.0.1:25274 0  <nil>}]
    		INFO: 2018/05/18 11:04:15 balancerWrapper: handle subconn state change: 0xc4305d11e0, CONNECTING
    		INFO: 2018/05/18 11:04:15 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc43b464660
    		I0518 11:04:15.063624     605 wrap.go:42] GET /apis/extensions/v1beta1/replicasets?limit=500&resourceVersion=0: (10.973234ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:35530]
    		I0518 11:04:15.064462     605 get.go:238] Starting watch for /apis/extensions/v1beta1/replicasets, rv=370 labels= fields= timeout=9m56s
    		INFO: 2018/05/18 11:04:15 balancerWrapper: handle subconn state change: 0xc4305d11e0, READY
    		INFO: 2018/05/18 11:04:15 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc43b464660
    		
    --- PASS: TestIntegration/TestApiGroups (34.91s)
    	runner_test.go:187: 
    			master_routes_test.go:430: Looking for build api group in server group discovery
    			master_routes_test.go:445: Looking for builds resource in resource discovery
    			master_routes_test.go:471: Creating test namespace "testapigroup664145806"
    			master_routes_test.go:478: GETting builds
    			master_routes_test.go:489: Creating a Build
    			master_routes_test.go:496: GETting builds again
    		I0518 11:04:18.221518     599 wrap.go:42] GET /apis/oauth.openshift.io/v1?timeout=32s: (287.858µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		
    		=== OUTPUT
    		/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44922]
    		I0518 11:04:18.039993     599 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/namespace-security-allocation-controller: (58.438264ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.040572     599 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/view?timeout=1m0s: (57.25275ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44922]
    		I0518 11:04:18.041092     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (48.466565ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.041339     599 wrap.go:42] PUT /api/v1/namespaces/kube-system/configmaps/kube-scheduler: (61.675869ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:44674]
    		I0518 11:04:18.041557     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (40.112397ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.041721     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (40.739996ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.041877     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (41.344761ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.042014     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (41.9607ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.042145     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (42.547125ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.042274     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (43.17554ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.042425     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (40.529251ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.042564     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (43.954269ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.042700     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (49.5794ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.042835     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (48.275816ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.042974     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (47.96828ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.043104     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (49.481392ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.043234     599 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (49.156027ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.045039     599 leaderelection.go:199] successfully renewed lease kube-system/kube-scheduler
    		I0518 11:04:18.047581     599 wrap.go:42] GET /api?timeout=1m0s: (601.382µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44922]
    		I0518 11:04:18.048028     599 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (2.398429ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.048864     599 wrap.go:42] GET /apis?timeout=1m0s: (816.65µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44922]
    		I0518 11:04:18.050339     599 wrap.go:42] GET /apis/build.openshift.io/v1?timeout=1m0s: (584.735µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44922]
    		I0518 11:04:18.066396     599 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (649.398µs) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.066655     599 client_builder.go:233] Verified credential for namespace-security-allocation-controller/openshift-infra
    		I0518 11:04:18.066876     599 controller_manager.go:201] Started "openshift.io/namespace-security-allocation"
    		I0518 11:04:18.066891     599 controller_manager.go:191] Starting "openshift.io/horizontalpodautoscaling"
    		I0518 11:04:18.067017     599 controller_utils.go:1019] Waiting for caches to sync for namespace-security-allocation-controller controller
    		I0518 11:04:18.082590     599 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/horizontal-pod-autoscaler: (14.099447ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.083114     599 wrap.go:42] POST /api/v1/namespaces?timeout=1m0s: (15.974944ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44922]
    		I0518 11:04:18.085235     599 wrap.go:42] GET /apis/build.openshift.io/v1: (437.485µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44922]
    		I0518 11:04:18.088428     599 wrap.go:42] GET /api/v1/namespaces/openshift-infra: (4.217372ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.100222     599 wrap.go:42] POST /api/v1/namespaces/openshift-infra/serviceaccounts: (10.252878ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.112272     599 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/horizontal-pod-autoscaler: (10.415627ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44674]
    		I0518 11:04:18.112364     599 wrap.go:42] PUT /api/v1/namespaces/testapigroup664145806/finalize: (22.596519ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.114002     599 wrap.go:42] POST /apis/authorization.k8s.io/v1/subjectaccessreviews: (843.342µs) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.120753     599 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (19.520752ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.121756     599 controller.go:537] quota admission added evaluator for: {build.openshift.io builds}
    		I0518 11:04:18.122090     599 controller.go:537] quota admission added evaluator for: {build.openshift.io builds}
    		I0518 11:04:18.122844     599 get.go:238] Starting watch for /api/v1/namespaces/openshift-infra/secrets, rv=336 labels= fields=type=kubernetes.io/service-account-token timeout=1h11m29.13495193s
    		I0518 11:04:18.124287     599 wrap.go:42] GET /api/v1/namespaces/testapigroup664145806/resourcequotas: (1.275812ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.125337     599 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (6.485718ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44674]
    		I0518 11:04:18.155517     599 wrap.go:42] POST /apis/build.openshift.io/v1/namespaces/testapigroup664145806/builds?timeout=1m0s: (66.838964ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44922]
    		I0518 11:04:18.187774     599 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/horizontal-pod-autoscaler: (61.255869ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:44674]
    		I0518 11:04:18.188031     599 wrap.go:42] GET /apis/build.openshift.io/v1/namespaces/testapigroup664145806/builds/foo: (31.14365ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44922]
    		I0518 11:04:18.188555     599 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/horizontal-pod-autoscaler-token-9gfkv: (61.070649ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.190548     599 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/horizontal-pod-autoscaler: (1.47805ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.190948     599 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (64.920972ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.191166     599 client_builder.go:233] Verified credential for horizontal-pod-autoscaler/openshift-infra
    		I0518 11:04:18.191499     599 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&resourceVersion=336&watch=true: (69.088061ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44674]
    		I0518 11:04:18.192077     599 wrap.go:42] GET /api?timeout=32s: (257.385µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.195024     599 wrap.go:42] GET /apis?timeout=32s: (304.261µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.195885     599 wrap.go:42] GET /api/v1?timeout=32s: (417.924µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.196570     599 wrap.go:42] GET /apis/apiregistration.k8s.io/v1?timeout=32s: (229.409µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.197167     599 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1?timeout=32s: (196.707µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.197885     599 wrap.go:42] GET /apis/extensions/v1beta1?timeout=32s: (333.793µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.198549     599 wrap.go:42] GET /apis/apps/v1?timeout=32s: (255.454µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.199222     599 wrap.go:42] GET /apis/apps/v1beta2?timeout=32s: (235.656µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.199857     599 wrap.go:42] GET /apis/apps/v1beta1?timeout=32s: (211.451µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.200517     599 wrap.go:42] GET /apis/events.k8s.io/v1beta1?timeout=32s: (245.896µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.201128     599 wrap.go:42] GET /apis/authentication.k8s.io/v1?timeout=32s: (204.308µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.201762     599 wrap.go:42] GET /apis/authentication.k8s.io/v1beta1?timeout=32s: (208.151µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.202386     599 wrap.go:42] GET /apis/authorization.k8s.io/v1?timeout=32s: (221.275µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.203006     599 wrap.go:42] GET /apis/authorization.k8s.io/v1beta1?timeout=32s: (204.38µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.203455     599 wrap.go:42] DELETE /api/v1/namespaces/testapigroup664145806?timeout=1m0s: (14.230331ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44922]
    		I0518 11:04:18.203669     599 wrap.go:42] GET /apis/autoscaling/v1?timeout=32s: (209.433µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		INFO: 2018/05/18 11:04:18 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc433c04660
    		INFO: 2018/05/18 11:04:18 dialing to target with scheme: ""
    		INFO: 2018/05/18 11:04:18 could not get resolver for scheme: ""
    		INFO: 2018/05/18 11:04:18 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/18 11:04:18 balancerWrapper: got update addr from Notify: [{127.0.0.1:25019 <nil>}]
    		INFO: 2018/05/18 11:04:18 ccBalancerWrapper: new subconn: [{127.0.0.1:25019 0  <nil>}]
    		INFO: 2018/05/18 11:04:18 balancerWrapper: handle subconn state change: 0xc427a66b40, CONNECTING
    		INFO: 2018/05/18 11:04:18 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc433c04660
    		I0518 11:04:18.205809     599 wrap.go:42] GET /apis/autoscaling/v2beta1?timeout=32s: (246.117µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.206491     599 wrap.go:42] GET /apis/batch/v1?timeout=32s: (289.104µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.207166     599 wrap.go:42] GET /apis/batch/v1beta1?timeout=32s: (216.156µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.207764     599 wrap.go:42] GET /apis/batch/v2alpha1?timeout=32s: (199.341µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.208379     599 wrap.go:42] GET /apis/certificates.k8s.io/v1beta1?timeout=32s: (238.682µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.209011     599 wrap.go:42] GET /apis/networking.k8s.io/v1?timeout=32s: (206.325µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.209685     599 wrap.go:42] GET /apis/policy/v1beta1?timeout=32s: (267.786µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.213434     599 wrap.go:42] GET /apis/authorization.openshift.io/v1?timeout=32s: (347.735µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.214143     599 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1?timeout=32s: (247.748µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.214845     599 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1beta1?timeout=32s: (238.625µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.215503     599 wrap.go:42] GET /apis/storage.k8s.io/v1?timeout=32s: (256.894µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.216156     599 wrap.go:42] GET /apis/storage.k8s.io/v1beta1?timeout=32s: (224.326µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.216928     599 wrap.go:42] GET /apis/admissionregistration.k8s.io/v1beta1?timeout=32s: (227.722µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.217637     599 wrap.go:42] GET /apis/apiextensions.k8s.io/v1beta1?timeout=32s: (326.804µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.218380     599 wrap.go:42] GET /apis/apps.openshift.io/v1?timeout=32s: (301.711µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.219087     599 wrap.go:42] GET /apis/build.openshift.io/v1?timeout=32s: (267.843µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.219799     599 wrap.go:42] GET /apis/image.openshift.io/v1?timeout=32s: (287.157µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		I0518 11:04:18.220652     599 wrap.go:42] GET /apis/network.openshift.io/v1?timeout=32s: (407.464µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:horizontal-pod-autoscaler] 127.0.0.1:44912]
    		INFO: 2018/05/18 11:04:18 balancerWrapper: handle subconn state change: 0xc427a66b40, READY
    		INFO: 2018/05/18 11:04:18 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc433c04660
    		
    --- PASS: TestIntegration/TestAllowedSCCViaRBAC (41.78s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		92] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.713555    1243 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0518 11:04:38.713565    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.713582    1243 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0518 11:04:38.713596    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.713603    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.713619    1243 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0518 11:04:38.713625    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.713638    1243 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0518 11:04:38.713657    1243 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0518 11:04:38.713686    1243 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0518 11:04:38.713700    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.713705    1243 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0518 11:04:38.713725    1243 admission.go:217] validating pod test3 (generate: ) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0518 11:04:38.713907    1243 admission.go:170] pod test3 (generate: ) validated against provider hostaccess
    		I0518 11:04:38.715579    1243 wrap.go:42] POST /api/v1/namespaces/project1/pods?timeout=1m0s: (5.32325ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36658]
    		I0518 11:04:38.716004    1243 factory.go:1147] About to try and schedule pod test3
    		I0518 11:04:38.716025    1243 scheduler.go:439] Attempting to schedule pod: project1/test3
    		I0518 11:04:38.716045    1243 disruption.go:328] addPod called on pod "test3"
    		I0518 11:04:38.716065    1243 disruption.go:403] No PodDisruptionBudgets found for pod test3, PodDisruptionBudget controller will avoid syncing.
    		I0518 11:04:38.716074    1243 disruption.go:331] No matching pdb for pod "test3"
    		I0518 11:04:38.716048    1243 scheduler.go:191] Failed to schedule pod: project1/test3
    		I0518 11:04:38.716805    1243 factory.go:1262] Unable to schedule project1 test3: no nodes are registered to the cluster; waiting
    		I0518 11:04:38.716868    1243 factory.go:1375] Updating pod condition for project1/test3 to (PodScheduled==False)
    		I0518 11:04:38.716211    1243 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"project1", Name:"test3"}
    		I0518 11:04:38.720244    1243 backoff_utils.go:79] Backing off 1s
    		I0518 11:04:38.722760    1243 pvc_protection_controller.go:276] Got event on pod project1/test3
    		I0518 11:04:38.726294    1243 admission.go:97] getting security context constraints for pod test4 (generate: ) in namespace project2 with user info &{user1 3d60a381-5a8b-11e8-a87c-0242ac110002 [system:authenticated:oauth system:authenticated] map[scopes.authorization.openshift.io:[]]}
    		I0518 11:04:38.730555    1243 admission.go:108] getting security context constraints for pod test4 (generate: ) with service account info &{system:serviceaccount:project2:default  [system:serviceaccounts system:serviceaccounts:project2] map[]}
    		I0518 11:04:38.738925    1243 wrap.go:42] POST /api/v1/namespaces/project1/events: (16.443988ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:36104]
    		I0518 11:04:38.742543    1243 wrap.go:42] PUT /api/v1/namespaces/project1/pods/test3/status: (21.929572ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:36104]
    		I0518 11:04:38.742788    1243 wrap.go:42] GET /api/v1/namespaces/project2: (7.621149ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36104]
    		I0518 11:04:38.743715    1243 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I0518 11:04:38.743736    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0518 11:04:38.743750    1243 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0518 11:04:38.743790    1243 admission.go:217] validating pod test4 (generate: ) against providers restricted
    		I0518 11:04:38.744059    1243 admission.go:179] unable to validate pod test4 (generate: ) against any security context constraint: [provider restricted: .spec.securityContext.hostPID: Invalid value: true: Host PID is not allowed to be used spec.containers[0].securityContext.hostPID: Invalid value: true: Host PID is not allowed to be used]
    		I0518 11:04:38.744494    1243 wrap.go:42] POST /api/v1/namespaces/project2/pods?timeout=1m0s: (21.435356ms) 403 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36658]
    		I0518 11:04:38.746041    1243 disruption.go:340] updatePod called on pod "test3"
    		I0518 11:04:38.746071    1243 disruption.go:403] No PodDisruptionBudgets found for pod test3, PodDisruptionBudget controller will avoid syncing.
    		I0518 11:04:38.746083    1243 disruption.go:343] No matching pdb for pod "test3"
    		I0518 11:04:38.746268    1243 pvc_protection_controller.go:276] Got event on pod project1/test3
    		I0518 11:04:38.748868    1243 admission.go:97] getting security context constraints for pod test5 (generate: ) in namespace project1 with user info &{user2 3d7027c6-5a8b-11e8-a87c-0242ac110002 [system:authenticated:oauth system:authenticated] map[scopes.authorization.openshift.io:[]]}
    		I0518 11:04:38.748905    1243 factory.go:1147] About to try and schedule pod test3
    		I0518 11:04:38.748927    1243 scheduler.go:439] Attempting to schedule pod: project1/test3
    		I0518 11:04:38.748954    1243 scheduler.go:191] Failed to schedule pod: project1/test3
    		I0518 11:04:38.749191    1243 factory.go:1262] Unable to schedule project1 test3: no nodes are registered to the cluster; waiting
    		I0518 11:04:38.749248    1243 factory.go:1375] Updating pod condition for project1/test3 to (PodScheduled==False)
    		W0518 11:04:38.750621    1243 factory.go:1304] Request for pod project1/test3 already in flight, abandoning
    		I0518 11:04:38.750913    1243 admission.go:108] getting security context constraints for pod test5 (generate: ) with service account info &{system:serviceaccount:project1:default  [system:serviceaccounts system:serviceaccounts:project1] map[]}
    		I0518 11:04:38.755804    1243 wrap.go:42] GET /api/v1/namespaces/project1: (1.932059ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36104]
    		I0518 11:04:38.756207    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.756253    1243 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0518 11:04:38.756269    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.756281    1243 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0518 11:04:38.756314    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.756335    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.756390    1243 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0518 11:04:38.756406    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.756418    1243 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0518 11:04:38.756439    1243 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0518 11:04:38.756489    1243 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0518 11:04:38.756505    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.756516    1243 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0518 11:04:38.756553    1243 admission.go:217] validating pod test5 (generate: ) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0518 11:04:38.757276    1243 admission.go:170] pod test5 (generate: ) validated against provider hostaccess
    		I0518 11:04:38.761791    1243 wrap.go:42] POST /api/v1/namespaces/project1/pods?timeout=1m0s: (15.149854ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36658]
    		I0518 11:04:38.762237    1243 disruption.go:328] addPod called on pod "test5"
    		I0518 11:04:38.762261    1243 disruption.go:403] No PodDisruptionBudgets found for pod test5, PodDisruptionBudget controller will avoid syncing.
    		I0518 11:04:38.762273    1243 disruption.go:331] No matching pdb for pod "test5"
    		I0518 11:04:38.762948    1243 pvc_protection_controller.go:276] Got event on pod project1/test5
    		I0518 11:04:38.763053    1243 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"project1", Name:"test5"}
    		I0518 11:04:38.763538    1243 factory.go:1147] About to try and schedule pod test5
    		I0518 11:04:38.763553    1243 scheduler.go:439] Attempting to schedule pod: project1/test5
    		I0518 11:04:38.763573    1243 scheduler.go:191] Failed to schedule pod: project1/test5
    		I0518 11:04:38.763603    1243 factory.go:1262] Unable to schedule project1 test5: no nodes are registered to the cluster; waiting
    		I0518 11:04:38.763634    1243 factory.go:1375] Updating pod condition for project1/test5 to (PodScheduled==False)
    		I0518 11:04:38.763944    1243 backoff_utils.go:79] Backing off 1s
    		I0518 11:04:38.764840    1243 admission.go:97] getting security context constraints for pod test6 (generate: ) in namespace project2 with user info &{user2 3d7027c6-5a8b-11e8-a87c-0242ac110002 [system:authenticated:oauth system:authenticated] map[scopes.authorization.openshift.io:[]]}
    		I0518 11:04:38.765611    1243 admission.go:108] getting security context constraints for pod test6 (generate: ) with service account info &{system:serviceaccount:project2:default  [system:serviceaccounts system:serviceaccounts:project2] map[]}
    		I0518 11:04:38.771448    1243 wrap.go:42] GET /api/v1/namespaces/project2: (3.804223ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36104]
    		I0518 11:04:38.771672    1243 wrap.go:42] PATCH /api/v1/namespaces/project1/events/test3.152fb8561e4b8de7: (17.063665ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:36104]
    		I0518 11:04:38.772093    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0518 11:04:38.772121    1243 wrap.go:42] PUT /api/v1/namespaces/project1/pods/test5/status: (7.935012ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:36104]
    		I0518 11:04:38.772139    1243 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I0518 11:04:38.772153    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0518 11:04:38.772164    1243 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0518 11:04:38.772193    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0518 11:04:38.772211    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0518 11:04:38.772235    1243 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I0518 11:04:38.772247    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0518 11:04:38.772257    1243 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0518 11:04:38.772276    1243 matcher.go:342] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0518 11:04:38.772309    1243 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I0518 11:04:38.772320    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0518 11:04:38.772330    1243 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0518 11:04:38.772390    1243 admission.go:217] validating pod test6 (generate: ) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0518 11:04:38.772890    1243 admission.go:170] pod test6 (generate: ) validated against provider hostaccess
    		I0518 11:04:38.773741    1243 disruption.go:340] updatePod called on pod "test5"
    		I0518 11:04:38.773765    1243 disruption.go:403] No PodDisruptionBudgets found for pod test5, PodDisruptionBudget controller will avoid syncing.
    		I0518 11:04:38.773776    1243 disruption.go:343] No matching pdb for pod "test5"
    		I0518 11:04:38.773973    1243 pvc_protection_controller.go:276] Got event on pod project1/test5
    		I0518 11:04:38.775694    1243 wrap.go:42] POST /api/v1/namespaces/project2/pods?timeout=1m0s: (11.794198ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36658]
    		I0518 11:04:38.777087    1243 disruption.go:328] addPod called on pod "test6"
    		I0518 11:04:38.777111    1243 disruption.go:403] No PodDisruptionBudgets found for pod test6, PodDisruptionBudget controller will avoid syncing.
    		I0518 11:04:38.777123    1243 disruption.go:331] No matching pdb for pod "test6"
    		I0518 11:04:38.777160    1243 pvc_protection_controller.go:276] Got event on pod project2/test6
    		I0518 11:04:38.777238    1243 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"project2", Name:"test6"}
    		I0518 11:04:38.778582    1243 factory.go:1147] About to try and schedule pod test5
    		I0518 11:04:38.778602    1243 scheduler.go:439] Attempting to schedule pod: project1/test5
    		I0518 11:04:38.778628    1243 scheduler.go:191] Failed to schedule pod: project1/test5
    		I0518 11:04:38.778677    1243 factory.go:1262] Unable to schedule project1 test5: no nodes are registered to the cluster; waiting
    		I0518 11:04:38.778740    1243 factory.go:1375] Updating pod condition for project1/test5 to (PodScheduled==False)
    		W0518 11:04:38.778824    1243 factory.go:1304] Request for pod project1/test5 already in flight, abandoning
    		I0518 11:04:38.779516    1243 factory.go:1147] About to try and schedule pod test6
    		I0518 11:04:38.779537    1243 scheduler.go:439] Attempting to schedule pod: project2/test6
    		I0518 11:04:38.779562    1243 scheduler.go:191] Failed to schedule pod: project2/test6
    		I0518 11:04:38.779605    1243 factory.go:1262] Unable to schedule project2 test6: no nodes are registered to the cluster; waiting
    		I0518 11:04:38.779665    1243 factory.go:1375] Updating pod condition for project2/test6 to (PodScheduled==False)
    		I0518 11:04:38.779750    1243 backoff_utils.go:79] Backing off 1s
    		I0518 11:04:38.780721    1243 controller.go:537] quota admission added evaluator for: {security.openshift.io podsecuritypolicyselfsubjectreviews}
    		I0518 11:04:38.793579    1243 wrap.go:42] POST /api/v1/namespaces/project1/events: (20.045697ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:36104]
    		I0518 11:04:38.793642    1243 wrap.go:42] GET /api/v1/namespaces/project1: (10.85552ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36104]
    		I0518 11:04:38.794090    1243 wrap.go:42] PUT /api/v1/namespaces/project2/pods/test6/status: (13.779615ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:36104]
    		I0518 11:04:38.794269    1243 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0518 11:04:38.794853    1243 wrap.go:42] POST /apis/security.openshift.io/v1/namespaces/project1/podsecuritypolicyselfsubjectreviews?timeout=1m0s: (15.663094ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36658]
    		I0518 11:04:38.796364    1243 disruption.go:340] updatePod called on pod "test6"
    		I0518 11:04:38.796392    1243 disruption.go:403] No PodDisruptionBudgets found for pod test6, PodDisruptionBudget controller will avoid syncing.
    		I0518 11:04:38.796403    1243 disruption.go:343] No matching pdb for pod "test6"
    		I0518 11:04:38.796427    1243 pvc_protection_controller.go:276] Got event on pod project2/test6
    		I0518 11:04:38.797558    1243 factory.go:1147] About to try and schedule pod test6
    		I0518 11:04:38.797580    1243 scheduler.go:439] Attempting to schedule pod: project2/test6
    		I0518 11:04:38.797608    1243 scheduler.go:191] Failed to schedule pod: project2/test6
    		I0518 11:04:38.797658    1243 factory.go:1262] Unable to schedule project2 test6: no nodes are registered to the cluster; waiting
    		I0518 11:04:38.797734    1243 factory.go:1375] Updating pod condition for project2/test6 to (PodScheduled==False)
    		W0518 11:04:38.797845    1243 factory.go:1304] Request for pod project2/test6 already in flight, abandoning
    		I0518 11:04:38.801074    1243 wrap.go:42] GET /api/v1/namespaces/project2: (1.981481ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36104]
    		I0518 11:04:38.801411    1243 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0518 11:04:38.801827    1243 wrap.go:42] POST /apis/security.openshift.io/v1/namespaces/project2/podsecuritypolicyselfsubjectreviews?timeout=1m0s: (5.167225ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:36658]
    		I0518 11:04:38.802713    1243 wrap.go:42] PATCH /api/v1/namespaces/project1/events/test5.152fb85621153c1c: (7.391456ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:36104]
    		INFO: 2018/05/18 11:04:38 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc4403c2cc0
    		INFO: 2018/05/18 11:04:38 dialing to target with scheme: ""
    		INFO: 2018/05/18 11:04:38 could not get resolver for scheme: ""
    		INFO: 2018/05/18 11:04:38 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/18 11:04:38 balancerWrapper: got update addr from Notify: [{127.0.0.1:20895 <nil>}]
    		INFO: 2018/05/18 11:04:38 ccBalancerWrapper: new subconn: [{127.0.0.1:20895 0  <nil>}]
    		INFO: 2018/05/18 11:04:38 balancerWrapper: handle subconn state change: 0xc430a02550, CONNECTING
    		INFO: 2018/05/18 11:04:38 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc4403c2cc0
    		I0518 11:04:38.818254    1243 wrap.go:42] POST /api/v1/namespaces/project2/events: (14.263726ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:36104]
    		INFO: 2018/05/18 11:04:38 balancerWrapper: handle subconn state change: 0xc430a02550, READY
    		INFO: 2018/05/18 11:04:38 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc4403c2cc0
    		INFO: 2018/05/18 11:04:38 balancerWrapper: got update addr from Notify: [{127.0.0.1:20895 <nil>}]
    		
    --- PASS: TestIntegration/TestAggregator (31.41s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		27.0.0.1:38438]
    		I0518 11:04:46.398602    1267 wrap.go:42] PUT /api/v1/namespaces/kube-system/serviceaccounts/default: (12.019259ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:38438]
    		I0518 11:04:46.399413    1267 wrap.go:42] POST /api/v1/namespaces/openshift-node/secrets: (5.490156ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:38438]
    		I0518 11:04:46.400209    1267 shared_informer.go:123] caches populated
    		I0518 11:04:46.400230    1267 controller_utils.go:1026] Caches are synced for garbage collector controller
    		I0518 11:04:46.400239    1267 garbagecollector.go:227] synced garbage collector
    		I0518 11:04:46.400256    1267 shared_informer.go:123] caches populated
    		I0518 11:04:46.400270    1267 controller_utils.go:1026] Caches are synced for certificate controller
    		I0518 11:04:46.402899    1267 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/default: (12.968356ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:38438]
    		I0518 11:04:46.406070    1267 wrap.go:42] PUT /api/v1/namespaces/openshift/serviceaccounts/default: (9.071207ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:38438]
    		I0518 11:04:46.408636    1267 wrap.go:42] PUT /api/v1/namespaces/openshift-node/serviceaccounts/default: (8.022452ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:38438]
    		I0518 11:04:46.411169    1267 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/service-serving-cert-controller-token-ztv2h: (4.202596ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:38438]
    		I0518 11:04:46.412570    1267 request.go:485] Throttling request took 146.995238ms, request: PUT:https://127.0.0.1:25450/api/v1/namespaces/openshift-infra/serviceaccounts/serviceaccount-pull-secrets-controller
    		I0518 11:04:46.415765    1267 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/serviceaccount-pull-secrets-controller: (2.91151ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:38598]
    		I0518 11:04:46.415990    1267 create_dockercfg_secrets.go:441] Creating token secret "serviceaccount-pull-secrets-controller-token-x97lc" for service account openshift-infra/serviceaccount-pull-secrets-controller
    		I0518 11:04:46.419132    1267 wrap.go:42] POST /apis/rbac.authorization.k8s.io/v1/namespaces/openshift-node/rolebindings: (64.4361ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:default-rolebindings-controller] 127.0.0.1:38598]
    		I0518 11:04:46.419550    1267 shared_informer.go:123] caches populated
    		I0518 11:04:46.419566    1267 controller_utils.go:1026] Caches are synced for daemon sets controller
    		I0518 11:04:46.434959    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.325267ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.445175    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.363833ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.445670    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.63857ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.446350    1267 request.go:485] Throttling request took 144.307417ms, request: PUT:https://127.0.0.1:25450/api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-controller
    		I0518 11:04:46.449145    1267 wrap.go:42] PUT /apis/rbac.authorization.k8s.io/v1/clusterroles/view: (54.798423ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:kube-system:clusterrole-aggregation-controller] 127.0.0.1:38598]
    		I0518 11:04:46.450511    1267 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-controller: (3.910695ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:38598]
    		I0518 11:04:46.450810    1267 create_dockercfg_secrets.go:441] Creating token secret "template-instance-controller-token-f8gnv" for service account openshift-infra/template-instance-controller
    		I0518 11:04:46.452496    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.149529ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.453028    1267 wrap.go:42] GET /api?timeout=32s: (335.624µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.454145    1267 wrap.go:42] GET /apis?timeout=32s: (461.472µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.454371    1267 wrap.go:42] PUT /apis/rbac.authorization.k8s.io/v1/clusterroles/edit: (60.32926ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:kube-system:clusterrole-aggregation-controller] 127.0.0.1:38598]
    		I0518 11:04:46.455757    1267 wrap.go:42] GET /api/v1?timeout=32s: (574.396µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.456602    1267 wrap.go:42] GET /apis/apiregistration.k8s.io/v1?timeout=32s: (303.802µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.457371    1267 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1?timeout=32s: (283.679µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.458098    1267 wrap.go:42] GET /apis/extensions/v1beta1?timeout=32s: (317.506µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.458752    1267 wrap.go:42] GET /apis/apps/v1?timeout=32s: (269.201µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.459443    1267 wrap.go:42] GET /apis/apps/v1beta2?timeout=32s: (293.074µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.459627    1267 wrap.go:42] PUT /apis/rbac.authorization.k8s.io/v1/clusterroles/admin: (65.303947ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:kube-system:clusterrole-aggregation-controller] 127.0.0.1:38598]
    		I0518 11:04:46.460444    1267 wrap.go:42] GET /apis/apps/v1beta1?timeout=32s: (305.088µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.461121    1267 wrap.go:42] GET /apis/events.k8s.io/v1beta1?timeout=32s: (236.06µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.462424    1267 wrap.go:42] GET /apis/authentication.k8s.io/v1?timeout=32s: (248.922µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.462966    1267 wrap.go:42] GET /apis/authentication.k8s.io/v1beta1?timeout=32s: (223.94µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.463624    1267 wrap.go:42] GET /apis/authorization.k8s.io/v1?timeout=32s: (268.919µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.463848    1267 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/admin?timeout=1m0s: (1.806567ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38600]
    		I0518 11:04:46.464245    1267 wrap.go:42] GET /apis/authorization.k8s.io/v1beta1?timeout=32s: (256.912µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.464852    1267 wrap.go:42] GET /apis/autoscaling/v1?timeout=32s: (245.437µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.465408    1267 wrap.go:42] GET /apis/autoscaling/v2beta1?timeout=32s: (245.898µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.466046    1267 wrap.go:42] GET /apis/batch/v1?timeout=32s: (225.071µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.466536    1267 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/edit?timeout=1m0s: (1.626152ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38600]
    		I0518 11:04:46.466608    1267 wrap.go:42] GET /apis/batch/v1beta1?timeout=32s: (220.366µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.467205    1267 wrap.go:42] GET /apis/batch/v2alpha1?timeout=32s: (240.554µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.469132    1267 wrap.go:42] GET /apis/certificates.k8s.io/v1beta1?timeout=32s: (373.525µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.469742    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.861491ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.469841    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.579762ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.469964    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (2.212187ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.470128    1267 wrap.go:42] GET /apis/networking.k8s.io/v1?timeout=32s: (342.53µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.470982    1267 wrap.go:42] GET /apis/policy/v1beta1?timeout=32s: (428.992µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.471898    1267 wrap.go:42] GET /apis/authorization.openshift.io/v1?timeout=32s: (439.098µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.472562    1267 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1?timeout=32s: (268.256µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.472995    1267 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/view?timeout=1m0s: (2.238774ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38600]
    		I0518 11:04:46.473380    1267 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1beta1?timeout=32s: (389.901µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.474138    1267 wrap.go:42] GET /apis/storage.k8s.io/v1?timeout=32s: (340.184µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.474836    1267 wrap.go:42] GET /apis/storage.k8s.io/v1beta1?timeout=32s: (321.456µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.475713    1267 wrap.go:42] GET /apis/admissionregistration.k8s.io/v1beta1?timeout=32s: (413.386µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.476824    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.164133ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.477030    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.896681ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.477618    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.465187ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.477841    1267 wrap.go:42] GET /apis/apiextensions.k8s.io/v1beta1?timeout=32s: (419.603µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.478299    1267 request.go:485] Throttling request took 145.356096ms, request: PUT:https://127.0.0.1:25450/api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-finalizer-controller
    		I0518 11:04:46.478328    1267 wrap.go:42] GET /api/v1/namespaces/default?timeout=1m0s: (1.481283ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38600]
    		I0518 11:04:46.479620    1267 wrap.go:42] GET /apis/apps.openshift.io/v1?timeout=32s: (499.585µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.481905    1267 wrap.go:42] GET /apis/build.openshift.io/v1?timeout=32s: (431.483µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.482824    1267 wrap.go:42] GET /api/v1/namespaces/default: (2.280765ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.482938    1267 wrap.go:42] GET /apis/image.openshift.io/v1?timeout=32s: (443.409µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.483032    1267 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-finalizer-controller: (4.252045ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:38598]
    		I0518 11:04:46.483265    1267 create_dockercfg_secrets.go:441] Creating token secret "template-instance-finalizer-controller-token-qxn9w" for service account openshift-infra/template-instance-finalizer-controller
    		I0518 11:04:46.483328    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (3.295478ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.483564    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (3.326365ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.483800    1267 wrap.go:42] GET /apis/project.openshift.io/v1/projects/default?timeout=1m0s: (4.398205ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38600]
    		I0518 11:04:46.483816    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (3.110416ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.484006    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (4.261696ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.484617    1267 wrap.go:42] GET /apis/network.openshift.io/v1?timeout=32s: (330.45µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.485638    1267 wrap.go:42] GET /apis/oauth.openshift.io/v1?timeout=32s: (404.339µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.486707    1267 wrap.go:42] GET /apis/project.openshift.io/v1?timeout=32s: (220.718µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.487269    1267 wrap.go:42] GET /apis/quota.openshift.io/v1?timeout=32s: (226.825µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.487673    1267 wrap.go:42] GET /api/v1/namespaces/default: (1.895612ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.487967    1267 wrap.go:42] GET /apis/project.openshift.io/v1/projects/default?timeout=1m0s: (3.365965ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38600]
    		I0518 11:04:46.488514    1267 wrap.go:42] GET /apis/route.openshift.io/v1?timeout=32s: (279.274µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.489387    1267 wrap.go:42] GET /apis/security.openshift.io/v1?timeout=32s: (364.482µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.490666    1267 wrap.go:42] GET /apis/template.openshift.io/v1?timeout=32s: (247.915µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.490997    1267 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.239132ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.491746    1267 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1/apiservices/v1.?timeout=1m0s: (2.724901ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38600]
    		I0518 11:04:46.492604    1267 wrap.go:42] GET /apis/user.openshift.io/v1?timeout=32s: (257.091µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38438]
    		I0518 11:04:46.493338    1267 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1/apiservices/v1.project.openshift.io?timeout=1m0s: (1.022142ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38600]
    		I0518 11:04:46.494855    1267 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1/apiservices/v1beta1.rbac.authorization.k8s.io?timeout=1m0s: (1.063748ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:38600]
    		INFO: 2018/05/18 11:04:46 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc43a13ed80
    		INFO: 2018/05/18 11:04:46 dialing to target with scheme: ""
    		INFO: 2018/05/18 11:04:46 could not get resolver for scheme: ""
    		INFO: 2018/05/18 11:04:46 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/18 11:04:46 balancerWrapper: got update addr from Notify: [{127.0.0.1:16658 <nil>}]
    		INFO: 2018/05/18 11:04:46 ccBalancerWrapper: new subconn: [{127.0.0.1:16658 0  <nil>}]
    		INFO: 2018/05/18 11:04:46 balancerWrapper: handle subconn state change: 0xc4332664d0, CONNECTING
    		INFO: 2018/05/18 11:04:46 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc43a13ed80
    		I0518 11:04:46.498714    1267 wrap.go:42] PUT /apis/rbac.authorization.k8s.io/v1/clusterroles/edit: (42.891625ms) 409 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:kube-system:clusterrole-aggregation-controller] 127.0.0.1:38598]
    		E0518 11:04:46.498911    1267 clusterroleaggregation_controller.go:180] edit failed with : Operation cannot be fulfilled on clusterroles.rbac.authorization.k8s.io "edit": the object has been modified; please apply your changes to the latest version and try again
    		INFO: 2018/05/18 11:04:46 balancerWrapper: handle subconn state change: 0xc4332664d0, READY
    		INFO: 2018/05/18 11:04:46 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc43a13ed80
    		INFO: 2018/05/18 11:04:46 balancerWrapper: got update addr from Notify: [{127.0.0.1:16658 <nil>}]
    		
    --- PASS: TestIntegration/TestAlwaysPullImagesOff (35.95s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		ount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:39374]
    		I0518 11:04:49.725427    1258 request.go:485] Throttling request took 61.704459ms, request: POST:https://127.0.0.1:20759/api/v1/namespaces/openshift/secrets
    		I0518 11:04:49.727389    1258 wrap.go:42] POST /api/v1/namespaces/openshift/secrets: (1.693818ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:39374]
    		I0518 11:04:49.759764    1258 wrap.go:42] PUT /api/v1/namespaces/openshift/serviceaccounts/deployer: (1.837666ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:39374]
    		I0518 11:04:49.789916    1258 request.go:485] Throttling request took 62.236088ms, request: POST:https://127.0.0.1:20759/api/v1/namespaces/integration/secrets
    		I0518 11:04:49.791846    1258 wrap.go:42] POST /api/v1/namespaces/integration/secrets: (1.712263ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:39374]
    		I0518 11:04:49.824630    1258 wrap.go:42] PUT /api/v1/namespaces/integration/serviceaccounts/builder: (2.076558ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:39374]
    		I0518 11:04:49.854453    1258 request.go:485] Throttling request took 62.294516ms, request: POST:https://127.0.0.1:20759/api/v1/namespaces/integration/secrets
    		I0518 11:04:49.856579    1258 wrap.go:42] POST /api/v1/namespaces/integration/secrets: (1.837506ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:39374]
    		I0518 11:04:49.889174    1258 wrap.go:42] PUT /api/v1/namespaces/integration/serviceaccounts/default: (2.206347ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:39374]
    		I0518 11:04:49.918964    1258 request.go:485] Throttling request took 62.091455ms, request: POST:https://127.0.0.1:20759/api/v1/namespaces/integration/secrets
    		I0518 11:04:49.922730    1258 wrap.go:42] POST /api/v1/namespaces/integration/secrets: (3.493052ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:39374]
    		I0518 11:04:49.953668    1258 wrap.go:42] PUT /api/v1/namespaces/integration/serviceaccounts/deployer: (2.178745ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:39374]
    		I0518 11:04:50.012944    1258 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.214964ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:39160]
    		I0518 11:04:50.012944    1258 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.247255ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:39160]
    		I0518 11:04:50.054357    1258 wrap.go:42] GET /api/v1/namespaces/kube-system/configmaps/kube-controller-manager: (1.167465ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:39160]
    		I0518 11:04:50.056887    1258 wrap.go:42] PUT /api/v1/namespaces/kube-system/configmaps/kube-controller-manager: (2.049621ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:39160]
    		I0518 11:04:50.057109    1258 leaderelection.go:199] successfully renewed lease kube-system/kube-controller-manager
    		I0518 11:04:50.216552    1258 wrap.go:42] GET /api/v1/namespaces/integration/serviceaccounts/default?timeout=1m0s: (1.382594ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:39384]
    		I0518 11:04:50.218366    1258 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-token-vmrw2?timeout=1m0s: (1.22633ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:39384]
    		I0518 11:04:50.219834    1258 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-dockercfg-cgzqs?timeout=1m0s: (963.275µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:39384]
    		I0518 11:04:50.222184    1258 wrap.go:42] GET /api/v1/namespaces/integration/limitranges: (989.707µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:39160]
    		I0518 11:04:50.222510    1258 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I0518 11:04:50.222567    1258 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I0518 11:04:50.224277    1258 wrap.go:42] GET /api/v1/namespaces/integration: (920.129µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:39160]
    		I0518 11:04:50.224586    1258 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0518 11:04:50.224647    1258 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0518 11:04:50.224661    1258 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0518 11:04:50.224668    1258 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0518 11:04:50.224689    1258 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0518 11:04:50.224702    1258 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0518 11:04:50.224712    1258 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0518 11:04:50.224722    1258 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0518 11:04:50.224727    1258 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0518 11:04:50.224738    1258 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0518 11:04:50.224756    1258 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0518 11:04:50.224766    1258 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0518 11:04:50.224771    1258 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0518 11:04:50.224785    1258 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0518 11:04:50.224856    1258 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I0518 11:04:50.226608    1258 wrap.go:42] POST /api/v1/namespaces/integration/pods?timeout=1m0s: (6.168219ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:39384]
    		I0518 11:04:50.226876    1258 factory.go:1147] About to try and schedule pod testm47sl
    		I0518 11:04:50.226892    1258 scheduler.go:439] Attempting to schedule pod: integration/testm47sl
    		I0518 11:04:50.226917    1258 scheduler.go:191] Failed to schedule pod: integration/testm47sl
    		I0518 11:04:50.226975    1258 factory.go:1262] Unable to schedule integration testm47sl: no nodes are registered to the cluster; waiting
    		I0518 11:04:50.227017    1258 factory.go:1375] Updating pod condition for integration/testm47sl to (PodScheduled==False)
    		I0518 11:04:50.227139    1258 pvc_protection_controller.go:276] Got event on pod integration/testm47sl
    		I0518 11:04:50.227177    1258 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"testm47sl"}
    		I0518 11:04:50.227210    1258 disruption.go:328] addPod called on pod "testm47sl"
    		I0518 11:04:50.227223    1258 disruption.go:403] No PodDisruptionBudgets found for pod testm47sl, PodDisruptionBudget controller will avoid syncing.
    		I0518 11:04:50.227232    1258 disruption.go:331] No matching pdb for pod "testm47sl"
    		I0518 11:04:50.227336    1258 backoff_utils.go:79] Backing off 1s
    		I0518 11:04:50.231849    1258 wrap.go:42] PUT /api/v1/namespaces/integration/pods/testm47sl/status: (4.431907ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:39160]
    		I0518 11:04:50.232064    1258 pvc_protection_controller.go:276] Got event on pod integration/testm47sl
    		I0518 11:04:50.232098    1258 disruption.go:340] updatePod called on pod "testm47sl"
    		I0518 11:04:50.232116    1258 disruption.go:403] No PodDisruptionBudgets found for pod testm47sl, PodDisruptionBudget controller will avoid syncing.
    		I0518 11:04:50.232124    1258 disruption.go:343] No matching pdb for pod "testm47sl"
    		I0518 11:04:50.232192    1258 factory.go:1147] About to try and schedule pod testm47sl
    		I0518 11:04:50.232216    1258 scheduler.go:439] Attempting to schedule pod: integration/testm47sl
    		I0518 11:04:50.232233    1258 scheduler.go:191] Failed to schedule pod: integration/testm47sl
    		I0518 11:04:50.232261    1258 factory.go:1262] Unable to schedule integration testm47sl: no nodes are registered to the cluster; waiting
    		I0518 11:04:50.232316    1258 factory.go:1375] Updating pod condition for integration/testm47sl to (PodScheduled==False)
    		W0518 11:04:50.232359    1258 factory.go:1304] Request for pod integration/testm47sl already in flight, abandoning
    		I0518 11:04:50.232515    1258 store.go:370] GuaranteedUpdate of /kubernetes.io/pods/integration/testm47sl failed because of a conflict, going to retry
    		I0518 11:04:50.233414    1258 wrap.go:42] POST /api/v1/namespaces/integration/events: (5.636543ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:39160]
    		I0518 11:04:50.235479    1258 disruption.go:340] updatePod called on pod "testm47sl"
    		I0518 11:04:50.235501    1258 disruption.go:403] No PodDisruptionBudgets found for pod testm47sl, PodDisruptionBudget controller will avoid syncing.
    		I0518 11:04:50.235510    1258 disruption.go:343] No matching pdb for pod "testm47sl"
    		I0518 11:04:50.235589    1258 pvc_protection_controller.go:276] Got event on pod integration/testm47sl
    		I0518 11:04:50.235624    1258 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testm47sl, uid 48bf4b7e-5a8b-11e8-a4db-0242ac110002, event type update
    		I0518 11:04:50.235637    1258 factory.go:1147] About to try and schedule pod testm47sl
    		I0518 11:04:50.235663    1258 scheduler.go:435] Skip schedule deleting pod: integration/testm47sl
    		I0518 11:04:50.235679    1258 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testm47sl, uid 48bf4b7e-5a8b-11e8-a4db-0242ac110002, event type update
    		I0518 11:04:50.237064    1258 wrap.go:42] DELETE /api/v1/namespaces/integration/pods/testm47sl?timeout=1m0s: (9.742237ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:39384]
    		I0518 11:04:50.237194    1258 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testm47sl, uid 48bf4b7e-5a8b-11e8-a4db-0242ac110002, event type delete
    		I0518 11:04:50.237320    1258 pvc_protection_controller.go:276] Got event on pod integration/testm47sl
    		I0518 11:04:50.237335    1258 deployment_controller.go:357] Pod testm47sl deleted.
    		I0518 11:04:50.237383    1258 disruption.go:369] deletePod called on pod "testm47sl"
    		I0518 11:04:50.237388    1258 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testm47sl, uid 48bf4b7e-5a8b-11e8-a4db-0242ac110002, event type delete
    		I0518 11:04:50.237396    1258 disruption.go:403] No PodDisruptionBudgets found for pod testm47sl, PodDisruptionBudget controller will avoid syncing.
    		I0518 11:04:50.237402    1258 disruption.go:372] No matching pdb for pod "testm47sl"
    		I0518 11:04:50.237416    1258 taint_manager.go:338] Noticed pod deletion: types.NamespacedName{Namespace:"integration", Name:"testm47sl"}
    		I0518 11:04:50.238042    1258 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I0518 11:04:50.238080    1258 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I0518 11:04:50.239313    1258 wrap.go:42] PATCH /api/v1/namespaces/integration/events/testm47sl.152fb858cc5a80f1: (4.984392ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:39160]
    		I0518 11:04:50.239919    1258 wrap.go:42] GET /api/v1/namespaces/integration: (1.0911ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:39160]
    		I0518 11:04:50.240127    1258 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0518 11:04:50.240164    1258 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0518 11:04:50.240176    1258 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0518 11:04:50.240186    1258 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0518 11:04:50.240212    1258 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0518 11:04:50.240231    1258 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0518 11:04:50.240247    1258 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0518 11:04:50.240261    1258 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0518 11:04:50.240269    1258 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0518 11:04:50.240288    1258 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0518 11:04:50.240317    1258 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0518 11:04:50.240332    1258 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0518 11:04:50.240340    1258 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0518 11:04:50.240385    1258 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0518 11:04:50.240434    1258 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I0518 11:04:50.242203    1258 factory.go:1147] About to try and schedule pod test5tlq5
    		I0518 11:04:50.242218    1258 scheduler.go:439] Attempting to schedule pod: integration/test5tlq5
    		I0518 11:04:50.242230    1258 scheduler.go:191] Failed to schedule pod: integration/test5tlq5
    		I0518 11:04:50.242241    1258 factory.go:1262] Unable to schedule integration test5tlq5: no nodes are registered to the cluster; waiting
    		I0518 11:04:50.242259    1258 factory.go:1375] Updating pod condition for integration/test5tlq5 to (PodScheduled==False)
    		I0518 11:04:50.242422    1258 wrap.go:42] POST /api/v1/namespaces/integration/pods?timeout=1m0s: (4.824907ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:39384]
    		I0518 11:04:50.242449    1258 backoff_utils.go:79] Backing off 1s
    		I0518 11:04:50.242794    1258 pvc_protection_controller.go:276] Got event on pod integration/test5tlq5
    		I0518 11:04:50.242819    1258 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"test5tlq5"}
    		I0518 11:04:50.242837    1258 disruption.go:328] addPod called on pod "test5tlq5"
    		I0518 11:04:50.242843    1258 wrap.go:42] POST /api/v1/namespaces/integration/events: (3.057859ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:39160]
    		I0518 11:04:50.242848    1258 disruption.go:403] No PodDisruptionBudgets found for pod test5tlq5, PodDisruptionBudget controller will avoid syncing.
    		I0518 11:04:50.242856    1258 disruption.go:331] No matching pdb for pod "test5tlq5"
    		INFO: 2018/05/18 11:04:50 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc43c44d200
    		INFO: 2018/05/18 11:04:50 dialing to target with scheme: ""
    		INFO: 2018/05/18 11:04:50 could not get resolver for scheme: ""
    		INFO: 2018/05/18 11:04:50 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/18 11:04:50 balancerWrapper: got update addr from Notify: [{127.0.0.1:11226 <nil>}]
    		INFO: 2018/05/18 11:04:50 ccBalancerWrapper: new subconn: [{127.0.0.1:11226 0  <nil>}]
    		INFO: 2018/05/18 11:04:50 balancerWrapper: handle subconn state change: 0xc429841950, CONNECTING
    		INFO: 2018/05/18 11:04:50 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc43c44d200
    		I0518 11:04:50.244561    1258 wrap.go:42] PUT /api/v1/namespaces/integration/pods/test5tlq5/status: (1.978866ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:39160]
    		I0518 11:04:50.244854    1258 pvc_protection_controller.go:276] Got event on pod integration/test5tlq5
    		I0518 11:04:50.244877    1258 disruption.go:340] updatePod called on pod "test5tlq5"
    		I0518 11:04:50.244888    1258 disruption.go:403] No PodDisruptionBudgets found for pod test5tlq5, PodDisruptionBudget controller will avoid syncing.
    		I0518 11:04:50.244896    1258 disruption.go:343] No matching pdb for pod "test5tlq5"
    		I0518 11:04:50.244978    1258 factory.go:1147] About to try and schedule pod test5tlq5
    		I0518 11:04:50.244991    1258 scheduler.go:439] Attempting to schedule pod: integration/test5tlq5
    		I0518 11:04:50.245006    1258 scheduler.go:191] Failed to schedule pod: integration/test5tlq5
    		I0518 11:04:50.245019    1258 factory.go:1262] Unable to schedule integration test5tlq5: no nodes are registered to the cluster; waiting
    		I0518 11:04:50.245055    1258 factory.go:1375] Updating pod condition for integration/test5tlq5 to (PodScheduled==False)
    		W0518 11:04:50.245083    1258 factory.go:1304] Request for pod integration/test5tlq5 already in flight, abandoning
    		I0518 11:04:50.247199    1258 wrap.go:42] POST /api/v1/namespaces/integration/events: (3.907902ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:39160]
    		I0518 11:04:50.251607    1258 wrap.go:42] PATCH /api/v1/namespaces/integration/events/test5tlq5.152fb858cd433606: (3.725783ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:39160]
    		INFO: 2018/05/18 11:04:50 balancerWrapper: handle subconn state change: 0xc429841950, READY
    		INFO: 2018/05/18 11:04:50 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc43c44d200
    		INFO: 2018/05/18 11:04:50 balancerWrapper: got update addr from Notify: [{127.0.0.1:11226 <nil>}]
    		
FAIL
exit status 1
FAIL	github.com/openshift/origin/test/integration/runner	2066.207s
[INFO] [11:04:51+0000] jUnit XML report placed at _output/scripts/test-integration/artifacts/gotest_report_iKiER.xml
Of 203 tests executed in 2066.206s, 199 succeeded, 4 failed, and 0 were skipped.

In suite "github.com/openshift/origin/test/integration/runner", test case "TestIntegration" failed:
runner_test.go:84: using existing binary

In suite "github.com/openshift/origin/test/integration/runner", test case "TestIntegration/TestConcurrentBuildImageChangeTriggerControllers" failed:
runner_test.go:129: FAILED TestConcurrentBuildImageChangeTriggerControllers, retrying:


In suite "github.com/openshift/origin/test/integration/runner", test case "TestIntegration/TestImageStreamImportScheduled" failed:
runner_test.go:129: FAILED TestImageStreamImportScheduled, retrying:

imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 1
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 2
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 3
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 4
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 5
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 6
imageimporter_test.go:771: unexpected object: <nil>


In suite "github.com/openshift/origin/test/integration/runner", test case "TestIntegration/TestOAuthServiceAccountClientEvent" failed:
runner_test.go:129: FAILED TestOAuthServiceAccountClientEvent, retrying:

oauth_serviceaccount_client_events_test.go:141: test-good-url: annotationPrefix serviceaccounts.openshift.io/oauth-redirecturi.one, annotation /oauthcallback
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:18879/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=http%!A(MISSING)%!F(MISSING)%!F(MISSING)127.0.0.1%!A(MISSING)42257%!F(MISSING)oauthcallback&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:18879/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=http%!A(MISSING)%!F(MISSING)%!F(MISSING)127.0.0.1%!A(MISSING)42257%!F(MISSING)oauthcallback&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:362: 302 Location: https://127.0.0.1:18879/oauth/authorize/approve?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=http%!A(MISSING)%!F(MISSING)%!F(MISSING)127.0.0.1%!A(MISSING)42257%!F(MISSING)oauthcallback&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project&then=..%!F(MISSING)authorize%!F(MISSING)client_id%!D(MISSING)system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default%!r(MISSING)edirect_uri%!D(MISSING)http%!A(MISSING)%!F(MISSING)%!F(MISSING)127.0.0.1%!A(MISSING)42257%!F(MISSING)oauthcallback%!r(MISSING)esponse_type%!D(MISSING)code%!s(MISSING)cope%!D(MISSING)user%!A(MISSING)info%!B(MISSING)role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:371: POST https://127.0.0.1:18879/oauth/authorize/approve
oauth_serviceaccount_client_events_test.go:362: 302 Location: https://127.0.0.1:18879/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=http%!A(MISSING)%!F(MISSING)%!F(MISSING)127.0.0.1%!A(MISSING)42257%!F(MISSING)oauthcallback&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:362: 302 Location: http://127.0.0.1:42257/oauthcallback?code=2ZRj3v143eHxk-YU-2DST7ckU2asuhQOW17ryce1pIc&state=
oauth_serviceaccount_client_events_test.go:141: test-bad-url: annotationPrefix serviceaccounts.openshift.io/oauth-redirecturi.one, annotation foo:foo
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:18879/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=foo%!A(MISSING)foo&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:396: Bad Request: HTTP/2.0 400 Bad Request
Content-Length: 155
Cache-Control: no-cache, no-store, max-age=0, must-revalidate
Content-Type: application/json
Date: Fri, 18 May 2018 10:49:52 GMT
Expires: Fri, 01 Jan 1990 00:00:00 GMT
Pragma: no-cache

{"error":"server_error","error_description":"The authorization server encountered an unexpected condition that prevented it from fulfilling the request."}
oauth_serviceaccount_client_events_test.go:141: test-bad-redirect-type-parse: annotationPrefix serviceaccounts.openshift.io/oauth-redirectreference.1, annotation {asdf":"adsf"}
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:18879/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=%!B(MISSING)asdf%3A%!a(MISSING)dsf%7D&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:396: Bad Request: HTTP/2.0 400 Bad Request
Content-Length: 155
Cache-Control: no-cache, no-store, max-age=0, must-revalidate
Content-Type: application/json
Date: Fri, 18 May 2018 10:49:52 GMT
Expires: Fri, 01 Jan 1990 00:00:00 GMT
Pragma: no-cache

{"error":"server_error","error_description":"The authorization server encountered an unexpected condition that prevented it from fulfilling the request."}
oauth_serviceaccount_client_events_test.go:141: test-bad-redirect-route-wrong-group: annotationPrefix serviceaccounts.openshift.io/oauth-redirectreference.1, annotation {"kind":"OAuthRedirectReference","apiVersion":"oauth.openshift.io/v1","metadata":{"creationTimestamp":null},"reference":{"group":"foo","kind":"Route","name":"route1"}}
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:18879/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=%!B(MISSING)%!k(MISSING)ind%3A%!O(MISSING)AuthRedirectReference%2C%!a(MISSING)piVersion%3A%!o(MISSING)auth.openshift.io%!F(MISSING)v1%2C%!m(MISSING)etadata%3A%!B(MISSING)%!c(MISSING)reationTimestamp%3Anull%!D(MISSING)%!C(MISSING)%!r(MISSING)eference%3A%!B(MISSING)%!g(MISSING)roup%3A%!f(MISSING)oo%2C%!k(MISSING)ind%3A%!R(MISSING)oute%2C%!n(MISSING)ame%3A%!r(MISSING)oute1%7D%!D(MISSING)%!A(MISSING)&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:396: Bad Request: HTTP/2.0 400 Bad Request
Content-Length: 155
Cache-Control: no-cache, no-store, max-age=0, must-revalidate
Content-Type: application/json
Date: Fri, 18 May 2018 10:49:52 GMT
Expires: Fri, 01 Jan 1990 00:00:00 GMT
Pragma: no-cache

{"error":"server_error","error_description":"The authorization server encountered an unexpected condition that prevented it from fulfilling the request."}
oauth_serviceaccount_client_events_test.go:163: test-bad-redirect-route-wrong-group: expected 1 events, found 0
[ERROR] [11:04:53+0000] hack/test-go.sh exited with code 1 after 00h 34m 29s
[ERROR] [11:04:53+0000] PID 1253: hack/test-integration.sh:18: `COVERAGE_SPEC=" " DETECT_RACES=false TMPDIR="${BASETMPDIR}" TIMEOUT=45m GOTEST_FLAGS="${gotest_flags}" "${OS_ROOT}/hack/test-go.sh" "test/integration/runner"` exited with status 1.
[INFO] [11:04:53+0000] 		Stack Trace: 
[INFO] [11:04:53+0000] 		  1: hack/test-integration.sh:18: `COVERAGE_SPEC=" " DETECT_RACES=false TMPDIR="${BASETMPDIR}" TIMEOUT=45m GOTEST_FLAGS="${gotest_flags}" "${OS_ROOT}/hack/test-go.sh" "test/integration/runner"`
[INFO] [11:04:53+0000]   Exiting with code 1.
make: *** [test-integration] Error 1
[WARNING] [11:04:54+0000] Copying _output/local/releases from the container failed!
[WARNING] [11:04:54+0000] Error response from daemon: lstat /var/lib/docker/overlay2/02799369de2224ebcc269c15527a5cbba1b9f7a64e7bfd26bf888ce837304f54/merged/go/src/github.com/openshift/origin/_output/local/releases: no such file or directory
[ERROR] [11:04:54+0000] PID 10830: hack/lib/build/environment.sh:172: `return "${exitcode}"` exited with status 2.
[INFO] [11:04:54+0000] 		Stack Trace: 
[INFO] [11:04:54+0000] 		  1: hack/lib/build/environment.sh:172: `return "${exitcode}"`
[INFO] [11:04:54+0000] 		  2: hack/lib/build/environment.sh:281: os::build::environment::withsource
[INFO] [11:04:54+0000] 		  3: hack/env:42: os::build::environment::run
[INFO] [11:04:54+0000]   Exiting with code 2.
++ export status=FAILURE
++ status=FAILURE
+ set +o xtrace
########## FINISHED STAGE: FAILURE: RUN INTEGRATION TESTS [00h 46m 31s] ##########
Build step 'Execute shell' marked build as failure
[PostBuildScript] - Execution post build scripts.
[workspace@2] $ /bin/bash /tmp/jenkins8546953729507298631.sh
########## STARTING STAGE: DOWNLOAD ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/gathered
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/gathered
+ mkdir -p /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/gathered
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo stat /data/src/github.com/openshift/origin/_output/scripts
  File: ‘/data/src/github.com/openshift/origin/_output/scripts’
  Size: 61        	Blocks: 0          IO Block: 4096   directory
Device: ca02h/51714d	Inode: 63256755    Links: 5
Access: (2755/drwxr-sr-x)  Uid: ( 1001/  origin)   Gid: ( 1003/origin-git)
Context: unconfined_u:object_r:container_file_t:s0
Access: 1970-01-01 00:00:00.000000000 +0000
Modify: 2018-05-18 10:19:56.000000000 +0000
Change: 2018-05-18 11:04:54.388308058 +0000
 Birth: -
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod -R o+rX /data/src/github.com/openshift/origin/_output/scripts
+ scp -r -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel:/data/src/github.com/openshift/origin/_output/scripts /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/gathered
+ tree /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/gathered
/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/gathered
└── scripts
    ├── shell
    │   ├── artifacts
    │   ├── logs
    │   │   ├── b20863ca830b6a484e2e17a02bfb162d61c028bb83fbff11e02050a20c2ce413.json
    │   │   └── scripts.log
    │   └── openshift.local.home
    ├── test-integration
    │   ├── artifacts
    │   │   ├── gotest_report_iKiER
    │   │   └── gotest_report_iKiER.xml
    │   ├── logs
    │   │   ├── raw_test_output.log
    │   │   ├── scripts.log
    │   │   └── test-go-err.log
    │   └── openshift.local.home
    └── test-tools
        ├── artifacts
        ├── logs
        │   ├── raw_test_output.log
        │   └── scripts.log
        └── openshift.local.home

13 directories, 9 files
+ exit 0
[workspace@2] $ /bin/bash /tmp/jenkins884317282060651379.sh
########## STARTING STAGE: GENERATE ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/generated
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/generated
+ mkdir /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/generated
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a 2>&1'
  WARNING: You're not using the default seccomp profile
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo cat /etc/sysconfig/docker /etc/sysconfig/docker-network /etc/sysconfig/docker-storage /etc/sysconfig/docker-storage-setup /etc/systemd/system/docker.service 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'oc get --raw /metrics --server=https://$( uname --nodename ):10250 --config=/etc/origin/master/admin.kubeconfig 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'oc get --raw /metrics --config=/etc/origin/master/admin.kubeconfig 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo df -T -h && sudo pvs && sudo vgs && sudo lvs && sudo findmnt --all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo yum list installed 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl --dmesg --no-pager --all --lines=all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl _PID=1 --no-pager --all --lines=all 2>&1'
+ tree /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/generated
/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/generated
├── avc_denials.log
├── containers.log
├── dmesg.log
├── docker.config
├── docker.info
├── filesystem.info
├── installed_packages.log
├── master-metrics.log
├── node-metrics.log
└── pid1.journal

0 directories, 10 files
+ exit 0
[workspace@2] $ /bin/bash /tmp/jenkins785160541737669943.sh
########## STARTING STAGE: FETCH SYSTEMD JOURNALS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/journals
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/journals
+ mkdir /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/journals
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit docker.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
+ tree /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/journals
/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/journals
├── dnsmasq.service
├── docker.service
└── systemd-journald.service

0 directories, 3 files
+ exit 0
[workspace@2] $ /bin/bash /tmp/jenkins1246404014129534464.sh
########## STARTING STAGE: ASSEMBLE GCS OUTPUT ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
+ trap 'exit 0' EXIT
+ mkdir -p gcs/artifacts gcs/artifacts/generated gcs/artifacts/journals gcs/artifacts/gathered
++ python -c 'import json; import urllib; print json.load(urllib.urlopen('\''https://ci.openshift.redhat.com/jenkins/job/test_pull_request_origin_integration/17930/api/json'\''))['\''result'\'']'
+ result=FAILURE
+ cat
++ date +%s
+ cat /var/lib/jenkins/jobs/test_pull_request_origin_integration/builds/17930/log
+ cp artifacts/generated/avc_denials.log artifacts/generated/containers.log artifacts/generated/dmesg.log artifacts/generated/docker.config artifacts/generated/docker.info artifacts/generated/filesystem.info artifacts/generated/installed_packages.log artifacts/generated/master-metrics.log artifacts/generated/node-metrics.log artifacts/generated/pid1.journal gcs/artifacts/generated/
+ cp artifacts/journals/dnsmasq.service artifacts/journals/docker.service artifacts/journals/systemd-journald.service gcs/artifacts/journals/
+ cp -r artifacts/gathered/scripts gcs/artifacts/
++ pwd
+ scp -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config -r /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/gcs openshiftdevel:/data
+ scp -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.config/gcloud/gcs-publisher-credentials.json openshiftdevel:/data/credentials.json
+ exit 0
[workspace@2] $ /bin/bash /tmp/jenkins8038033865763838394.sh
########## STARTING STAGE: PUSH THE ARTIFACTS AND METADATA ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ mktemp
+ script=/tmp/tmp.4XaFBFJYmh
+ cat
+ chmod +x /tmp/tmp.4XaFBFJYmh
+ scp -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.4XaFBFJYmh openshiftdevel:/tmp/tmp.4XaFBFJYmh
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 300 /tmp/tmp.4XaFBFJYmh"'
+ cd /home/origin
+ trap 'exit 0' EXIT
+ [[ -n {"type":"presubmit","job":"test_pull_request_origin_integration","buildid":"0933ceb7-5a84-11e8-ae4a-0a58ac100556","refs":{"org":"openshift","repo":"origin","base_ref":"master","base_sha":"a397fb8f1b365e4248a2c8226a227cb00346879e","pulls":[{"number":19759,"author":"wozniakjan","sha":"bbcd6fe94fd915e4a8d7b4db98ddbc7cfe1a0c4e"}]}} ]]
++ jq --compact-output .buildid
+ [[ "0933ceb7-5a84-11e8-ae4a-0a58ac100556" =~ ^"[0-9]+"$ ]]
+ echo 'Using BUILD_NUMBER'
Using BUILD_NUMBER
++ jq --compact-output '.buildid |= "17930"'
+ JOB_SPEC='{"type":"presubmit","job":"test_pull_request_origin_integration","buildid":"17930","refs":{"org":"openshift","repo":"origin","base_ref":"master","base_sha":"a397fb8f1b365e4248a2c8226a227cb00346879e","pulls":[{"number":19759,"author":"wozniakjan","sha":"bbcd6fe94fd915e4a8d7b4db98ddbc7cfe1a0c4e"}]}}'
+ docker run -e 'JOB_SPEC={"type":"presubmit","job":"test_pull_request_origin_integration","buildid":"17930","refs":{"org":"openshift","repo":"origin","base_ref":"master","base_sha":"a397fb8f1b365e4248a2c8226a227cb00346879e","pulls":[{"number":19759,"author":"wozniakjan","sha":"bbcd6fe94fd915e4a8d7b4db98ddbc7cfe1a0c4e"}]}}' -v /data:/data:z registry.svc.ci.openshift.org/ci/gcsupload:latest --dry-run=false --gcs-path=gs://origin-ci-test --gcs-credentials-file=/data/credentials.json --path-strategy=single --default-org=openshift --default-repo=origin /data/gcs/artifacts /data/gcs/build-log.txt /data/gcs/finished.json
Unable to find image 'registry.svc.ci.openshift.org/ci/gcsupload:latest' locally
Trying to pull repository registry.svc.ci.openshift.org/ci/gcsupload ... 
latest: Pulling from registry.svc.ci.openshift.org/ci/gcsupload
6d987f6f4279: Already exists
4cccebe844ee: Already exists
cafcba51f636: Pulling fs layer
cafcba51f636: Verifying Checksum
cafcba51f636: Download complete
cafcba51f636: Pull complete
Digest: sha256:c452798b56e3f4649c557c3ff7273126042065e3be152689a1ffb880638e617d
Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/gcsupload:latest
{"component":"gcsupload","level":"info","msg":"Gathering artifacts from artifact directory: /data/gcs/artifacts","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/avc_denials.log in artifact directory. Uploading as artifacts/generated/avc_denials.log\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/containers.log in artifact directory. Uploading as artifacts/generated/containers.log\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/dmesg.log in artifact directory. Uploading as artifacts/generated/dmesg.log\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.config in artifact directory. Uploading as artifacts/generated/docker.config\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.info in artifact directory. Uploading as artifacts/generated/docker.info\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/filesystem.info in artifact directory. Uploading as artifacts/generated/filesystem.info\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/installed_packages.log in artifact directory. Uploading as artifacts/generated/installed_packages.log\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/master-metrics.log in artifact directory. Uploading as artifacts/generated/master-metrics.log\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/node-metrics.log in artifact directory. Uploading as artifacts/generated/node-metrics.log\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/pid1.journal in artifact directory. Uploading as artifacts/generated/pid1.journal\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/dnsmasq.service in artifact directory. Uploading as artifacts/journals/dnsmasq.service\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/docker.service in artifact directory. Uploading as artifacts/journals/docker.service\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/systemd-journald.service in artifact directory. Uploading as artifacts/journals/systemd-journald.service\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/shell/logs/b20863ca830b6a484e2e17a02bfb162d61c028bb83fbff11e02050a20c2ce413.json in artifact directory. Uploading as artifacts/scripts/shell/logs/b20863ca830b6a484e2e17a02bfb162d61c028bb83fbff11e02050a20c2ce413.json\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/shell/logs/scripts.log in artifact directory. Uploading as artifacts/scripts/shell/logs/scripts.log\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/artifacts/gotest_report_iKiER in artifact directory. Uploading as artifacts/scripts/test-integration/artifacts/gotest_report_iKiER\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/artifacts/gotest_report_iKiER.xml in artifact directory. Uploading as artifacts/scripts/test-integration/artifacts/gotest_report_iKiER.xml\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/logs/raw_test_output.log in artifact directory. Uploading as artifacts/scripts/test-integration/logs/raw_test_output.log\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/logs/scripts.log in artifact directory. Uploading as artifacts/scripts/test-integration/logs/scripts.log\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/logs/test-go-err.log in artifact directory. Uploading as artifacts/scripts/test-integration/logs/test-go-err.log\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-tools/logs/raw_test_output.log in artifact directory. Uploading as artifacts/scripts/test-tools/logs/raw_test_output.log\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-tools/logs/scripts.log in artifact directory. Uploading as artifacts/scripts/test-tools/logs/scripts.log\n","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/shell/logs/scripts.log","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/pid1.journal","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/shell/logs/b20863ca830b6a484e2e17a02bfb162d61c028bb83fbff11e02050a20c2ce413.json","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration/17930.txt","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/docker.info","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/filesystem.info","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration/latest-build.txt","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/installed_packages.log","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-tools/logs/scripts.log","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/latest-build.txt","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/docker.config","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-integration/logs/raw_test_output.log","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-integration/logs/test-go-err.log","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/master-metrics.log","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/journals/systemd-journald.service","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/journals/docker.service","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-integration/artifacts/gotest_report_iKiER","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-integration/logs/scripts.log","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/build-log.txt","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/finished.json","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/avc_denials.log","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-integration/artifacts/gotest_report_iKiER.xml","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-tools/logs/raw_test_output.log","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/journals/dnsmasq.service","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/containers.log","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/dmesg.log","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/node-metrics.log","level":"info","msg":"Queued for upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/master-metrics.log","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-integration/logs/scripts.log","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/shell/logs/scripts.log","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-tools/logs/raw_test_output.log","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:47Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/installed_packages.log","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/latest-build.txt","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/journals/systemd-journald.service","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-tools/logs/scripts.log","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/pid1.journal","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/journals/dnsmasq.service","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/dmesg.log","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/finished.json","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/docker.info","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/filesystem.info","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/node-metrics.log","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/shell/logs/b20863ca830b6a484e2e17a02bfb162d61c028bb83fbff11e02050a20c2ce413.json","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/docker.config","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-integration/artifacts/gotest_report_iKiER","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-integration/logs/test-go-err.log","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/avc_denials.log","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration/latest-build.txt","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/generated/containers.log","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration/17930.txt","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/journals/docker.service","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/build-log.txt","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-integration/artifacts/gotest_report_iKiER.xml","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/17930/artifacts/scripts/test-integration/logs/raw_test_output.log","level":"info","msg":"Finished upload","time":"2018-05-18T11:05:48Z"}
{"component":"gcsupload","level":"info","msg":"Finished upload to GCS","time":"2018-05-18T11:05:48Z"}
+ exit 0
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: PUSH THE ARTIFACTS AND METADATA [00h 00m 12s] ##########
[workspace@2] $ /bin/bash /tmp/jenkins487479355098535519.sh
########## STARTING STAGE: DEPROVISION CLOUD RESOURCES ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
+ oct deprovision

PLAYBOOK: main.yml *************************************************************
4 plays in /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml

PLAY [ensure we have the parameters necessary to deprovision virtual hosts] ****

TASK [ensure all required variables are set] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:9
skipping: [localhost] => (item=origin_ci_inventory_dir)  => {
    "changed": false, 
    "generated_timestamp": "2018-05-18 07:05:50.103212", 
    "item": "origin_ci_inventory_dir", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_region)  => {
    "changed": false, 
    "generated_timestamp": "2018-05-18 07:05:50.106461", 
    "item": "origin_ci_aws_region", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

PLAY [deprovision virtual hosts in EC2] ****************************************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

TASK [deprovision a virtual EC2 host] ******************************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:28
included: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml for localhost

TASK [update the SSH configuration to remove AWS EC2 specifics] ****************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:2
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-05-18 07:05:50.866262", 
    "msg": ""
}

TASK [rename EC2 instance for termination reaper] ******************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:8
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-05-18 07:05:51.437739", 
    "msg": "Tags {'Name': 'oct-terminate'} created for resource i-04f3f5405b1a4332d."
}

TASK [tear down the EC2 instance] **********************************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:15
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-05-18 07:05:52.289640", 
    "instance_ids": [
        "i-04f3f5405b1a4332d"
    ], 
    "instances": [
        {
            "ami_launch_index": "0", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-0a6a96dc47c90e6ad"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-011c39732a5ea03d3"
                }
            }, 
            "dns_name": "ec2-52-87-201-252.compute-1.amazonaws.com", 
            "ebs_optimized": false, 
            "groups": {
                "sg-7e73221a": "default"
            }, 
            "hypervisor": "xen", 
            "id": "i-04f3f5405b1a4332d", 
            "image_id": "ami-0f07d2d9a03af96ec", 
            "instance_type": "m4.xlarge", 
            "kernel": null, 
            "key_name": "libra", 
            "launch_time": "2018-05-18T10:13:20.000Z", 
            "placement": "us-east-1d", 
            "private_dns_name": "ip-172-18-7-32.ec2.internal", 
            "private_ip": "172.18.7.32", 
            "public_dns_name": "ec2-52-87-201-252.compute-1.amazonaws.com", 
            "public_ip": "52.87.201.252", 
            "ramdisk": null, 
            "region": "us-east-1", 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "running", 
            "state_code": 16, 
            "tags": {
                "Name": "oct-terminate", 
                "openshift_etcd": "", 
                "openshift_master": "", 
                "openshift_node": ""
            }, 
            "tenancy": "default", 
            "virtualization_type": "hvm"
        }
    ], 
    "tagged_instances": []
}

TASK [remove the serialized host variables] ************************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:22
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-05-18 07:05:52.531482", 
    "path": "/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/host_vars/172.18.7.32.yml", 
    "state": "absent"
}

PLAY [deprovision virtual hosts locally manged by Vagrant] *********************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

PLAY [clean up local configuration for deprovisioned instances] ****************

TASK [remove inventory configuration directory] ********************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:61
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-05-18 07:05:52.972203", 
    "path": "/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory", 
    "state": "absent"
}

PLAY RECAP *********************************************************************
localhost                  : ok=8    changed=4    unreachable=0    failed=0   

+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPROVISION CLOUD RESOURCES [00h 00m 05s] ##########
Archiving artifacts
[BFA] Scanning build for known causes...
[BFA] Found failure cause(s):
[BFA] Job Stage Failed
[BFA] Unit or Integration Test Failed
[BFA] Done. 38s
[WS-CLEANUP] Deleting project workspace...[WS-CLEANUP] done
Finished: FAILURE