Console Output

Skipping 13,237 KB.. Full Log
    		INFO: 2018/05/22 10:13:53 balancerWrapper: handle subconn state change: 0xc42f506c10, READY
    		INFO: 2018/05/22 10:13:53 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc439f53f80
    		
    --- PASS: TestIntegration/TestApiGroupPreferredVersions (33.83s)
    	runner_test.go:187: 
    			master_routes_test.go:382: Looking for build api group in server group discovery
    		
    		=== OUTPUT
    		for resource "apps.openshift.io/v1, Resource=deploymentconfigs"
    		I0522 10:13:53.548881    1218 resource_quota_monitor.go:228] QuotaMonitor created object count evaluator for {apps.openshift.io deploymentconfigs}
    		I0522 10:13:53.548901    1218 resource_quota_monitor.go:176] QuotaMonitor using a shared informer for resource "/v1, Resource=limitranges"
    		I0522 10:13:53.548920    1218 resource_quota_monitor.go:228] QuotaMonitor created object count evaluator for { limitranges}
    		I0522 10:13:53.548938    1218 resource_quota_monitor.go:176] QuotaMonitor using a shared informer for resource "/v1, Resource=endpoints"
    		I0522 10:13:53.548959    1218 resource_quota_monitor.go:228] QuotaMonitor created object count evaluator for { endpoints}
    		I0522 10:13:53.548970    1218 resource_quota_monitor.go:243] quota synced monitors; added 36, kept 0, removed 0
    		E0522 10:13:53.548993    1218 reconciliation_controller.go:116] initial monitor sync has error: couldn't start monitor for resource {"network.openshift.io" "v1" "egressnetworkpolicies"}: unable to monitor quota for resource "network.openshift.io/v1, Resource=egressnetworkpolicies"
    		I0522 10:13:53.549008    1218 controller_manager.go:201] Started "openshift.io/cluster-quota-reconciliation"
    		I0522 10:13:53.549020    1218 controller_manager.go:191] Starting "openshift.io/build-config-change"
    		I0522 10:13:53.549148    1218 reconciliation_controller.go:129] Starting the cluster quota reconciliation controller
    		I0522 10:13:53.549160    1218 controller_utils.go:1019] Waiting for caches to sync for cluster resource quota controller
    		I0522 10:13:53.549185    1218 clusterquotamapping.go:160] Starting ClusterQuotaMappingController controller
    		I0522 10:13:53.549204    1218 resource_quota_monitor.go:301] QuotaMonitor running
    		I0522 10:13:53.550595    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/build-config-change-controller: (1.298898ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.552207    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra: (1.166128ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.565174    1218 wrap.go:42] POST /api/v1/namespaces/openshift-infra/serviceaccounts: (12.465454ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.568584    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/build-config-change-controller: (2.62845ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.575946    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (9.498471ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.578393    1218 get.go:238] Starting watch for /api/v1/namespaces/openshift-infra/secrets, rv=338 labels= fields=type=kubernetes.io/service-account-token timeout=1h33m48.068316423s
    		I0522 10:13:53.580354    1218 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (4.254695ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.606273    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/build-config-change-controller-token-bp6t6: (22.611022ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.608029    1218 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/build-config-change-controller: (26.462737ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.608801    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/build-config-change-controller: (1.774495ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.609335    1218 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (27.178092ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.609598    1218 client_builder.go:233] Verified credential for build-config-change-controller/openshift-infra
    		I0522 10:13:53.610129    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&resourceVersion=338&watch=true: (32.153855ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.611346    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/build-config-change-controller: (1.167004ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.614136    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (2.134399ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.615436    1218 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (537.337µs) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.620128    1218 client_builder.go:233] Verified credential for build-config-change-controller/openshift-infra
    		I0522 10:13:53.620267    1218 controller_manager.go:201] Started "openshift.io/build-config-change"
    		I0522 10:13:53.620284    1218 controller_manager.go:191] Starting "openshift.io/deployer"
    		I0522 10:13:53.621536    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/deployer-controller: (1.061002ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.623077    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra: (1.130306ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.627417    1218 wrap.go:42] POST /api/v1/namespaces/openshift-infra/serviceaccounts: (3.864291ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.630077    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/deployer-controller: (1.320872ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.630318    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (2.055753ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.631316    1218 get.go:238] Starting watch for /api/v1/namespaces/openshift-infra/secrets, rv=341 labels= fields=type=kubernetes.io/service-account-token timeout=1h58m26.414247015s
    		I0522 10:13:53.640531    1218 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (4.05866ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.646072    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/deployer-controller-token-njql9: (4.102728ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.648174    1218 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/deployer-controller: (6.984654ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.649270    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/deployer-controller: (2.510309ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.649590    1218 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (8.839808ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.649873    1218 client_builder.go:233] Verified credential for deployer-controller/openshift-infra
    		I0522 10:13:53.650317    1218 controller_manager.go:201] Started "openshift.io/deployer"
    		I0522 10:13:53.650331    1218 controller_manager.go:191] Starting "openshift.io/deploymentconfig"
    		I0522 10:13:53.650453    1218 factory.go:77] Starting deployer controller
    		I0522 10:13:53.650620    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&resourceVersion=341&watch=true: (19.691854ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.651700    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/deploymentconfig-controller: (1.035662ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.653202    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra: (1.085746ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.656288    1218 wrap.go:42] POST /api/v1/namespaces/openshift-infra/serviceaccounts: (2.601099ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.658657    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/deploymentconfig-controller: (1.589823ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.659921    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (2.374792ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.661013    1218 get.go:238] Starting watch for /api/v1/namespaces/openshift-infra/secrets, rv=344 labels= fields=type=kubernetes.io/service-account-token timeout=1h56m0.062420202s
    		I0522 10:13:53.668866    1218 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (5.013602ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.685716    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/deploymentconfig-controller-token-2kr5t: (1.690045ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.690239    1218 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/deploymentconfig-controller: (7.558649ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.693659    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/deploymentconfig-controller: (7.0055ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.694019    1218 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (12.427888ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.694287    1218 client_builder.go:233] Verified credential for deploymentconfig-controller/openshift-infra
    		I0522 10:13:53.694864    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&resourceVersion=344&watch=true: (34.253846ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.696815    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/deploymentconfig-controller: (1.903934ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.699684    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (2.423262ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.700968    1218 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (584.577µs) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.701344    1218 client_builder.go:233] Verified credential for deploymentconfig-controller/openshift-infra
    		I0522 10:13:53.701531    1218 controller_manager.go:201] Started "openshift.io/deploymentconfig"
    		I0522 10:13:53.701551    1218 controller_manager.go:191] Starting "openshift.io/templateinstancefinalizer"
    		I0522 10:13:53.701671    1218 factory.go:79] Starting deploymentconfig controller
    		I0522 10:13:53.702866    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-finalizer-controller: (1.093554ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.704459    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra: (1.120922ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.717197    1218 wrap.go:42] POST /api/v1/namespaces/openshift-infra/serviceaccounts: (12.118201ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.721350    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (2.82299ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.721617    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-finalizer-controller: (3.611413ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.729948    1218 get.go:238] Starting watch for /api/v1/namespaces/openshift-infra/secrets, rv=347 labels= fields=type=kubernetes.io/service-account-token timeout=1h15m17.783421518s
    		I0522 10:13:53.733816    1218 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (5.155972ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.738159    1218 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/admin?timeout=1m0s: (6.525692ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57480]
    		I0522 10:13:53.741727    1218 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-finalizer-controller: (5.995538ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.742632    1218 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/edit?timeout=1m0s: (2.470504ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57480]
    		I0522 10:13:53.743299    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/template-instance-finalizer-controller-token-2tc2h: (2.620139ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.746136    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-finalizer-controller: (2.383796ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.747067    1218 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/view?timeout=1m0s: (2.036436ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57480]
    		I0522 10:13:53.747385    1218 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (12.531123ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.747689    1218 client_builder.go:233] Verified credential for template-instance-finalizer-controller/openshift-infra
    		I0522 10:13:53.750559    1218 wrap.go:42] GET /api?timeout=1m0s: (584.439µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57480]
    		I0522 10:13:53.750801    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&resourceVersion=347&watch=true: (21.266796ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.752765    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-finalizer-controller: (1.918712ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.755543    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (2.318283ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.757049    1218 wrap.go:42] POST /apis/authentication.k8s.io/v1/tokenreviews: (604.822µs) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.757302    1218 client_builder.go:233] Verified credential for template-instance-finalizer-controller/openshift-infra
    		I0522 10:13:53.757554    1218 controller_manager.go:201] Started "openshift.io/templateinstancefinalizer"
    		I0522 10:13:53.757570    1218 controller_manager.go:191] Starting "openshift.io/unidling"
    		I0522 10:13:53.757687    1218 templateinstance_finalizer.go:214] TemplateInstanceFinalizer controller waiting for cache sync
    		I0522 10:13:53.758888    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/unidling-controller: (1.101349ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.760404    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra: (1.088859ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.762385    1218 wrap.go:42] GET /apis?timeout=1m0s: (10.345806ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57480]
    		I0522 10:13:53.763268    1218 wrap.go:42] POST /api/v1/namespaces/openshift-infra/serviceaccounts: (2.367848ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		INFO: 2018/05/22 10:13:53 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc436732ba0
    		INFO: 2018/05/22 10:13:53 dialing to target with scheme: ""
    		INFO: 2018/05/22 10:13:53 could not get resolver for scheme: ""
    		INFO: 2018/05/22 10:13:53 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/22 10:13:53 balancerWrapper: got update addr from Notify: [{127.0.0.1:12796 <nil>}]
    		INFO: 2018/05/22 10:13:53 ccBalancerWrapper: new subconn: [{127.0.0.1:12796 0  <nil>}]
    		INFO: 2018/05/22 10:13:53 balancerWrapper: handle subconn state change: 0xc42a80bca0, CONNECTING
    		INFO: 2018/05/22 10:13:53 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc436732ba0
    		I0522 10:13:53.777552    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/serviceaccounts/unidling-controller: (13.414138ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:57146]
    		I0522 10:13:53.778524    1218 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets?fieldSelector=type%3Dkubernetes.io%2Fservice-account-token&limit=500: (13.877288ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:57146]
    		I0522 10:13:53.784168    1218 get.go:238] Starting watch for /api/v1/namespaces/openshift-infra/secrets, rv=350 labels= fields=type=kubernetes.io/service-account-token timeout=1h39m56.892890544s
    		INFO: 2018/05/22 10:13:53 balancerWrapper: handle subconn state change: 0xc42a80bca0, READY
    		INFO: 2018/05/22 10:13:53 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc436732ba0
    		
    --- PASS: TestIntegration/TestAlwaysPullImagesOn (43.54s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		ation.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.614042    1197 wrap.go:42] PUT /api/v1/namespaces/kube-system/serviceaccounts/default: (2.520132ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.614590    1197 create_dockercfg_secrets.go:478] Creating dockercfg secret "default-dockercfg-2h78w" for service account integration/default
    		I0522 10:13:57.660280    1197 wrap.go:42] POST /api/v1/namespaces/openshift/secrets: (12.606245ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.690672    1197 wrap.go:42] PUT /api/v1/namespaces/openshift/serviceaccounts/default: (7.033956ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.691235    1197 create_dockercfg_secrets.go:478] Creating dockercfg secret "deployer-dockercfg-2t7q2" for service account integration/deployer
    		I0522 10:13:57.728897    1197 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (23.063658ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.774423    1197 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/default: (23.818593ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.778266    1197 wrap.go:42] POST /api/v1/namespaces/openshift-node/secrets: (4.187119ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.818848    1197 wrap.go:42] PUT /api/v1/namespaces/openshift-node/serviceaccounts/default: (16.020114ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.838567    1197 request.go:485] Throttling request took 59.732994ms, request: POST:https://127.0.0.1:16719/api/v1/namespaces/integration/secrets
    		I0522 10:13:57.844237    1197 wrap.go:42] POST /api/v1/namespaces/integration/secrets: (5.398008ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.872313    1197 wrap.go:42] PUT /api/v1/namespaces/integration/serviceaccounts/builder: (5.530859ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.898423    1197 request.go:485] Throttling request took 53.621742ms, request: POST:https://127.0.0.1:16719/api/v1/namespaces/integration/secrets
    		I0522 10:13:57.904266    1197 wrap.go:42] POST /api/v1/namespaces/integration/secrets: (5.558853ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.933264    1197 wrap.go:42] PUT /api/v1/namespaces/integration/serviceaccounts/default: (2.465431ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.965528    1197 request.go:485] Throttling request took 60.666386ms, request: POST:https://127.0.0.1:16719/api/v1/namespaces/integration/secrets
    		I0522 10:13:57.972940    1197 wrap.go:42] POST /api/v1/namespaces/integration/secrets: (2.198871ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:57.987972    1197 wrap.go:42] GET /api/v1/namespaces/kube-system/configmaps/kube-scheduler: (17.838432ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:44802]
    		I0522 10:13:58.008939    1197 wrap.go:42] PUT /api/v1/namespaces/kube-system/configmaps/kube-scheduler: (10.646177ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:44802]
    		I0522 10:13:58.009195    1197 leaderelection.go:199] successfully renewed lease kube-system/kube-scheduler
    		I0522 10:13:58.019440    1197 wrap.go:42] PUT /api/v1/namespaces/integration/serviceaccounts/deployer: (14.697164ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45460]
    		I0522 10:13:58.067745    1197 wrap.go:42] GET /api/v1/namespaces/integration/serviceaccounts/default?timeout=1m0s: (1.152338ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45476]
    		I0522 10:13:58.069616    1197 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-token-9c4n9?timeout=1m0s: (1.33086ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45476]
    		I0522 10:13:58.071390    1197 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-dockercfg-2h78w?timeout=1m0s: (1.097973ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45476]
    		I0522 10:13:58.074096    1197 wrap.go:42] GET /api/v1/namespaces/integration/limitranges: (1.134001ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44802]
    		I0522 10:13:58.074391    1197 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I0522 10:13:58.074447    1197 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I0522 10:13:58.076520    1197 wrap.go:42] GET /api/v1/namespaces/integration: (1.100024ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44802]
    		I0522 10:13:58.076801    1197 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:13:58.076841    1197 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 10:13:58.076851    1197 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:13:58.076859    1197 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:13:58.076876    1197 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:13:58.076886    1197 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:13:58.076902    1197 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 10:13:58.076910    1197 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:13:58.076917    1197 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:13:58.076929    1197 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:13:58.076948    1197 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 10:13:58.076956    1197 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:13:58.076963    1197 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:13:58.076980    1197 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 10:13:58.077062    1197 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I0522 10:13:58.098784    1197 wrap.go:42] POST /api/v1/namespaces/integration/pods?timeout=1m0s: (26.715016ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45476]
    		I0522 10:13:58.099523    1197 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (22.754021ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44802]
    		I0522 10:13:58.100601    1197 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"testmk9d8"}
    		I0522 10:13:58.100764    1197 factory.go:1147] About to try and schedule pod testmk9d8
    		I0522 10:13:58.100776    1197 scheduler.go:439] Attempting to schedule pod: integration/testmk9d8
    		I0522 10:13:58.100794    1197 scheduler.go:191] Failed to schedule pod: integration/testmk9d8
    		I0522 10:13:58.100850    1197 factory.go:1262] Unable to schedule integration testmk9d8: no nodes are registered to the cluster; waiting
    		I0522 10:13:58.100882    1197 factory.go:1375] Updating pod condition for integration/testmk9d8 to (PodScheduled==False)
    		I0522 10:13:58.101175    1197 pvc_protection_controller.go:276] Got event on pod integration/testmk9d8
    		I0522 10:13:58.101252    1197 disruption.go:328] addPod called on pod "testmk9d8"
    		I0522 10:13:58.101264    1197 disruption.go:403] No PodDisruptionBudgets found for pod testmk9d8, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:13:58.101271    1197 disruption.go:331] No matching pdb for pod "testmk9d8"
    		I0522 10:13:58.101301    1197 backoff_utils.go:79] Backing off 1s
    		I0522 10:13:58.106574    1197 wrap.go:42] PUT /api/v1/namespaces/integration/pods/testmk9d8/status: (4.869772ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:44802]
    		I0522 10:13:58.106670    1197 pvc_protection_controller.go:276] Got event on pod integration/testmk9d8
    		I0522 10:13:58.106727    1197 disruption.go:340] updatePod called on pod "testmk9d8"
    		I0522 10:13:58.106738    1197 disruption.go:403] No PodDisruptionBudgets found for pod testmk9d8, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:13:58.106745    1197 disruption.go:343] No matching pdb for pod "testmk9d8"
    		I0522 10:13:58.107175    1197 factory.go:1147] About to try and schedule pod testmk9d8
    		I0522 10:13:58.107202    1197 scheduler.go:439] Attempting to schedule pod: integration/testmk9d8
    		I0522 10:13:58.107218    1197 scheduler.go:191] Failed to schedule pod: integration/testmk9d8
    		I0522 10:13:58.107233    1197 factory.go:1262] Unable to schedule integration testmk9d8: no nodes are registered to the cluster; waiting
    		I0522 10:13:58.107258    1197 factory.go:1375] Updating pod condition for integration/testmk9d8 to (PodScheduled==False)
    		W0522 10:13:58.107302    1197 factory.go:1304] Request for pod integration/testmk9d8 already in flight, abandoning
    		I0522 10:13:58.121253    1197 store.go:370] GuaranteedUpdate of /kubernetes.io/pods/integration/testmk9d8 failed because of a conflict, going to retry
    		I0522 10:13:58.144005    1197 wrap.go:42] POST /api/v1/namespaces/integration/events: (42.155048ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:44802]
    		I0522 10:13:58.145632    1197 factory.go:1147] About to try and schedule pod testmk9d8
    		I0522 10:13:58.145658    1197 scheduler.go:435] Skip schedule deleting pod: integration/testmk9d8
    		I0522 10:13:58.145726    1197 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testmk9d8, uid d72d8142-5da8-11e8-b17b-0242ac110002, event type update
    		I0522 10:13:58.145897    1197 pvc_protection_controller.go:276] Got event on pod integration/testmk9d8
    		I0522 10:13:58.145946    1197 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testmk9d8, uid d72d8142-5da8-11e8-b17b-0242ac110002, event type update
    		I0522 10:13:58.145980    1197 disruption.go:340] updatePod called on pod "testmk9d8"
    		I0522 10:13:58.145990    1197 disruption.go:403] No PodDisruptionBudgets found for pod testmk9d8, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:13:58.145997    1197 disruption.go:343] No matching pdb for pod "testmk9d8"
    		I0522 10:13:58.147750    1197 wrap.go:42] DELETE /api/v1/namespaces/integration/pods/testmk9d8?timeout=1m0s: (44.776281ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45476]
    		I0522 10:13:58.147862    1197 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testmk9d8, uid d72d8142-5da8-11e8-b17b-0242ac110002, event type delete
    		I0522 10:13:58.147943    1197 taint_manager.go:338] Noticed pod deletion: types.NamespacedName{Namespace:"integration", Name:"testmk9d8"}
    		I0522 10:13:58.148042    1197 pvc_protection_controller.go:276] Got event on pod integration/testmk9d8
    		I0522 10:13:58.148077    1197 deployment_controller.go:357] Pod testmk9d8 deleted.
    		I0522 10:13:58.148099    1197 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testmk9d8, uid d72d8142-5da8-11e8-b17b-0242ac110002, event type delete
    		I0522 10:13:58.148125    1197 disruption.go:369] deletePod called on pod "testmk9d8"
    		I0522 10:13:58.148135    1197 disruption.go:403] No PodDisruptionBudgets found for pod testmk9d8, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:13:58.148142    1197 disruption.go:372] No matching pdb for pod "testmk9d8"
    		I0522 10:13:58.148856    1197 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I0522 10:13:58.148892    1197 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I0522 10:13:58.169761    1197 wrap.go:42] GET /api/v1/namespaces/integration: (12.543674ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44802]
    		I0522 10:13:58.171596    1197 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:13:58.171627    1197 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 10:13:58.171636    1197 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:13:58.171644    1197 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:13:58.171661    1197 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:13:58.171671    1197 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:13:58.171684    1197 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 10:13:58.171692    1197 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:13:58.171699    1197 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:13:58.171710    1197 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:13:58.171729    1197 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 10:13:58.171737    1197 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:13:58.171744    1197 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:13:58.171761    1197 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 10:13:58.171812    1197 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I0522 10:13:58.181536    1197 wrap.go:42] PATCH /api/v1/namespaces/integration/events/testmk9d8.1530efe470b1af0e: (35.413423ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:44802]
    		I0522 10:13:58.182560    1197 wrap.go:42] POST /api/v1/namespaces/integration/pods?timeout=1m0s: (34.267267ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45476]
    		I0522 10:13:58.183699    1197 factory.go:1147] About to try and schedule pod testm4hmh
    		I0522 10:13:58.183712    1197 scheduler.go:439] Attempting to schedule pod: integration/testm4hmh
    		I0522 10:13:58.183727    1197 scheduler.go:191] Failed to schedule pod: integration/testm4hmh
    		I0522 10:13:58.183742    1197 factory.go:1262] Unable to schedule integration testm4hmh: no nodes are registered to the cluster; waiting
    		I0522 10:13:58.183766    1197 factory.go:1375] Updating pod condition for integration/testm4hmh to (PodScheduled==False)
    		INFO: 2018/05/22 10:13:58 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc433edef00
    		I0522 10:13:58.184202    1197 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"testm4hmh"}
    		INFO: 2018/05/22 10:13:58 dialing to target with scheme: ""
    		INFO: 2018/05/22 10:13:58 could not get resolver for scheme: ""
    		I0522 10:13:58.184283    1197 backoff_utils.go:79] Backing off 1s
    		INFO: 2018/05/22 10:13:58 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/22 10:13:58 balancerWrapper: got update addr from Notify: [{127.0.0.1:22910 <nil>}]
    		INFO: 2018/05/22 10:13:58 ccBalancerWrapper: new subconn: [{127.0.0.1:22910 0  <nil>}]
    		I0522 10:13:58.184339    1197 pvc_protection_controller.go:276] Got event on pod integration/testm4hmh
    		I0522 10:13:58.184388    1197 disruption.go:328] addPod called on pod "testm4hmh"
    		I0522 10:13:58.184399    1197 disruption.go:403] No PodDisruptionBudgets found for pod testm4hmh, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:13:58.184406    1197 disruption.go:331] No matching pdb for pod "testm4hmh"
    		INFO: 2018/05/22 10:13:58 balancerWrapper: handle subconn state change: 0xc427233290, CONNECTING
    		INFO: 2018/05/22 10:13:58 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc433edef00
    		I0522 10:13:58.194055    1197 wrap.go:42] PUT /api/v1/namespaces/integration/pods/testm4hmh/status: (9.432247ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:44802]
    		I0522 10:13:58.194800    1197 pvc_protection_controller.go:276] Got event on pod integration/testm4hmh
    		I0522 10:13:58.194851    1197 disruption.go:340] updatePod called on pod "testm4hmh"
    		I0522 10:13:58.194862    1197 disruption.go:403] No PodDisruptionBudgets found for pod testm4hmh, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:13:58.194869    1197 disruption.go:343] No matching pdb for pod "testm4hmh"
    		INFO: 2018/05/22 10:13:58 balancerWrapper: handle subconn state change: 0xc427233290, READY
    		INFO: 2018/05/22 10:13:58 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc433edef00
    		
    --- PASS: TestIntegration/TestAggregator (33.12s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		.k8s.io/v1/clusterroles/view?timeout=1m0s: (10.783148ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45676]
    		I0522 10:14:26.674090    1255 request.go:485] Throttling request took 62.194524ms, request: PUT:https://127.0.0.1:17193/api/v1/namespaces/openshift-infra/serviceaccounts/service-serving-cert-controller
    		I0522 10:14:26.677007    1255 wrap.go:42] GET /apis/events.k8s.io/v1beta1?timeout=32s: (270.674µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.678429    1255 wrap.go:42] PUT /api/v1/namespaces/kube-system/configmaps/openshift-master-controllers: (12.172899ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.679820    1255 leaderelection.go:199] successfully renewed lease kube-system/openshift-master-controllers
    		I0522 10:14:26.680659    1255 wrap.go:42] GET /apis/authentication.k8s.io/v1?timeout=32s: (362.992µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.681564    1255 wrap.go:42] GET /apis/authentication.k8s.io/v1beta1?timeout=32s: (354.975µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.682428    1255 wrap.go:42] GET /apis/authorization.k8s.io/v1?timeout=32s: (393.074µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.683383    1255 wrap.go:42] GET /apis/authorization.k8s.io/v1beta1?timeout=32s: (345.943µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.686540    1255 wrap.go:42] GET /apis/autoscaling/v1?timeout=32s: (886.908µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.686805    1255 wrap.go:42] GET /api/v1/namespaces/default?timeout=1m0s: (7.356607ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45676]
    		I0522 10:14:26.687142    1255 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/image-trigger-controller-token-rwspr: (9.527023ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.692543    1255 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/horizontal-pod-autoscaler-token-7p42q: (12.276506ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.693069    1255 wrap.go:42] POST /apis/rbac.authorization.k8s.io/v1/namespaces/openshift-infra/rolebindings: (386.390912ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:default-rolebindings-controller] 127.0.0.1:45668]
    		I0522 10:14:26.694042    1255 wrap.go:42] GET /apis/autoscaling/v2beta1?timeout=32s: (348.327µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.694840    1255 wrap.go:42] GET /apis/batch/v1?timeout=32s: (380.077µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.695694    1255 wrap.go:42] GET /apis/batch/v1beta1?timeout=32s: (398.584µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.698287    1255 wrap.go:42] GET /apis/batch/v2alpha1?timeout=32s: (2.137074ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.698497    1255 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/service-serving-cert-controller: (18.485394ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45668]
    		I0522 10:14:26.699210    1255 create_dockercfg_secrets.go:441] Creating token secret "service-serving-cert-controller-token-c5xh4" for service account openshift-infra/service-serving-cert-controller
    		I0522 10:14:26.699376    1255 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/deployer-controller-token-m7fpc: (22.259821ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.699585    1255 request.go:485] Throttling request took 87.418119ms, request: PUT:https://127.0.0.1:17193/api/v1/namespaces/openshift-infra/serviceaccounts/serviceaccount-controller
    		I0522 10:14:26.699673    1255 wrap.go:42] GET /api/v1/namespaces/default: (9.197255ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45378]
    		I0522 10:14:26.700609    1255 wrap.go:42] GET /apis/certificates.k8s.io/v1beta1?timeout=32s: (364.997µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.704983    1255 wrap.go:42] GET /apis/networking.k8s.io/v1?timeout=32s: (331.841µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.705897    1255 trace.go:76] Trace[726075432]: "GuaranteedUpdate etcd3: *rbac.ClusterRole" (started: 2018-05-22 10:14:26.156760172 +0000 UTC m=+32.195022495) (total time: 549.067872ms):
    		Trace[726075432]: [16.751037ms] [16.751037ms] initial value restored
    		Trace[726075432]: [549.067872ms] [532.316835ms] END
    		I0522 10:14:26.705983    1255 trace.go:76] Trace[207262936]: "Update /apis/rbac.authorization.k8s.io/v1/clusterroles/edit" (started: 2018-05-22 10:14:26.154969758 +0000 UTC m=+32.193231973) (total time: 550.9768ms):
    		Trace[207262936]: [1.651121ms] [1.651121ms] About to convert to expected version
    		Trace[207262936]: [1.761817ms] [110.696µs] Conversion done
    		Trace[207262936]: [1.767582ms] [5.765µs] About to store object in database
    		Trace[207262936]: [550.9768ms] [549.209218ms] END
    		I0522 10:14:26.706070    1255 wrap.go:42] PUT /apis/rbac.authorization.k8s.io/v1/clusterroles/edit: (551.324879ms) 409 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:kube-system:clusterrole-aggregation-controller] 127.0.0.1:45668]
    		I0522 10:14:26.706499    1255 wrap.go:42] GET /api/v1/namespaces/default/services/kubernetes: (5.890809ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45378]
    		I0522 10:14:26.706623    1255 wrap.go:42] GET /api/v1/namespaces/default: (14.61006ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.707097    1255 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/serviceaccount-controller: (6.834324ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45668]
    		I0522 10:14:26.707385    1255 wrap.go:42] GET /apis/project.openshift.io/v1/projects/default?timeout=1m0s: (16.517162ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45676]
    		I0522 10:14:26.708154    1255 wrap.go:42] GET /apis/policy/v1beta1?timeout=32s: (381.931µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		E0522 10:14:26.708306    1255 clusterroleaggregation_controller.go:180] edit failed with : Operation cannot be fulfilled on clusterroles.rbac.authorization.k8s.io "edit": the object has been modified; please apply your changes to the latest version and try again
    		I0522 10:14:26.709590    1255 create_dockercfg_secrets.go:441] Creating token secret "serviceaccount-controller-token-m7qrh" for service account openshift-infra/serviceaccount-controller
    		I0522 10:14:26.710680    1255 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/node-bootstrapper-token-jzkpb: (8.267416ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.710968    1255 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/origin-namespace-controller-token-7lwdn: (8.041851ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.711221    1255 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/pv-recycler-controller-token-8xxzp: (10.609477ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.711938    1255 wrap.go:42] GET /apis/authorization.openshift.io/v1?timeout=32s: (483.292µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.713457    1255 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1?timeout=32s: (361.337µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.714348    1255 wrap.go:42] GET /apis/rbac.authorization.k8s.io/v1beta1?timeout=32s: (396.12µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.715168    1255 wrap.go:42] GET /apis/storage.k8s.io/v1?timeout=32s: (357.234µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.715928    1255 wrap.go:42] GET /apis/storage.k8s.io/v1beta1?timeout=32s: (334.673µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.716758    1255 wrap.go:42] GET /apis/admissionregistration.k8s.io/v1beta1?timeout=32s: (388.358µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.717623    1255 wrap.go:42] GET /apis/apiextensions.k8s.io/v1beta1?timeout=32s: (464.748µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.737190    1255 request.go:485] Throttling request took 123.122574ms, request: PUT:https://127.0.0.1:17193/api/v1/namespaces/openshift-infra/serviceaccounts/serviceaccount-pull-secrets-controller
    		I0522 10:14:26.746642    1255 wrap.go:42] GET /apis/apps.openshift.io/v1?timeout=32s: (27.647196ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.758559    1255 wrap.go:42] GET /apis/build.openshift.io/v1?timeout=32s: (947.522µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.764023    1255 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/image-trigger-controller-token-rwspr: (66.636259ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.764073    1255 wrap.go:42] GET /apis/image.openshift.io/v1?timeout=32s: (901.498µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.767212    1255 request.go:485] Throttling request took 143.232895ms, request: POST:https://127.0.0.1:17193/api/v1/namespaces/openshift-infra/secrets
    		I0522 10:14:26.779796    1255 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/image-import-controller-token-ffg6x: (88.008097ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.808315    1255 request.go:485] Throttling request took 142.07407ms, request: POST:https://127.0.0.1:17193/api/v1/namespaces/openshift-infra/secrets
    		I0522 10:14:26.809415    1255 wrap.go:42] GET /api/v1/namespaces/default: (97.233878ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.810181    1255 wrap.go:42] GET /apis/network.openshift.io/v1?timeout=32s: (547.247µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.812696    1255 wrap.go:42] GET /apis/project.openshift.io/v1/projects/default?timeout=1m0s: (102.240931ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45676]
    		I0522 10:14:26.817717    1255 wrap.go:42] GET /apis/oauth.openshift.io/v1?timeout=32s: (515.277µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.820738    1255 wrap.go:42] GET /apis/project.openshift.io/v1?timeout=32s: (507.585µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.822027    1255 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (51.498312ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45668]
    		I0522 10:14:26.823120    1255 wrap.go:42] GET /apis/quota.openshift.io/v1?timeout=32s: (456.752µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.824421    1255 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/serviceaccount-pull-secrets-controller: (73.688758ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45668]
    		I0522 10:14:26.824551    1255 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/ingress-to-route-controller-token-vdzlw: (55.856152ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.827994    1255 create_dockercfg_secrets.go:441] Creating token secret "serviceaccount-pull-secrets-controller-token-cgm82" for service account openshift-infra/serviceaccount-pull-secrets-controller
    		I0522 10:14:26.832088    1255 wrap.go:42] GET /apis/route.openshift.io/v1?timeout=32s: (927.026µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.834839    1255 wrap.go:42] GET /apis/security.openshift.io/v1?timeout=32s: (907.7µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.835678    1255 request.go:485] Throttling request took 136.246499ms, request: POST:https://127.0.0.1:17193/api/v1/namespaces/openshift-infra/secrets
    		I0522 10:14:26.846752    1255 wrap.go:42] GET /apis/template.openshift.io/v1?timeout=32s: (694.178µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.859890    1255 wrap.go:42] GET /apis/user.openshift.io/v1?timeout=32s: (559.204µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45334]
    		I0522 10:14:26.867100    1255 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1/apiservices/v1.?timeout=1m0s: (48.146704ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45676]
    		I0522 10:14:26.867338    1255 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (48.847759ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45668]
    		I0522 10:14:26.867695    1255 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/pv-recycler-controller-token-8xxzp: (149.135151ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.868186    1255 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (30.794695ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45668]
    		I0522 10:14:26.870804    1255 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/origin-namespace-controller-token-7lwdn: (130.025731ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.871084    1255 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/secrets/node-bootstrapper-token-jzkpb: (153.533726ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.877300    1255 request.go:485] Throttling request took 167.501241ms, request: POST:https://127.0.0.1:17193/api/v1/namespaces/openshift-infra/secrets
    		I0522 10:14:26.884608    1255 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1/apiservices/v1.project.openshift.io?timeout=1m0s: (14.484669ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45676]
    		I0522 10:14:26.884810    1255 request.go:485] Throttling request took 59.650207ms, request: PUT:https://127.0.0.1:17193/api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-controller
    		I0522 10:14:26.890414    1255 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/namespace-security-allocation-controller-token-hvc4k: (43.438082ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.895186    1255 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (16.822997ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45668]
    		I0522 10:14:26.906159    1255 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/resourcequota-controller-token-nsjfh: (7.927753ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.906786    1255 wrap.go:42] GET /api/v1/namespaces/openshift-infra/secrets/service-ingress-ip-controller-token-kxsj6: (7.925456ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/tokens-controller] 127.0.0.1:45334]
    		I0522 10:14:26.907396    1255 wrap.go:42] GET /apis/apiregistration.k8s.io/v1beta1/apiservices/v1beta1.rbac.authorization.k8s.io?timeout=1m0s: (11.261889ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:45676]
    		INFO: 2018/05/22 10:14:26 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc43ea1a0c0
    		INFO: 2018/05/22 10:14:26 dialing to target with scheme: ""
    		INFO: 2018/05/22 10:14:26 could not get resolver for scheme: ""
    		INFO: 2018/05/22 10:14:26 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/22 10:14:26 balancerWrapper: got update addr from Notify: [{127.0.0.1:25393 <nil>}]
    		INFO: 2018/05/22 10:14:26 ccBalancerWrapper: new subconn: [{127.0.0.1:25393 0  <nil>}]
    		INFO: 2018/05/22 10:14:26 balancerWrapper: handle subconn state change: 0xc42ab37f40, CONNECTING
    		INFO: 2018/05/22 10:14:26 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc43ea1a0c0
    		I0522 10:14:26.924027    1255 wrap.go:42] PUT /api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-controller: (38.142003ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45668]
    		I0522 10:14:26.936217    1255 create_dockercfg_secrets.go:441] Creating token secret "template-instance-controller-token-z74wj" for service account openshift-infra/template-instance-controller
    		I0522 10:14:26.945076    1255 request.go:485] Throttling request took 116.998823ms, request: POST:https://127.0.0.1:17193/api/v1/namespaces/openshift-infra/secrets
    		I0522 10:14:26.949280    1255 request.go:485] Throttling request took 80.734737ms, request: PUT:https://127.0.0.1:17193/api/v1/namespaces/openshift-infra/serviceaccounts/template-instance-finalizer-controller
    		I0522 10:14:26.959097    1255 wrap.go:42] POST /api/v1/namespaces/openshift-infra/secrets: (13.430647ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:45668]
    		INFO: 2018/05/22 10:14:26 balancerWrapper: handle subconn state change: 0xc42ab37f40, READY
    		INFO: 2018/05/22 10:14:26 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc43ea1a0c0
    		
    --- PASS: TestIntegration/TestAllowedSCCViaRBAC (39.23s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		espace project1
    		I0522 10:14:29.932214    1237 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 10:14:29.932220    1237 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 10:14:29.932242    1237 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 10:14:29.932255    1237 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 10:14:29.932274    1237 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0522 10:14:29.932286    1237 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 10:14:29.932291    1237 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 10:14:29.932298    1237 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 10:14:29.932310    1237 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0522 10:14:29.932320    1237 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 10:14:29.932324    1237 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 10:14:29.932336    1237 admission.go:217] validating pod test3 (generate: ) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 10:14:29.932499    1237 admission.go:170] pod test3 (generate: ) validated against provider hostaccess
    		I0522 10:14:29.934241    1237 wrap.go:42] POST /api/v1/namespaces/project1/pods?timeout=1m0s: (4.979924ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44880]
    		I0522 10:14:29.934550    1237 factory.go:1147] About to try and schedule pod test3
    		I0522 10:14:29.934568    1237 scheduler.go:439] Attempting to schedule pod: project1/test3
    		I0522 10:14:29.934589    1237 scheduler.go:191] Failed to schedule pod: project1/test3
    		I0522 10:14:29.934647    1237 factory.go:1262] Unable to schedule project1 test3: no nodes are registered to the cluster; waiting
    		I0522 10:14:29.934675    1237 disruption.go:328] addPod called on pod "test3"
    		I0522 10:14:29.934687    1237 disruption.go:403] No PodDisruptionBudgets found for pod test3, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:14:29.934692    1237 disruption.go:331] No matching pdb for pod "test3"
    		I0522 10:14:29.934689    1237 factory.go:1375] Updating pod condition for project1/test3 to (PodScheduled==False)
    		I0522 10:14:29.934723    1237 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"project1", Name:"test3"}
    		I0522 10:14:29.934826    1237 backoff_utils.go:79] Backing off 1s
    		I0522 10:14:29.934835    1237 pvc_protection_controller.go:276] Got event on pod project1/test3
    		I0522 10:14:29.935620    1237 admission.go:97] getting security context constraints for pod test4 (generate: ) in namespace project2 with user info &{user1 e5b007a9-5da8-11e8-b90d-0242ac110002 [system:authenticated:oauth system:authenticated] map[scopes.authorization.openshift.io:[]]}
    		I0522 10:14:29.936117    1237 admission.go:108] getting security context constraints for pod test4 (generate: ) with service account info &{system:serviceaccount:project2:default  [system:serviceaccounts system:serviceaccounts:project2] map[]}
    		I0522 10:14:29.937819    1237 disruption.go:340] updatePod called on pod "test3"
    		I0522 10:14:29.937828    1237 wrap.go:42] PUT /api/v1/namespaces/project1/pods/test3/status: (2.873826ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:44572]
    		I0522 10:14:29.937841    1237 disruption.go:403] No PodDisruptionBudgets found for pod test3, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:14:29.937849    1237 disruption.go:343] No matching pdb for pod "test3"
    		I0522 10:14:29.938014    1237 pvc_protection_controller.go:276] Got event on pod project1/test3
    		I0522 10:14:29.938028    1237 wrap.go:42] GET /api/v1/namespaces/project2: (1.352654ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44572]
    		I0522 10:14:29.938204    1237 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I0522 10:14:29.938221    1237 factory.go:1147] About to try and schedule pod test3
    		I0522 10:14:29.938226    1237 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 10:14:29.938231    1237 scheduler.go:439] Attempting to schedule pod: project1/test3
    		I0522 10:14:29.938234    1237 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0522 10:14:29.938246    1237 scheduler.go:191] Failed to schedule pod: project1/test3
    		I0522 10:14:29.938279    1237 factory.go:1262] Unable to schedule project1 test3: no nodes are registered to the cluster; waiting
    		I0522 10:14:29.938321    1237 factory.go:1375] Updating pod condition for project1/test3 to (PodScheduled==False)
    		I0522 10:14:29.938251    1237 admission.go:217] validating pod test4 (generate: ) against providers restricted
    		I0522 10:14:29.938374    1237 admission.go:179] unable to validate pod test4 (generate: ) against any security context constraint: [provider restricted: .spec.securityContext.hostPID: Invalid value: true: Host PID is not allowed to be used spec.containers[0].securityContext.hostPID: Invalid value: true: Host PID is not allowed to be used]
    		W0522 10:14:29.938446    1237 factory.go:1304] Request for pod project1/test3 already in flight, abandoning
    		I0522 10:14:29.938527    1237 wrap.go:42] POST /api/v1/namespaces/project2/pods?timeout=1m0s: (3.594966ms) 403 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44880]
    		I0522 10:14:29.938906    1237 wrap.go:42] POST /api/v1/namespaces/project1/events: (3.703963ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:44572]
    		I0522 10:14:29.939542    1237 admission.go:97] getting security context constraints for pod test5 (generate: ) in namespace project1 with user info &{user2 e5b94426-5da8-11e8-b90d-0242ac110002 [system:authenticated:oauth system:authenticated] map[scopes.authorization.openshift.io:[]]}
    		I0522 10:14:29.939761    1237 admission.go:108] getting security context constraints for pod test5 (generate: ) with service account info &{system:serviceaccount:project1:default  [system:serviceaccounts system:serviceaccounts:project1] map[]}
    		I0522 10:14:29.942910    1237 wrap.go:42] GET /api/v1/namespaces/project1: (2.406508ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44572]
    		I0522 10:14:29.943089    1237 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 10:14:29.943116    1237 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0522 10:14:29.943122    1237 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 10:14:29.943127    1237 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 10:14:29.943138    1237 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 10:14:29.943149    1237 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 10:14:29.943159    1237 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0522 10:14:29.943168    1237 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 10:14:29.943173    1237 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 10:14:29.943180    1237 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 10:14:29.943196    1237 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace project1
    		I0522 10:14:29.943204    1237 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 10:14:29.943212    1237 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace project1
    		I0522 10:14:29.943229    1237 admission.go:217] validating pod test5 (generate: ) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 10:14:29.943346    1237 admission.go:170] pod test5 (generate: ) validated against provider hostaccess
    		I0522 10:14:29.943912    1237 wrap.go:42] PATCH /api/v1/namespaces/project1/events/test3.1530efebda22f566: (3.972508ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:44572]
    		I0522 10:14:29.944725    1237 wrap.go:42] POST /api/v1/namespaces/project1/pods?timeout=1m0s: (5.624115ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44880]
    		I0522 10:14:29.944897    1237 factory.go:1147] About to try and schedule pod test5
    		I0522 10:14:29.944918    1237 scheduler.go:439] Attempting to schedule pod: project1/test5
    		I0522 10:14:29.944933    1237 scheduler.go:191] Failed to schedule pod: project1/test5
    		I0522 10:14:29.944954    1237 factory.go:1262] Unable to schedule project1 test5: no nodes are registered to the cluster; waiting
    		I0522 10:14:29.944985    1237 factory.go:1375] Updating pod condition for project1/test5 to (PodScheduled==False)
    		I0522 10:14:29.945006    1237 disruption.go:328] addPod called on pod "test5"
    		I0522 10:14:29.945033    1237 disruption.go:403] No PodDisruptionBudgets found for pod test5, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:14:29.945042    1237 disruption.go:331] No matching pdb for pod "test5"
    		I0522 10:14:29.945117    1237 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"project1", Name:"test5"}
    		I0522 10:14:29.945149    1237 backoff_utils.go:79] Backing off 1s
    		I0522 10:14:29.945186    1237 pvc_protection_controller.go:276] Got event on pod project1/test5
    		I0522 10:14:29.945790    1237 admission.go:97] getting security context constraints for pod test6 (generate: ) in namespace project2 with user info &{user2 e5b94426-5da8-11e8-b90d-0242ac110002 [system:authenticated:oauth system:authenticated] map[scopes.authorization.openshift.io:[]]}
    		I0522 10:14:29.946024    1237 admission.go:108] getting security context constraints for pod test6 (generate: ) with service account info &{system:serviceaccount:project2:default  [system:serviceaccounts system:serviceaccounts:project2] map[]}
    		I0522 10:14:29.947921    1237 wrap.go:42] PUT /api/v1/namespaces/project1/pods/test5/status: (2.598764ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:44572]
    		I0522 10:14:29.948219    1237 disruption.go:340] updatePod called on pod "test5"
    		I0522 10:14:29.948238    1237 disruption.go:403] No PodDisruptionBudgets found for pod test5, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:14:29.948246    1237 disruption.go:343] No matching pdb for pod "test5"
    		I0522 10:14:29.948283    1237 pvc_protection_controller.go:276] Got event on pod project1/test5
    		I0522 10:14:29.948343    1237 factory.go:1147] About to try and schedule pod test5
    		I0522 10:14:29.948355    1237 scheduler.go:439] Attempting to schedule pod: project1/test5
    		I0522 10:14:29.948367    1237 scheduler.go:191] Failed to schedule pod: project1/test5
    		I0522 10:14:29.948386    1237 factory.go:1262] Unable to schedule project1 test5: no nodes are registered to the cluster; waiting
    		I0522 10:14:29.948418    1237 factory.go:1375] Updating pod condition for project1/test5 to (PodScheduled==False)
    		W0522 10:14:29.948442    1237 factory.go:1304] Request for pod project1/test5 already in flight, abandoning
    		I0522 10:14:29.949793    1237 wrap.go:42] GET /api/v1/namespaces/project2: (3.018138ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44572]
    		I0522 10:14:29.949938    1237 wrap.go:42] POST /api/v1/namespaces/project1/events: (4.563979ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:44572]
    		I0522 10:14:29.950096    1237 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 10:14:29.950128    1237 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I0522 10:14:29.950140    1237 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 10:14:29.950154    1237 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0522 10:14:29.950175    1237 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 10:14:29.950188    1237 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 10:14:29.950204    1237 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I0522 10:14:29.950214    1237 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 10:14:29.950222    1237 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0522 10:14:29.950236    1237 matcher.go:342] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0522 10:14:29.950254    1237 matcher.go:279] got preallocated values for min: 1000070000, max: 1000079999 for uid range in namespace project2
    		I0522 10:14:29.950267    1237 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 10:14:29.950274    1237 matcher.go:322] got preallocated value for groups: 1000070000/10000 in namespace project2
    		I0522 10:14:29.950297    1237 admission.go:217] validating pod test6 (generate: ) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 10:14:29.950436    1237 admission.go:170] pod test6 (generate: ) validated against provider hostaccess
    		I0522 10:14:29.951851    1237 wrap.go:42] POST /api/v1/namespaces/project2/pods?timeout=1m0s: (6.482901ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44880]
    		I0522 10:14:29.951857    1237 factory.go:1147] About to try and schedule pod test6
    		I0522 10:14:29.951861    1237 disruption.go:328] addPod called on pod "test6"
    		I0522 10:14:29.951871    1237 scheduler.go:439] Attempting to schedule pod: project2/test6
    		I0522 10:14:29.951879    1237 disruption.go:403] No PodDisruptionBudgets found for pod test6, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:14:29.951887    1237 scheduler.go:191] Failed to schedule pod: project2/test6
    		I0522 10:14:29.951905    1237 factory.go:1262] Unable to schedule project2 test6: no nodes are registered to the cluster; waiting
    		I0522 10:14:29.951915    1237 pvc_protection_controller.go:276] Got event on pod project2/test6
    		I0522 10:14:29.951931    1237 factory.go:1375] Updating pod condition for project2/test6 to (PodScheduled==False)
    		I0522 10:14:29.951887    1237 disruption.go:331] No matching pdb for pod "test6"
    		I0522 10:14:29.951983    1237 backoff_utils.go:79] Backing off 1s
    		I0522 10:14:29.951928    1237 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"project2", Name:"test6"}
    		I0522 10:14:29.953328    1237 controller.go:537] quota admission added evaluator for: {security.openshift.io podsecuritypolicyselfsubjectreviews}
    		I0522 10:14:29.954870    1237 wrap.go:42] PATCH /api/v1/namespaces/project1/events/test5.1530efebdac02c9b: (4.306608ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:44572]
    		I0522 10:14:29.955111    1237 wrap.go:42] GET /api/v1/namespaces/project1: (1.14046ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44572]
    		I0522 10:14:29.955130    1237 disruption.go:340] updatePod called on pod "test6"
    		I0522 10:14:29.955145    1237 disruption.go:403] No PodDisruptionBudgets found for pod test6, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:14:29.955153    1237 disruption.go:343] No matching pdb for pod "test6"
    		I0522 10:14:29.955112    1237 wrap.go:42] PUT /api/v1/namespaces/project2/pods/test6/status: (2.722675ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:44572]
    		I0522 10:14:29.955293    1237 pvc_protection_controller.go:276] Got event on pod project2/test6
    		I0522 10:14:29.955526    1237 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace project1
    		I0522 10:14:29.955580    1237 factory.go:1147] About to try and schedule pod test6
    		I0522 10:14:29.955593    1237 scheduler.go:439] Attempting to schedule pod: project2/test6
    		I0522 10:14:29.955608    1237 scheduler.go:191] Failed to schedule pod: project2/test6
    		I0522 10:14:29.955625    1237 factory.go:1262] Unable to schedule project2 test6: no nodes are registered to the cluster; waiting
    		I0522 10:14:29.955652    1237 factory.go:1375] Updating pod condition for project2/test6 to (PodScheduled==False)
    		W0522 10:14:29.955690    1237 factory.go:1304] Request for pod project2/test6 already in flight, abandoning
    		I0522 10:14:29.955725    1237 wrap.go:42] POST /apis/security.openshift.io/v1/namespaces/project1/podsecuritypolicyselfsubjectreviews?timeout=1m0s: (3.087118ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44880]
    		I0522 10:14:29.959159    1237 wrap.go:42] GET /api/v1/namespaces/project2: (1.748677ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44572]
    		I0522 10:14:29.959312    1237 wrap.go:42] POST /api/v1/namespaces/project2/events: (3.838439ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:44572]
    		I0522 10:14:29.959609    1237 matcher.go:292] got preallocated value for level: s0:c8,c7 for selinux options in namespace project2
    		I0522 10:14:29.959761    1237 wrap.go:42] POST /apis/security.openshift.io/v1/namespaces/project2/podsecuritypolicyselfsubjectreviews?timeout=1m0s: (3.519074ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:44880]
    		INFO: 2018/05/22 10:14:29 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc433dedb00
    		INFO: 2018/05/22 10:14:29 dialing to target with scheme: ""
    		INFO: 2018/05/22 10:14:29 could not get resolver for scheme: ""
    		INFO: 2018/05/22 10:14:29 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/22 10:14:29 balancerWrapper: got update addr from Notify: [{127.0.0.1:28897 <nil>}]
    		INFO: 2018/05/22 10:14:29 ccBalancerWrapper: new subconn: [{127.0.0.1:28897 0  <nil>}]
    		INFO: 2018/05/22 10:14:29 balancerWrapper: handle subconn state change: 0xc42b712660, CONNECTING
    		INFO: 2018/05/22 10:14:29 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc433dedb00
    		I0522 10:14:29.963398    1237 wrap.go:42] PATCH /api/v1/namespaces/project2/events/test6.1530efebdb2a2406: (3.315306ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:44572]
    		INFO: 2018/05/22 10:14:29 balancerWrapper: handle subconn state change: 0xc42b712660, READY
    		INFO: 2018/05/22 10:14:29 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc433dedb00
    		INFO: 2018/05/22 10:14:29 balancerWrapper: got update addr from Notify: [{127.0.0.1:28897 <nil>}]
    		
    --- PASS: TestIntegration/TestAlwaysPullImagesOff (39.25s)
    	runner_test.go:187: 
    		
    		=== OUTPUT
    		42384]
    		I0522 10:14:32.732918    1247 wrap.go:42] PUT /api/v1/namespaces/kube-public/serviceaccounts/deployer: (1.989209ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:42384]
    		I0522 10:14:32.733115    1247 create_dockercfg_secrets.go:478] Creating dockercfg secret "builder-dockercfg-q5lfg" for service account openshift-infra/builder
    		I0522 10:14:32.762932    1247 request.go:485] Throttling request took 62.718384ms, request: POST:https://127.0.0.1:15131/api/v1/namespaces/default/secrets
    		I0522 10:14:32.764596    1247 wrap.go:42] POST /api/v1/namespaces/default/secrets: (1.451795ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:42384]
    		I0522 10:14:32.784576    1247 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (883.889µs) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.793964    1247 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (907.225µs) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.795310    1247 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (948.976µs) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.795750    1247 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.058944ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.797567    1247 wrap.go:42] PUT /api/v1/namespaces/default/serviceaccounts/deployer: (2.1635ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:42384]
    		I0522 10:14:32.797815    1247 create_dockercfg_secrets.go:478] Creating dockercfg secret "deployer-dockercfg-mt4q7" for service account openshift/deployer
    		I0522 10:14:32.797823    1247 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (960.574µs) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.800267    1247 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.122514ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.800427    1247 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.31692ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.800605    1247 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.399477ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.800908    1247 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.065068ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.801229    1247 wrap.go:42] GET /api/v1/namespaces/openshift-web-console/configmaps/webconsole-config: (1.052092ms) 404 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.809435    1247 wrap.go:42] GET /api/v1/namespaces/integration/serviceaccounts/default?timeout=1m0s: (3.234801ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42388]
    		I0522 10:14:32.810956    1247 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-token-rfh9f?timeout=1m0s: (1.08422ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42388]
    		I0522 10:14:32.812366    1247 wrap.go:42] GET /api/v1/namespaces/integration/secrets/default-dockercfg-hzvh6?timeout=1m0s: (922.244µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42388]
    		I0522 10:14:32.814577    1247 wrap.go:42] GET /api/v1/namespaces/integration/limitranges: (969.827µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.814899    1247 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I0522 10:14:32.814955    1247 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I0522 10:14:32.816550    1247 wrap.go:42] GET /api/v1/namespaces/integration: (897.282µs) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.816775    1247 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:14:32.816813    1247 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 10:14:32.816825    1247 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:14:32.816831    1247 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:14:32.816867    1247 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:14:32.816886    1247 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:14:32.816899    1247 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 10:14:32.816909    1247 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:14:32.816914    1247 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:14:32.816921    1247 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:14:32.816938    1247 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 10:14:32.816947    1247 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:14:32.816952    1247 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:14:32.816965    1247 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 10:14:32.817034    1247 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I0522 10:14:32.818670    1247 wrap.go:42] POST /api/v1/namespaces/integration/pods?timeout=1m0s: (5.764142ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42388]
    		I0522 10:14:32.818832    1247 factory.go:1147] About to try and schedule pod testqr8x2
    		I0522 10:14:32.818843    1247 scheduler.go:439] Attempting to schedule pod: integration/testqr8x2
    		I0522 10:14:32.818857    1247 scheduler.go:191] Failed to schedule pod: integration/testqr8x2
    		I0522 10:14:32.818903    1247 factory.go:1262] Unable to schedule integration testqr8x2: no nodes are registered to the cluster; waiting
    		I0522 10:14:32.818933    1247 factory.go:1375] Updating pod condition for integration/testqr8x2 to (PodScheduled==False)
    		I0522 10:14:32.819067    1247 pvc_protection_controller.go:276] Got event on pod integration/testqr8x2
    		I0522 10:14:32.819101    1247 backoff_utils.go:79] Backing off 1s
    		I0522 10:14:32.819111    1247 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"testqr8x2"}
    		I0522 10:14:32.819146    1247 disruption.go:328] addPod called on pod "testqr8x2"
    		I0522 10:14:32.819157    1247 disruption.go:403] No PodDisruptionBudgets found for pod testqr8x2, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:14:32.819164    1247 disruption.go:331] No matching pdb for pod "testqr8x2"
    		I0522 10:14:32.821940    1247 pvc_protection_controller.go:276] Got event on pod integration/testqr8x2
    		I0522 10:14:32.821995    1247 wrap.go:42] PUT /api/v1/namespaces/integration/pods/testqr8x2/status: (2.591549ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:42026]
    		I0522 10:14:32.822015    1247 disruption.go:340] updatePod called on pod "testqr8x2"
    		I0522 10:14:32.822028    1247 disruption.go:403] No PodDisruptionBudgets found for pod testqr8x2, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:14:32.822043    1247 disruption.go:343] No matching pdb for pod "testqr8x2"
    		I0522 10:14:32.822367    1247 factory.go:1147] About to try and schedule pod testqr8x2
    		I0522 10:14:32.822380    1247 scheduler.go:439] Attempting to schedule pod: integration/testqr8x2
    		I0522 10:14:32.822396    1247 scheduler.go:191] Failed to schedule pod: integration/testqr8x2
    		I0522 10:14:32.822419    1247 factory.go:1262] Unable to schedule integration testqr8x2: no nodes are registered to the cluster; waiting
    		I0522 10:14:32.822452    1247 factory.go:1375] Updating pod condition for integration/testqr8x2 to (PodScheduled==False)
    		W0522 10:14:32.822502    1247 factory.go:1304] Request for pod integration/testqr8x2 already in flight, abandoning
    		I0522 10:14:32.822864    1247 store.go:370] GuaranteedUpdate of /kubernetes.io/pods/integration/testqr8x2 failed because of a conflict, going to retry
    		I0522 10:14:32.822954    1247 wrap.go:42] POST /api/v1/namespaces/integration/events: (3.509562ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:42026]
    		I0522 10:14:32.825446    1247 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testqr8x2, uid ebe266ca-5da8-11e8-bf17-0242ac110002, event type update
    		I0522 10:14:32.825660    1247 factory.go:1147] About to try and schedule pod testqr8x2
    		I0522 10:14:32.825687    1247 scheduler.go:435] Skip schedule deleting pod: integration/testqr8x2
    		I0522 10:14:32.825808    1247 pvc_protection_controller.go:276] Got event on pod integration/testqr8x2
    		I0522 10:14:32.825828    1247 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testqr8x2, uid ebe266ca-5da8-11e8-bf17-0242ac110002, event type update
    		I0522 10:14:32.825837    1247 disruption.go:340] updatePod called on pod "testqr8x2"
    		I0522 10:14:32.825849    1247 disruption.go:403] No PodDisruptionBudgets found for pod testqr8x2, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:14:32.825856    1247 disruption.go:343] No matching pdb for pod "testqr8x2"
    		I0522 10:14:32.827308    1247 wrap.go:42] DELETE /api/v1/namespaces/integration/pods/testqr8x2?timeout=1m0s: (7.774301ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42388]
    		I0522 10:14:32.827377    1247 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testqr8x2, uid ebe266ca-5da8-11e8-bf17-0242ac110002, event type delete
    		I0522 10:14:32.827388    1247 pvc_protection_controller.go:276] Got event on pod integration/testqr8x2
    		I0522 10:14:32.827413    1247 taint_manager.go:338] Noticed pod deletion: types.NamespacedName{Namespace:"integration", Name:"testqr8x2"}
    		I0522 10:14:32.827438    1247 disruption.go:369] deletePod called on pod "testqr8x2"
    		I0522 10:14:32.827455    1247 disruption.go:403] No PodDisruptionBudgets found for pod testqr8x2, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:14:32.827463    1247 disruption.go:372] No matching pdb for pod "testqr8x2"
    		I0522 10:14:32.827498    1247 deployment_controller.go:357] Pod testqr8x2 deleted.
    		I0522 10:14:32.827503    1247 resource_quota_monitor.go:352] QuotaMonitor process object: /v1, Resource=pods, namespace integration, name testqr8x2, uid ebe266ca-5da8-11e8-bf17-0242ac110002, event type delete
    		I0522 10:14:32.827638    1247 request.go:485] Throttling request took 62.764891ms, request: POST:https://127.0.0.1:15131/api/v1/namespaces/kube-system/secrets
    		I0522 10:14:32.828352    1247 admission.go:97] getting security context constraints for pod  (generate: test) in namespace integration with user info &{system:admin  [system:masters system:cluster-admins system:authenticated] map[]}
    		I0522 10:14:32.828403    1247 admission.go:108] getting security context constraints for pod  (generate: test) with service account info &{system:serviceaccount:integration:default  [system:serviceaccounts system:serviceaccounts:integration] map[]}
    		I0522 10:14:32.830169    1247 wrap.go:42] PATCH /api/v1/namespaces/integration/events/testqr8x2.1530efec860d1a02: (6.261812ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:42026]
    		I0522 10:14:32.830199    1247 wrap.go:42] POST /api/v1/namespaces/kube-system/secrets: (2.309358ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/system:serviceaccount:openshift-infra:serviceaccount-pull-secrets-controller] 127.0.0.1:42384]
    		I0522 10:14:32.830421    1247 wrap.go:42] GET /api/v1/namespaces/integration: (1.245444ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42026]
    		I0522 10:14:32.830733    1247 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:14:32.830752    1247 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 10:14:32.830758    1247 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:14:32.830763    1247 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:14:32.830773    1247 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:14:32.830780    1247 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:14:32.830789    1247 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 10:14:32.830793    1247 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:14:32.830798    1247 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:14:32.830805    1247 matcher.go:342] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:14:32.830823    1247 matcher.go:279] got preallocated values for min: 1000060000, max: 1000069999 for uid range in namespace integration
    		I0522 10:14:32.830832    1247 matcher.go:292] got preallocated value for level: s0:c8,c2 for selinux options in namespace integration
    		I0522 10:14:32.830839    1247 matcher.go:322] got preallocated value for groups: 1000060000/10000 in namespace integration
    		I0522 10:14:32.830858    1247 admission.go:217] validating pod  (generate: test) against providers anyuid,restricted,nonroot,hostmount-anyuid,hostnetwork,hostaccess,privileged
    		I0522 10:14:32.830900    1247 admission.go:170] pod  (generate: test) validated against provider anyuid
    		I0522 10:14:32.832995    1247 wrap.go:42] POST /api/v1/namespaces/integration/events: (2.369012ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:42026]
    		I0522 10:14:32.832999    1247 wrap.go:42] POST /api/v1/namespaces/integration/pods?timeout=1m0s: (5.169315ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8] 127.0.0.1:42388]
    		I0522 10:14:32.833634    1247 disruption.go:328] addPod called on pod "testl2xmx"
    		I0522 10:14:32.833651    1247 factory.go:1147] About to try and schedule pod testl2xmx
    		I0522 10:14:32.833660    1247 disruption.go:403] No PodDisruptionBudgets found for pod testl2xmx, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:14:32.833662    1247 scheduler.go:439] Attempting to schedule pod: integration/testl2xmx
    		I0522 10:14:32.833668    1247 disruption.go:331] No matching pdb for pod "testl2xmx"
    		I0522 10:14:32.833678    1247 scheduler.go:191] Failed to schedule pod: integration/testl2xmx
    		I0522 10:14:32.833696    1247 factory.go:1262] Unable to schedule integration testl2xmx: no nodes are registered to the cluster; waiting
    		I0522 10:14:32.833725    1247 taint_manager.go:345] Noticed pod update: types.NamespacedName{Namespace:"integration", Name:"testl2xmx"}
    		I0522 10:14:32.833735    1247 factory.go:1375] Updating pod condition for integration/testl2xmx to (PodScheduled==False)
    		I0522 10:14:32.833701    1247 pvc_protection_controller.go:276] Got event on pod integration/testl2xmx
    		I0522 10:14:32.833914    1247 backoff_utils.go:79] Backing off 1s
    		INFO: 2018/05/22 10:14:32 ccBalancerWrapper: updating state and picker called by balancer: IDLE, 0xc432da7200
    		INFO: 2018/05/22 10:14:32 dialing to target with scheme: ""
    		INFO: 2018/05/22 10:14:32 could not get resolver for scheme: ""
    		INFO: 2018/05/22 10:14:32 balancerWrapper: is pickfirst: false
    		INFO: 2018/05/22 10:14:32 balancerWrapper: got update addr from Notify: [{127.0.0.1:23291 <nil>}]
    		INFO: 2018/05/22 10:14:32 ccBalancerWrapper: new subconn: [{127.0.0.1:23291 0  <nil>}]
    		INFO: 2018/05/22 10:14:32 balancerWrapper: handle subconn state change: 0xc42c858c90, CONNECTING
    		INFO: 2018/05/22 10:14:32 ccBalancerWrapper: updating state and picker called by balancer: CONNECTING, 0xc432da7200
    		I0522 10:14:32.836387    1247 wrap.go:42] PUT /api/v1/namespaces/integration/pods/testl2xmx/status: (2.224214ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/scheduler] 127.0.0.1:42026]
    		I0522 10:14:32.836793    1247 pvc_protection_controller.go:276] Got event on pod integration/testl2xmx
    		I0522 10:14:32.836818    1247 disruption.go:340] updatePod called on pod "testl2xmx"
    		I0522 10:14:32.836825    1247 disruption.go:403] No PodDisruptionBudgets found for pod testl2xmx, PodDisruptionBudget controller will avoid syncing.
    		I0522 10:14:32.836830    1247 disruption.go:343] No matching pdb for pod "testl2xmx"
    		I0522 10:14:32.837071    1247 factory.go:1147] About to try and schedule pod testl2xmx
    		I0522 10:14:32.837083    1247 scheduler.go:439] Attempting to schedule pod: integration/testl2xmx
    		I0522 10:14:32.837098    1247 scheduler.go:191] Failed to schedule pod: integration/testl2xmx
    		I0522 10:14:32.837118    1247 factory.go:1262] Unable to schedule integration testl2xmx: no nodes are registered to the cluster; waiting
    		I0522 10:14:32.837144    1247 factory.go:1375] Updating pod condition for integration/testl2xmx to (PodScheduled==False)
    		W0522 10:14:32.837174    1247 factory.go:1304] Request for pod integration/testl2xmx already in flight, abandoning
    		I0522 10:14:32.837574    1247 wrap.go:42] POST /api/v1/namespaces/integration/events: (3.606604ms) 201 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:42026]
    		I0522 10:14:32.841563    1247 wrap.go:42] PATCH /api/v1/namespaces/integration/events/testl2xmx.1530efec86eef79c: (3.436517ms) 200 [[integration.test/v1.10.0+b81c8f8 (linux/amd64) kubernetes/b81c8f8/leader-election] 127.0.0.1:42026]
    		INFO: 2018/05/22 10:14:32 balancerWrapper: handle subconn state change: 0xc42c858c90, READY
    		INFO: 2018/05/22 10:14:32 ccBalancerWrapper: updating state and picker called by balancer: READY, 0xc432da7200
    		INFO: 2018/05/22 10:14:32 balancerWrapper: got update addr from Notify: [{127.0.0.1:23291 <nil>}]
    		
FAIL
exit status 1
FAIL	github.com/openshift/origin/test/integration/runner	1952.003s
[INFO] [10:14:33+0000] jUnit XML report placed at _output/scripts/test-integration/artifacts/gotest_report_dsSDQ.xml
Of 203 tests executed in 1952.003s, 199 succeeded, 4 failed, and 0 were skipped.

In suite "github.com/openshift/origin/test/integration/runner", test case "TestIntegration" failed:
runner_test.go:84: using existing binary

In suite "github.com/openshift/origin/test/integration/runner", test case "TestIntegration/TestConcurrentBuildImageChangeTriggerControllers" failed:
runner_test.go:129: FAILED TestConcurrentBuildImageChangeTriggerControllers, retrying:

controllers.go:531: expected watch event type ADDED, got 


In suite "github.com/openshift/origin/test/integration/runner", test case "TestIntegration/TestImageStreamImportScheduled" failed:
runner_test.go:129: FAILED TestImageStreamImportScheduled, retrying:

imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 1
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 2
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 3
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 4
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 5
imageimporter_test.go:623: got GET /v2/
imageimporter_test.go:623: got GET /v2/test/image/manifests/latest
imageimporter_test.go:630: serving 6
imageimporter_test.go:771: unexpected object: <nil>


In suite "github.com/openshift/origin/test/integration/runner", test case "TestIntegration/TestOAuthServiceAccountClientEvent" failed:
runner_test.go:129: FAILED TestOAuthServiceAccountClientEvent, retrying:

oauth_serviceaccount_client_events_test.go:141: test-bad-url-parse: annotationPrefix serviceaccounts.openshift.io/oauth-redirecturi.one, annotation ::
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:13340/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=%!A(MISSING)%!A(MISSING)&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:396: Bad Request: HTTP/2.0 400 Bad Request
Content-Length: 155
Cache-Control: no-cache, no-store, max-age=0, must-revalidate
Content-Type: application/json
Date: Tue, 22 May 2018 09:54:08 GMT
Expires: Fri, 01 Jan 1990 00:00:00 GMT
Pragma: no-cache

{"error":"server_error","error_description":"The authorization server encountered an unexpected condition that prevented it from fulfilling the request."}
oauth_serviceaccount_client_events_test.go:141: test-bad-redirect-route-not-found: annotationPrefix serviceaccounts.openshift.io/oauth-redirectreference.1, annotation {"kind":"OAuthRedirectReference","apiVersion":"oauth.openshift.io/v1","metadata":{"creationTimestamp":null},"reference":{"group":"route.openshift.io","kind":"Route","name":"route1"}}
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:13340/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=%!B(MISSING)%!k(MISSING)ind%3A%!O(MISSING)AuthRedirectReference%2C%!a(MISSING)piVersion%3A%!o(MISSING)auth.openshift.io%!F(MISSING)v1%2C%!m(MISSING)etadata%3A%!B(MISSING)%!c(MISSING)reationTimestamp%3Anull%!D(MISSING)%!C(MISSING)%!r(MISSING)eference%3A%!B(MISSING)%!g(MISSING)roup%3A%!r(MISSING)oute.openshift.io%2C%!k(MISSING)ind%3A%!R(MISSING)oute%2C%!n(MISSING)ame%3A%!r(MISSING)oute1%7D%!D(MISSING)%!A(MISSING)&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:396: Bad Request: HTTP/2.0 400 Bad Request
Content-Length: 155
Cache-Control: no-cache, no-store, max-age=0, must-revalidate
Content-Type: application/json
Date: Tue, 22 May 2018 09:54:08 GMT
Expires: Fri, 01 Jan 1990 00:00:00 GMT
Pragma: no-cache

{"error":"server_error","error_description":"The authorization server encountered an unexpected condition that prevented it from fulfilling the request."}
oauth_serviceaccount_client_events_test.go:141: test-bad-redirect-route-wrong-group: annotationPrefix serviceaccounts.openshift.io/oauth-redirectreference.1, annotation {"kind":"OAuthRedirectReference","apiVersion":"oauth.openshift.io/v1","metadata":{"creationTimestamp":null},"reference":{"group":"foo","kind":"Route","name":"route1"}}
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:13340/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=%!B(MISSING)%!k(MISSING)ind%3A%!O(MISSING)AuthRedirectReference%2C%!a(MISSING)piVersion%3A%!o(MISSING)auth.openshift.io%!F(MISSING)v1%2C%!m(MISSING)etadata%3A%!B(MISSING)%!c(MISSING)reationTimestamp%3Anull%!D(MISSING)%!C(MISSING)%!r(MISSING)eference%3A%!B(MISSING)%!g(MISSING)roup%3A%!f(MISSING)oo%2C%!k(MISSING)ind%3A%!R(MISSING)oute%2C%!n(MISSING)ame%3A%!r(MISSING)oute1%7D%!D(MISSING)%!A(MISSING)&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:396: Bad Request: HTTP/2.0 400 Bad Request
Content-Length: 155
Cache-Control: no-cache, no-store, max-age=0, must-revalidate
Content-Type: application/json
Date: Tue, 22 May 2018 09:54:08 GMT
Expires: Fri, 01 Jan 1990 00:00:00 GMT
Pragma: no-cache

{"error":"server_error","error_description":"The authorization server encountered an unexpected condition that prevented it from fulfilling the request."}
oauth_serviceaccount_client_events_test.go:141: test-bad-redirect-reference-kind: annotationPrefix serviceaccounts.openshift.io/oauth-redirectreference.1, annotation {"kind":"OAuthRedirectReference","apiVersion":"oauth.openshift.io/v1","metadata":{"creationTimestamp":null},"reference":{"group":"route.openshift.io","kind":"foo","name":"route1"}}
oauth_serviceaccount_client_events_test.go:371: GET https://127.0.0.1:13340/oauth/authorize?client_id=system%!A(MISSING)serviceaccount%!A(MISSING)test-project%!A(MISSING)default&redirect_uri=%!B(MISSING)%!k(MISSING)ind%3A%!O(MISSING)AuthRedirectReference%2C%!a(MISSING)piVersion%3A%!o(MISSING)auth.openshift.io%!F(MISSING)v1%2C%!m(MISSING)etadata%3A%!B(MISSING)%!c(MISSING)reationTimestamp%3Anull%!D(MISSING)%!C(MISSING)%!r(MISSING)eference%3A%!B(MISSING)%!g(MISSING)roup%3A%!r(MISSING)oute.openshift.io%2C%!k(MISSING)ind%3A%!f(MISSING)oo%2C%!n(MISSING)ame%3A%!r(MISSING)oute1%7D%!D(MISSING)%!A(MISSING)&response_type=code&scope=user%!A(MISSING)info+role%!A(MISSING)edit%!A(MISSING)test-project
oauth_serviceaccount_client_events_test.go:396: Bad Request: HTTP/2.0 400 Bad Request
Content-Length: 155
Cache-Control: no-cache, no-store, max-age=0, must-revalidate
Content-Type: application/json
Date: Tue, 22 May 2018 09:54:08 GMT
Expires: Fri, 01 Jan 1990 00:00:00 GMT
Pragma: no-cache

{"error":"server_error","error_description":"The authorization server encountered an unexpected condition that prevented it from fulfilling the request."}
oauth_serviceaccount_client_events_test.go:163: test-bad-redirect-reference-kind: expected 1 events, found 0
[ERROR] [10:14:34+0000] hack/test-go.sh exited with code 1 after 00h 32m 34s
[ERROR] [10:14:34+0000] PID 1244: hack/test-integration.sh:18: `COVERAGE_SPEC=" " DETECT_RACES=false TMPDIR="${BASETMPDIR}" TIMEOUT=45m GOTEST_FLAGS="${gotest_flags}" "${OS_ROOT}/hack/test-go.sh" "test/integration/runner"` exited with status 1.
[INFO] [10:14:34+0000] 		Stack Trace: 
[INFO] [10:14:34+0000] 		  1: hack/test-integration.sh:18: `COVERAGE_SPEC=" " DETECT_RACES=false TMPDIR="${BASETMPDIR}" TIMEOUT=45m GOTEST_FLAGS="${gotest_flags}" "${OS_ROOT}/hack/test-go.sh" "test/integration/runner"`
[INFO] [10:14:34+0000]   Exiting with code 1.
make: *** [test-integration] Error 1
[WARNING] [10:14:35+0000] Copying _output/local/releases from the container failed!
[WARNING] [10:14:35+0000] Error response from daemon: lstat /var/lib/docker/overlay2/e89f36655e5b0612a22dce940ae90ec58b1518aa696822e141e3ff9ce2c56eba/merged/go/src/github.com/openshift/origin/_output/local/releases: no such file or directory
[ERROR] [10:14:35+0000] PID 10821: hack/lib/build/environment.sh:172: `return "${exitcode}"` exited with status 2.
[INFO] [10:14:35+0000] 		Stack Trace: 
[INFO] [10:14:35+0000] 		  1: hack/lib/build/environment.sh:172: `return "${exitcode}"`
[INFO] [10:14:35+0000] 		  2: hack/lib/build/environment.sh:281: os::build::environment::withsource
[INFO] [10:14:35+0000] 		  3: hack/env:42: os::build::environment::run
[INFO] [10:14:35+0000]   Exiting with code 2.
++ export status=FAILURE
++ status=FAILURE
+ set +o xtrace
########## FINISHED STAGE: FAILURE: RUN INTEGRATION TESTS [00h 44m 35s] ##########
Build step 'Execute shell' marked build as failure
[PostBuildScript] - Executing post build scripts.
[workspace@2] $ /bin/bash /tmp/jenkins5192873762772446130.sh
########## STARTING STAGE: DOWNLOAD ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/gathered
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/gathered
+ mkdir -p /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/gathered
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo stat /data/src/github.com/openshift/origin/_output/scripts
  File: ‘/data/src/github.com/openshift/origin/_output/scripts’
  Size: 61        	Blocks: 0          IO Block: 4096   directory
Device: ca02h/51714d	Inode: 88263220    Links: 5
Access: (2755/drwxr-sr-x)  Uid: ( 1001/  origin)   Gid: ( 1003/origin-git)
Context: unconfined_u:object_r:container_file_t:s0
Access: 1970-01-01 00:00:00.000000000 +0000
Modify: 2018-05-22 09:31:41.000000000 +0000
Change: 2018-05-22 10:14:35.535072053 +0000
 Birth: -
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod -R o+rX /data/src/github.com/openshift/origin/_output/scripts
+ scp -r -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel:/data/src/github.com/openshift/origin/_output/scripts /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/gathered
+ tree /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/gathered
/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/gathered
└── scripts
    ├── shell
    │   ├── artifacts
    │   ├── logs
    │   │   ├── e60ed31da18e4c5bf0655908b0120756453b3b9d5a04860f59d1f41c6d22693b.json
    │   │   └── scripts.log
    │   └── openshift.local.home
    ├── test-integration
    │   ├── artifacts
    │   │   ├── gotest_report_dsSDQ
    │   │   └── gotest_report_dsSDQ.xml
    │   ├── logs
    │   │   ├── raw_test_output.log
    │   │   ├── scripts.log
    │   │   └── test-go-err.log
    │   └── openshift.local.home
    └── test-tools
        ├── artifacts
        ├── logs
        │   ├── raw_test_output.log
        │   └── scripts.log
        └── openshift.local.home

13 directories, 9 files
+ exit 0
[workspace@2] $ /bin/bash /tmp/jenkins7071010773567773410.sh
########## STARTING STAGE: GENERATE ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/generated
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/generated
+ mkdir /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/generated
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a 2>&1'
  WARNING: You're not using the default seccomp profile
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo cat /etc/sysconfig/docker /etc/sysconfig/docker-network /etc/sysconfig/docker-storage /etc/sysconfig/docker-storage-setup /etc/systemd/system/docker.service 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'oc get --raw /metrics --server=https://$( uname --nodename ):10250 --config=/etc/origin/master/admin.kubeconfig 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'oc get --raw /metrics --config=/etc/origin/master/admin.kubeconfig 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo df -T -h && sudo pvs && sudo vgs && sudo lvs && sudo findmnt --all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo yum list installed 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl --dmesg --no-pager --all --lines=all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl _PID=1 --no-pager --all --lines=all 2>&1'
+ tree /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/generated
/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/generated
├── avc_denials.log
├── containers.log
├── dmesg.log
├── docker.config
├── docker.info
├── filesystem.info
├── installed_packages.log
├── master-metrics.log
├── node-metrics.log
└── pid1.journal

0 directories, 10 files
+ exit 0
[workspace@2] $ /bin/bash /tmp/jenkins9050883427586586922.sh
########## STARTING STAGE: FETCH SYSTEMD JOURNALS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/journals
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/journals
+ mkdir /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/journals
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit docker.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
+ tree /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/journals
/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/artifacts/journals
├── dnsmasq.service
├── docker.service
└── systemd-journald.service

0 directories, 3 files
+ exit 0
[workspace@2] $ /bin/bash /tmp/jenkins575262929990040653.sh
########## STARTING STAGE: ASSEMBLE GCS OUTPUT ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
+ trap 'exit 0' EXIT
+ mkdir -p gcs/artifacts gcs/artifacts/generated gcs/artifacts/journals gcs/artifacts/gathered
++ python -c 'import json; import urllib; print json.load(urllib.urlopen('\''https://ci.openshift.redhat.com/jenkins/job/test_pull_request_origin_integration/18011/api/json'\''))['\''result'\'']'
+ result=FAILURE
+ cat
++ date +%s
+ cat /var/lib/jenkins/jobs/test_pull_request_origin_integration/builds/18011/log
+ cp artifacts/generated/avc_denials.log artifacts/generated/containers.log artifacts/generated/dmesg.log artifacts/generated/docker.config artifacts/generated/docker.info artifacts/generated/filesystem.info artifacts/generated/installed_packages.log artifacts/generated/master-metrics.log artifacts/generated/node-metrics.log artifacts/generated/pid1.journal gcs/artifacts/generated/
+ cp artifacts/journals/dnsmasq.service artifacts/journals/docker.service artifacts/journals/systemd-journald.service gcs/artifacts/journals/
+ cp -r artifacts/gathered/scripts gcs/artifacts/
++ pwd
+ scp -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config -r /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/gcs openshiftdevel:/data
+ scp -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.config/gcloud/gcs-publisher-credentials.json openshiftdevel:/data/credentials.json
+ exit 0
[workspace@2] $ /bin/bash /tmp/jenkins4354167589270373733.sh
########## STARTING STAGE: PUSH THE ARTIFACTS AND METADATA ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ mktemp
+ script=/tmp/tmp.SKN3S1VfWJ
+ cat
+ chmod +x /tmp/tmp.SKN3S1VfWJ
+ scp -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.SKN3S1VfWJ openshiftdevel:/tmp/tmp.SKN3S1VfWJ
+ ssh -F /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 300 /tmp/tmp.SKN3S1VfWJ"'
+ cd /home/origin
+ trap 'exit 0' EXIT
+ [[ -n {"type":"presubmit","job":"test_pull_request_origin_integration","buildid":"14cd55ae-5da2-11e8-ae4a-0a58ac100556","refs":{"org":"openshift","repo":"origin","base_ref":"master","base_sha":"b89edde21a383ff0ab88317b8d74aedf892239cf","pulls":[{"number":19759,"author":"wozniakjan","sha":"cd106aa6bae02a5e0af98e558989117b28e98957"}]}} ]]
++ jq --compact-output .buildid
+ [[ "14cd55ae-5da2-11e8-ae4a-0a58ac100556" =~ ^"[0-9]+"$ ]]
+ echo 'Using BUILD_NUMBER'
Using BUILD_NUMBER
++ jq --compact-output '.buildid |= "18011"'
+ JOB_SPEC='{"type":"presubmit","job":"test_pull_request_origin_integration","buildid":"18011","refs":{"org":"openshift","repo":"origin","base_ref":"master","base_sha":"b89edde21a383ff0ab88317b8d74aedf892239cf","pulls":[{"number":19759,"author":"wozniakjan","sha":"cd106aa6bae02a5e0af98e558989117b28e98957"}]}}'
+ docker run -e 'JOB_SPEC={"type":"presubmit","job":"test_pull_request_origin_integration","buildid":"18011","refs":{"org":"openshift","repo":"origin","base_ref":"master","base_sha":"b89edde21a383ff0ab88317b8d74aedf892239cf","pulls":[{"number":19759,"author":"wozniakjan","sha":"cd106aa6bae02a5e0af98e558989117b28e98957"}]}}' -v /data:/data:z registry.svc.ci.openshift.org/ci/gcsupload:latest --dry-run=false --gcs-path=gs://origin-ci-test --gcs-credentials-file=/data/credentials.json --path-strategy=single --default-org=openshift --default-repo=origin /data/gcs/artifacts /data/gcs/build-log.txt /data/gcs/finished.json
Unable to find image 'registry.svc.ci.openshift.org/ci/gcsupload:latest' locally
Trying to pull repository registry.svc.ci.openshift.org/ci/gcsupload ... 
latest: Pulling from registry.svc.ci.openshift.org/ci/gcsupload
6d987f6f4279: Already exists
4cccebe844ee: Already exists
cafcba51f636: Pulling fs layer
cafcba51f636: Verifying Checksum
cafcba51f636: Download complete
cafcba51f636: Pull complete
Digest: sha256:c452798b56e3f4649c557c3ff7273126042065e3be152689a1ffb880638e617d
Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/gcsupload:latest
{"component":"gcsupload","level":"info","msg":"Gathering artifacts from artifact directory: /data/gcs/artifacts","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/avc_denials.log in artifact directory. Uploading as artifacts/generated/avc_denials.log\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/containers.log in artifact directory. Uploading as artifacts/generated/containers.log\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/dmesg.log in artifact directory. Uploading as artifacts/generated/dmesg.log\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.config in artifact directory. Uploading as artifacts/generated/docker.config\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.info in artifact directory. Uploading as artifacts/generated/docker.info\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/filesystem.info in artifact directory. Uploading as artifacts/generated/filesystem.info\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/installed_packages.log in artifact directory. Uploading as artifacts/generated/installed_packages.log\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/master-metrics.log in artifact directory. Uploading as artifacts/generated/master-metrics.log\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/node-metrics.log in artifact directory. Uploading as artifacts/generated/node-metrics.log\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/pid1.journal in artifact directory. Uploading as artifacts/generated/pid1.journal\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/dnsmasq.service in artifact directory. Uploading as artifacts/journals/dnsmasq.service\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/docker.service in artifact directory. Uploading as artifacts/journals/docker.service\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/systemd-journald.service in artifact directory. Uploading as artifacts/journals/systemd-journald.service\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/shell/logs/e60ed31da18e4c5bf0655908b0120756453b3b9d5a04860f59d1f41c6d22693b.json in artifact directory. Uploading as artifacts/scripts/shell/logs/e60ed31da18e4c5bf0655908b0120756453b3b9d5a04860f59d1f41c6d22693b.json\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/shell/logs/scripts.log in artifact directory. Uploading as artifacts/scripts/shell/logs/scripts.log\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/artifacts/gotest_report_dsSDQ in artifact directory. Uploading as artifacts/scripts/test-integration/artifacts/gotest_report_dsSDQ\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/artifacts/gotest_report_dsSDQ.xml in artifact directory. Uploading as artifacts/scripts/test-integration/artifacts/gotest_report_dsSDQ.xml\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/logs/raw_test_output.log in artifact directory. Uploading as artifacts/scripts/test-integration/logs/raw_test_output.log\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/logs/scripts.log in artifact directory. Uploading as artifacts/scripts/test-integration/logs/scripts.log\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-integration/logs/test-go-err.log in artifact directory. Uploading as artifacts/scripts/test-integration/logs/test-go-err.log\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-tools/logs/raw_test_output.log in artifact directory. Uploading as artifacts/scripts/test-tools/logs/raw_test_output.log\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/test-tools/logs/scripts.log in artifact directory. Uploading as artifacts/scripts/test-tools/logs/scripts.log\n","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-tools/logs/raw_test_output.log","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/docker.info","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/filesystem.info","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/journals/systemd-journald.service","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/shell/logs/scripts.log","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-integration/logs/test-go-err.log","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/master-metrics.log","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-integration/artifacts/gotest_report_dsSDQ","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/finished.json","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration/latest-build.txt","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/containers.log","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/pid1.journal","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/dmesg.log","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/installed_packages.log","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/shell/logs/e60ed31da18e4c5bf0655908b0120756453b3b9d5a04860f59d1f41c6d22693b.json","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-integration/logs/scripts.log","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration/18011.txt","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/avc_denials.log","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/latest-build.txt","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/docker.config","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/journals/dnsmasq.service","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-integration/artifacts/gotest_report_dsSDQ.xml","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-integration/logs/raw_test_output.log","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-tools/logs/scripts.log","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/node-metrics.log","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/journals/docker.service","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/build-log.txt","level":"info","msg":"Queued for upload","time":"2018-05-22T10:15:00Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/node-metrics.log","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/journals/systemd-journald.service","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-integration/artifacts/gotest_report_dsSDQ","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/avc_denials.log","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-tools/logs/raw_test_output.log","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/shell/logs/scripts.log","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/containers.log","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-integration/logs/test-go-err.log","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/latest-build.txt","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/docker.info","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/journals/dnsmasq.service","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration/18011.txt","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/installed_packages.log","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/filesystem.info","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/master-metrics.log","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/shell/logs/e60ed31da18e4c5bf0655908b0120756453b3b9d5a04860f59d1f41c6d22693b.json","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/pid1.journal","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/finished.json","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/docker.config","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-tools/logs/scripts.log","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/generated/dmesg.log","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/directory/test_pull_request_origin_integration/latest-build.txt","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-integration/logs/scripts.log","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/journals/docker.service","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/build-log.txt","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-integration/logs/raw_test_output.log","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","dest":"pr-logs/pull/19759/test_pull_request_origin_integration/18011/artifacts/scripts/test-integration/artifacts/gotest_report_dsSDQ.xml","level":"info","msg":"Finished upload","time":"2018-05-22T10:15:01Z"}
{"component":"gcsupload","level":"info","msg":"Finished upload to GCS","time":"2018-05-22T10:15:01Z"}
+ exit 0
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: PUSH THE ARTIFACTS AND METADATA [00h 00m 05s] ##########
[workspace@2] $ /bin/bash /tmp/jenkins3089150237580885861.sh
########## STARTING STAGE: DEPROVISION CLOUD RESOURCES ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83
++ export PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config
+ oct deprovision

PLAYBOOK: main.yml *************************************************************
4 plays in /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml

PLAY [ensure we have the parameters necessary to deprovision virtual hosts] ****

TASK [ensure all required variables are set] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:9
skipping: [localhost] => (item=origin_ci_inventory_dir)  => {
    "changed": false, 
    "generated_timestamp": "2018-05-22 06:15:02.967460", 
    "item": "origin_ci_inventory_dir", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_region)  => {
    "changed": false, 
    "generated_timestamp": "2018-05-22 06:15:02.970809", 
    "item": "origin_ci_aws_region", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

PLAY [deprovision virtual hosts in EC2] ****************************************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

TASK [deprovision a virtual EC2 host] ******************************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:28
included: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml for localhost

TASK [update the SSH configuration to remove AWS EC2 specifics] ****************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:2
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-05-22 06:15:03.749807", 
    "msg": ""
}

TASK [rename EC2 instance for termination reaper] ******************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:8
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-05-22 06:15:04.332650", 
    "msg": "Tags {'Name': 'oct-terminate'} created for resource i-0db401b9bde50f9b8."
}

TASK [tear down the EC2 instance] **********************************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:15
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-05-22 06:15:05.184298", 
    "instance_ids": [
        "i-0db401b9bde50f9b8"
    ], 
    "instances": [
        {
            "ami_launch_index": "0", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-0e0383d6365f34ff1"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-06583daab31aa3cd6"
                }
            }, 
            "dns_name": "ec2-35-153-183-186.compute-1.amazonaws.com", 
            "ebs_optimized": false, 
            "groups": {
                "sg-7e73221a": "default"
            }, 
            "hypervisor": "xen", 
            "id": "i-0db401b9bde50f9b8", 
            "image_id": "ami-0f07d2d9a03af96ec", 
            "instance_type": "m4.xlarge", 
            "kernel": null, 
            "key_name": "libra", 
            "launch_time": "2018-05-22T09:26:22.000Z", 
            "placement": "us-east-1d", 
            "private_dns_name": "ip-172-18-6-57.ec2.internal", 
            "private_ip": "172.18.6.57", 
            "public_dns_name": "ec2-35-153-183-186.compute-1.amazonaws.com", 
            "public_ip": "35.153.183.186", 
            "ramdisk": null, 
            "region": "us-east-1", 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "running", 
            "state_code": 16, 
            "tags": {
                "Name": "oct-terminate", 
                "openshift_etcd": "", 
                "openshift_master": "", 
                "openshift_node": ""
            }, 
            "tenancy": "default", 
            "virtualization_type": "hvm"
        }
    ], 
    "tagged_instances": []
}

TASK [remove the serialized host variables] ************************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:22
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-05-22 06:15:05.475628", 
    "path": "/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory/host_vars/172.18.6.57.yml", 
    "state": "absent"
}

PLAY [deprovision virtual hosts locally manged by Vagrant] *********************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

PLAY [clean up local configuration for deprovisioned instances] ****************

TASK [remove inventory configuration directory] ********************************
task path: /var/lib/jenkins/origin-ci-tool/7a5c5e83c372ad2e6b3b64b3efa16fe2cb37ef83/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:61
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-05-22 06:15:05.949088", 
    "path": "/var/lib/jenkins/jobs/test_pull_request_origin_integration/workspace@2/.config/origin-ci-tool/inventory", 
    "state": "absent"
}

PLAY RECAP *********************************************************************
localhost                  : ok=8    changed=4    unreachable=0    failed=0   

+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPROVISION CLOUD RESOURCES [00h 00m 05s] ##########
Archiving artifacts
[BFA] Scanning build for known causes...
[BFA] Found failure cause(s):
[BFA] Unit or Integration Test Failed
[BFA] Job Stage Failed
[BFA] Done. 22s
[WS-CLEANUP] Deleting project workspace...[WS-CLEANUP] done
Finished: FAILURE