Console Output

Skipping 1,416 KB.. Full Log
Jul  9 14:39:54.131: INFO: (17) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:portname1/: foo (200; 52.600577ms)
Jul  9 14:39:54.131: INFO: (17) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:443/: tls baz (200; 52.595164ms)
Jul  9 14:39:54.131: INFO: (17) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:tlsportname2/: tls qux (200; 52.481988ms)
Jul  9 14:39:54.131: INFO: (17) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:81/: bar (200; 52.590731ms)
Jul  9 14:39:54.131: INFO: (17) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:162/: bar (200; 52.611822ms)
Jul  9 14:39:54.131: INFO: (17) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:162/: bar (200; 52.57743ms)
Jul  9 14:39:54.134: INFO: (17) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:81/: bar (200; 55.737381ms)
Jul  9 14:39:54.134: INFO: (17) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:80/: foo (200; 55.700213ms)
Jul  9 14:39:54.155: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:160/: foo (200; 20.886206ms)
Jul  9 14:39:54.155: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:1080/proxy/: <a href="/api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:1080/proxy/... (200; 21.39112ms)
Jul  9 14:39:54.156: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:162/proxy/: bar (200; 22.320767ms)
Jul  9 14:39:54.157: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:portname1/: foo (200; 22.973099ms)
Jul  9 14:39:54.157: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:portname1/proxy/: foo (200; 22.795498ms)
Jul  9 14:39:54.168: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:tlsportname1/: tls baz (200; 33.686104ms)
Jul  9 14:39:54.168: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:160/: foo (200; 33.652631ms)
Jul  9 14:39:54.168: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:tlsportname2/: tls qux (200; 33.754527ms)
Jul  9 14:39:54.177: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:tlsportname1/proxy/: tls baz (200; 42.393123ms)
Jul  9 14:39:54.177: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/https:proxy-service-wms7n-8bfbf:462/proxy/: tls qux (200; 42.613208ms)
Jul  9 14:39:54.177: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:portname1/: foo (200; 42.826402ms)
Jul  9 14:39:54.177: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:tlsportname2/proxy/: tls qux (200; 42.950354ms)
Jul  9 14:39:54.177: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:portname1/proxy/: foo (200; 42.968833ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:portname2/: bar (200; 55.563278ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:portname2/: bar (200; 55.339379ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:162/proxy/: bar (200; 55.374421ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf/proxy/: <a href="/api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf/proxy/rewriteme"... (200; 55.707872ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:1080/proxy/: <a href="/api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:1080/proxy/rewri... (200; 55.568909ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:1080/: <a href="/api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:1080/... (200; 55.657848ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:portname2/proxy/: bar (200; 55.712207ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:portname2/proxy/: bar (200; 55.588925ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:160/proxy/: foo (200; 55.659995ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:1080/: <a href="/api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:1080/rewri... (200; 55.579321ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/https:proxy-service-wms7n-8bfbf:460/proxy/: tls baz (200; 55.545387ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:160/proxy/: foo (200; 55.581425ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/https:proxy-service-wms7n-8bfbf:443/proxy/: <a href="/api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/https:proxy-service-wms7n-8bfbf:443/proxy/... (200; 55.721338ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:162/: bar (200; 55.649133ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:162/: bar (200; 55.771136ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:81/: bar (200; 56.061303ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:81/: bar (200; 56.013063ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:443/: tls baz (200; 55.830977ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:444/: tls qux (200; 56.051864ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:80/: foo (200; 56.075478ms)
Jul  9 14:39:54.190: INFO: (18) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:80/: foo (200; 55.887936ms)
Jul  9 14:39:54.219: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:162/proxy/: bar (200; 28.662412ms)
Jul  9 14:39:54.219: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:portname2/: bar (200; 28.93312ms)
Jul  9 14:39:54.219: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:portname1/proxy/: foo (200; 28.958431ms)
Jul  9 14:39:54.220: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:portname2/proxy/: bar (200; 29.193153ms)
Jul  9 14:39:54.223: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:1080/: <a href="/api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:1080/rewri... (200; 32.717959ms)
Jul  9 14:39:54.233: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:160/: foo (200; 42.366378ms)
Jul  9 14:39:54.233: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:80/: foo (200; 42.581133ms)
Jul  9 14:39:54.233: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/https:proxy-service-wms7n-8bfbf:460/proxy/: tls baz (200; 42.477814ms)
Jul  9 14:39:54.233: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:tlsportname2/proxy/: tls qux (200; 42.561243ms)
Jul  9 14:39:54.233: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:tlsportname1/proxy/: tls baz (200; 42.621949ms)
Jul  9 14:39:54.237: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:162/: bar (200; 45.969819ms)
Jul  9 14:39:54.242: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:160/proxy/: foo (200; 51.326848ms)
Jul  9 14:39:54.242: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:1080/: <a href="/api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:1080/... (200; 51.206766ms)
Jul  9 14:39:54.242: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/https:proxy-service-wms7n-8bfbf:443/proxy/: <a href="/api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/https:proxy-service-wms7n-8bfbf:443/proxy/... (200; 51.629057ms)
Jul  9 14:39:54.242: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:1080/proxy/: <a href="/api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:1080/proxy/... (200; 51.535871ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:160/proxy/: foo (200; 60.385203ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:160/: foo (200; 60.586575ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/https:proxy-service-wms7n-8bfbf:462/proxy/: tls qux (200; 60.448437ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:portname2/proxy/: bar (200; 60.503355ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:portname2/: bar (200; 60.400547ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:portname1/proxy/: foo (200; 60.589467ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:portname1/: foo (200; 60.410312ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:444/: tls qux (200; 60.620408ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:81/: bar (200; 60.734574ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:81/: bar (200; 60.433539ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:tlsportname2/: tls qux (200; 60.717659ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/proxy-service-wms7n:portname1/: foo (200; 60.777374ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:443/: tls baz (200; 60.958069ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:162/: bar (200; 60.741695ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/http:proxy-service-wms7n:80/: foo (200; 60.636749ms)
Jul  9 14:39:54.251: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf/proxy/: <a href="/api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf/proxy/rewriteme"... (200; 61.13284ms)
Jul  9 14:39:54.252: INFO: (19) /api/v1/proxy/namespaces/e2e-tests-proxy-6ksf7/services/https:proxy-service-wms7n:tlsportname1/: tls baz (200; 61.095849ms)
Jul  9 14:39:54.252: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:1080/proxy/: <a href="/api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/proxy-service-wms7n-8bfbf:1080/proxy/rewri... (200; 60.648521ms)
Jul  9 14:39:54.252: INFO: (19) /api/v1/namespaces/e2e-tests-proxy-6ksf7/pods/http:proxy-service-wms7n-8bfbf:162/proxy/: bar (200; 60.703777ms)
STEP: deleting { ReplicationController} proxy-service-wms7n in namespace e2e-tests-proxy-6ksf7
Jul  9 14:39:54.438: INFO: Deleting { ReplicationController} proxy-service-wms7n took: 117.86271ms
Jul  9 14:39:54.438: INFO: Terminating { ReplicationController} proxy-service-wms7n pods took: 33.13µs
Jul  9 14:40:02.738: INFO: Garbage collecting { ReplicationController} proxy-service-wms7n pods took: 8.418059241s
[AfterEach] version v1
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:40:02.738: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-proxy-6ksf7" for this suite.
Jul  9 14:40:08.818: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:40:09.528: INFO: namespace: e2e-tests-proxy-6ksf7, resource: bindings, ignored listing per whitelist
Jul  9 14:40:10.289: INFO: namespace e2e-tests-proxy-6ksf7 deletion completed in 7.521045397s

• [SLOW TEST:29.516 seconds]
[sig-network] Proxy
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
  version v1
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:60
    should proxy through a service and a pod  [Conformance]
    /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
[sig-storage] Downward API volume 
  should provide node allocatable (memory) as default memory limit if the limit is not set  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [sig-storage] Downward API volume
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:40:10.289: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Downward API volume
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:38
[It] should provide node allocatable (memory) as default memory limit if the limit is not set  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
STEP: Creating a pod to test downward API volume plugin
Jul  9 14:40:10.445: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fb444ee2-8385-11e8-b2e3-0e8bc4815172" in namespace "e2e-tests-downward-api-p2h2q" to be "success or failure"
Jul  9 14:40:10.460: INFO: Pod "downwardapi-volume-fb444ee2-8385-11e8-b2e3-0e8bc4815172": Phase="Pending", Reason="", readiness=false. Elapsed: 15.105866ms
Jul  9 14:40:12.477: INFO: Pod "downwardapi-volume-fb444ee2-8385-11e8-b2e3-0e8bc4815172": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031918592s
Jul  9 14:40:14.493: INFO: Pod "downwardapi-volume-fb444ee2-8385-11e8-b2e3-0e8bc4815172": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.048197303s
STEP: Saw pod success
Jul  9 14:40:14.493: INFO: Pod "downwardapi-volume-fb444ee2-8385-11e8-b2e3-0e8bc4815172" satisfied condition "success or failure"
Jul  9 14:40:14.509: INFO: Trying to get logs from node prtest-7d230a7-201-ig-n-jsvv pod downwardapi-volume-fb444ee2-8385-11e8-b2e3-0e8bc4815172 container client-container: <nil>
STEP: delete the pod
Jul  9 14:40:14.552: INFO: Waiting for pod downwardapi-volume-fb444ee2-8385-11e8-b2e3-0e8bc4815172 to disappear
Jul  9 14:40:14.568: INFO: Pod downwardapi-volume-fb444ee2-8385-11e8-b2e3-0e8bc4815172 no longer exists
[AfterEach] [sig-storage] Downward API volume
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:40:14.568: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-downward-api-p2h2q" for this suite.
Jul  9 14:40:20.650: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:40:21.710: INFO: namespace: e2e-tests-downward-api-p2h2q, resource: bindings, ignored listing per whitelist
Jul  9 14:40:22.117: INFO: namespace e2e-tests-downward-api-p2h2q deletion completed in 7.517543113s

• [SLOW TEST:11.828 seconds]
[sig-storage] Downward API volume
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:33
  should provide node allocatable (memory) as default memory limit if the limit is not set  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
SSSSSSSSSSSSSSSSS
------------------------------
[sig-network] DNS 
  should provide DNS for services  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [sig-network] DNS
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:40:22.118: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[It] should provide DNS for services  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
STEP: Creating a test headless service
STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(dig +notcp +noall +answer +search dns-test-service A)" && echo OK > /results/wheezy_udp@dns-test-service;test -n "$$(dig +tcp +noall +answer +search dns-test-service A)" && echo OK > /results/wheezy_tcp@dns-test-service;test -n "$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-n4xdd A)" && echo OK > /results/wheezy_udp@dns-test-service.e2e-tests-dns-n4xdd;test -n "$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-n4xdd A)" && echo OK > /results/wheezy_tcp@dns-test-service.e2e-tests-dns-n4xdd;test -n "$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-n4xdd.svc A)" && echo OK > /results/wheezy_udp@dns-test-service.e2e-tests-dns-n4xdd.svc;test -n "$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-n4xdd.svc A)" && echo OK > /results/wheezy_tcp@dns-test-service.e2e-tests-dns-n4xdd.svc;test -n "$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-n4xdd.svc SRV)" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.e2e-tests-dns-n4xdd.svc;test -n "$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-n4xdd.svc SRV)" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.e2e-tests-dns-n4xdd.svc;test -n "$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-n4xdd.svc SRV)" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.e2e-tests-dns-n4xdd.svc;test -n "$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-n4xdd.svc SRV)" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.e2e-tests-dns-n4xdd.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-n4xdd.pod.cluster.local"}');test -n "$$(dig +notcp +noall +answer +search $${podARec} A)" && echo OK > /results/wheezy_udp@PodARecord;test -n "$$(dig +tcp +noall +answer +search $${podARec} A)" && echo OK > /results/wheezy_tcp@PodARecord;test -n "$$(dig +notcp +noall +answer +search 110.35.30.172.in-addr.arpa. PTR)" && echo OK > /results/172.30.35.110_udp@PTR;test -n "$$(dig +tcp +noall +answer +search 110.35.30.172.in-addr.arpa. PTR)" && echo OK > /results/172.30.35.110_tcp@PTR;sleep 1; done

STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(dig +notcp +noall +answer +search dns-test-service A)" && echo OK > /results/jessie_udp@dns-test-service;test -n "$$(dig +tcp +noall +answer +search dns-test-service A)" && echo OK > /results/jessie_tcp@dns-test-service;test -n "$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-n4xdd A)" && echo OK > /results/jessie_udp@dns-test-service.e2e-tests-dns-n4xdd;test -n "$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-n4xdd A)" && echo OK > /results/jessie_tcp@dns-test-service.e2e-tests-dns-n4xdd;test -n "$$(dig +notcp +noall +answer +search dns-test-service.e2e-tests-dns-n4xdd.svc A)" && echo OK > /results/jessie_udp@dns-test-service.e2e-tests-dns-n4xdd.svc;test -n "$$(dig +tcp +noall +answer +search dns-test-service.e2e-tests-dns-n4xdd.svc A)" && echo OK > /results/jessie_tcp@dns-test-service.e2e-tests-dns-n4xdd.svc;test -n "$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-n4xdd.svc SRV)" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.e2e-tests-dns-n4xdd.svc;test -n "$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.e2e-tests-dns-n4xdd.svc SRV)" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.e2e-tests-dns-n4xdd.svc;test -n "$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-n4xdd.svc SRV)" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.e2e-tests-dns-n4xdd.svc;test -n "$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.e2e-tests-dns-n4xdd.svc SRV)" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.e2e-tests-dns-n4xdd.svc;podARec=$$(hostname -i| awk -F. '{print $$1"-"$$2"-"$$3"-"$$4".e2e-tests-dns-n4xdd.pod.cluster.local"}');test -n "$$(dig +notcp +noall +answer +search $${podARec} A)" && echo OK > /results/jessie_udp@PodARecord;test -n "$$(dig +tcp +noall +answer +search $${podARec} A)" && echo OK > /results/jessie_tcp@PodARecord;test -n "$$(dig +notcp +noall +answer +search 110.35.30.172.in-addr.arpa. PTR)" && echo OK > /results/172.30.35.110_udp@PTR;test -n "$$(dig +tcp +noall +answer +search 110.35.30.172.in-addr.arpa. PTR)" && echo OK > /results/172.30.35.110_tcp@PTR;sleep 1; done

STEP: creating a pod to probe DNS
STEP: submitting the pod to kubernetes
STEP: retrieving the pod
STEP: looking for the results for each expected name from probers
Jul  9 14:40:44.915: INFO: DNS probes using dns-test-025cc620-8386-11e8-b2e3-0e8bc4815172 succeeded

STEP: deleting the pod
STEP: deleting the test service
STEP: deleting the test headless service
[AfterEach] [sig-network] DNS
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:40:45.068: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-dns-n4xdd" for this suite.
Jul  9 14:40:51.150: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:40:52.413: INFO: namespace: e2e-tests-dns-n4xdd, resource: bindings, ignored listing per whitelist
Jul  9 14:40:52.628: INFO: namespace e2e-tests-dns-n4xdd deletion completed in 7.528128778s

• [SLOW TEST:30.510 seconds]
[sig-network] DNS
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
  should provide DNS for services  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
S
------------------------------
[sig-cli] Kubectl client [k8s.io] Update Demo 
  should create and stop a replication controller  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [sig-cli] Kubectl client
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:40:52.628: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-cli] Kubectl client
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:244
[BeforeEach] [k8s.io] Update Demo
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/kubectl.go:264
[It] should create and stop a replication controller  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
STEP: creating a replication controller
Jul  9 14:40:52.775: INFO: Running '/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/bin/linux/amd64/kubectl --kubeconfig=/tmp/cluster-admin.kubeconfig create -f - --namespace=e2e-tests-kubectl-4lw8w'
Jul  9 14:40:53.101: INFO: stderr: ""
Jul  9 14:40:53.101: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" created\n"
STEP: waiting for all containers in name=update-demo pods to come up.
Jul  9 14:40:53.101: INFO: Running '/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/bin/linux/amd64/kubectl --kubeconfig=/tmp/cluster-admin.kubeconfig get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-4lw8w'
Jul  9 14:40:53.277: INFO: stderr: ""
Jul  9 14:40:53.277: INFO: stdout: "update-demo-nautilus-245f8 update-demo-nautilus-2mbl4 "
Jul  9 14:40:53.277: INFO: Running '/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/bin/linux/amd64/kubectl --kubeconfig=/tmp/cluster-admin.kubeconfig get pods update-demo-nautilus-245f8 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-4lw8w'
Jul  9 14:40:53.452: INFO: stderr: ""
Jul  9 14:40:53.452: INFO: stdout: ""
Jul  9 14:40:53.452: INFO: update-demo-nautilus-245f8 is created but not running
Jul  9 14:40:58.452: INFO: Running '/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/bin/linux/amd64/kubectl --kubeconfig=/tmp/cluster-admin.kubeconfig get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo --namespace=e2e-tests-kubectl-4lw8w'
Jul  9 14:40:58.630: INFO: stderr: ""
Jul  9 14:40:58.630: INFO: stdout: "update-demo-nautilus-245f8 update-demo-nautilus-2mbl4 "
Jul  9 14:40:58.630: INFO: Running '/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/bin/linux/amd64/kubectl --kubeconfig=/tmp/cluster-admin.kubeconfig get pods update-demo-nautilus-245f8 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-4lw8w'
Jul  9 14:40:58.804: INFO: stderr: ""
Jul  9 14:40:58.804: INFO: stdout: "true"
Jul  9 14:40:58.804: INFO: Running '/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/bin/linux/amd64/kubectl --kubeconfig=/tmp/cluster-admin.kubeconfig get pods update-demo-nautilus-245f8 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-4lw8w'
Jul  9 14:40:58.979: INFO: stderr: ""
Jul  9 14:40:58.979: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus-amd64:1.0"
Jul  9 14:40:58.979: INFO: validating pod update-demo-nautilus-245f8
Jul  9 14:40:58.999: INFO: got data: {
  "image": "nautilus.jpg"
}

Jul  9 14:40:58.999: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
Jul  9 14:40:58.999: INFO: update-demo-nautilus-245f8 is verified up and running
Jul  9 14:40:58.999: INFO: Running '/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/bin/linux/amd64/kubectl --kubeconfig=/tmp/cluster-admin.kubeconfig get pods update-demo-nautilus-2mbl4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-4lw8w'
Jul  9 14:40:59.173: INFO: stderr: ""
Jul  9 14:40:59.173: INFO: stdout: "true"
Jul  9 14:40:59.173: INFO: Running '/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/bin/linux/amd64/kubectl --kubeconfig=/tmp/cluster-admin.kubeconfig get pods update-demo-nautilus-2mbl4 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}} --namespace=e2e-tests-kubectl-4lw8w'
Jul  9 14:40:59.350: INFO: stderr: ""
Jul  9 14:40:59.350: INFO: stdout: "gcr.io/kubernetes-e2e-test-images/nautilus-amd64:1.0"
Jul  9 14:40:59.350: INFO: validating pod update-demo-nautilus-2mbl4
Jul  9 14:40:59.369: INFO: got data: {
  "image": "nautilus.jpg"
}

Jul  9 14:40:59.369: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg .
Jul  9 14:40:59.369: INFO: update-demo-nautilus-2mbl4 is verified up and running
STEP: using delete to clean up resources
Jul  9 14:40:59.369: INFO: Running '/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/bin/linux/amd64/kubectl --kubeconfig=/tmp/cluster-admin.kubeconfig delete --grace-period=0 --force -f - --namespace=e2e-tests-kubectl-4lw8w'
Jul  9 14:40:59.652: INFO: stderr: "warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n"
Jul  9 14:40:59.652: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" deleted\n"
Jul  9 14:40:59.652: INFO: Running '/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/bin/linux/amd64/kubectl --kubeconfig=/tmp/cluster-admin.kubeconfig get rc,svc -l name=update-demo --no-headers --namespace=e2e-tests-kubectl-4lw8w'
Jul  9 14:40:59.845: INFO: stderr: "No resources found.\n"
Jul  9 14:40:59.845: INFO: stdout: ""
Jul  9 14:40:59.845: INFO: Running '/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/bin/linux/amd64/kubectl --kubeconfig=/tmp/cluster-admin.kubeconfig get pods -l name=update-demo --namespace=e2e-tests-kubectl-4lw8w -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}'
Jul  9 14:41:00.020: INFO: stderr: ""
Jul  9 14:41:00.021: INFO: stdout: ""
[AfterEach] [sig-cli] Kubectl client
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:41:00.021: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-kubectl-4lw8w" for this suite.
Jul  9 14:41:22.100: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:41:23.383: INFO: namespace: e2e-tests-kubectl-4lw8w, resource: bindings, ignored listing per whitelist
Jul  9 14:41:23.568: INFO: namespace e2e-tests-kubectl-4lw8w deletion completed in 23.518392505s

• [SLOW TEST:30.940 seconds]
[sig-cli] Kubectl client
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/kubectl/framework.go:22
  [k8s.io] Update Demo
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:643
    should create and stop a replication controller  [Conformance]
    /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
S
------------------------------
[sig-storage] Projected 
  should update labels on modification [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [sig-storage] Projected
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:41:23.568: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Projected
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected.go:858
[It] should update labels on modification [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
STEP: Creating the pod
Jul  9 14:41:28.345: INFO: Successfully updated pod "labelsupdate26f2769f-8386-11e8-b2e3-0e8bc4815172"
[AfterEach] [sig-storage] Projected
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:41:30.389: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-projected-9vgr6" for this suite.
Jul  9 14:41:52.469: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:41:53.142: INFO: namespace: e2e-tests-projected-9vgr6, resource: bindings, ignored listing per whitelist
Jul  9 14:41:53.990: INFO: namespace e2e-tests-projected-9vgr6 deletion completed in 23.571843135s

• [SLOW TEST:30.422 seconds]
[sig-storage] Projected
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected.go:34
  should update labels on modification [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
SSSSSS
------------------------------
[sig-storage] Projected 
  should provide container's memory limit [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [sig-storage] Projected
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:41:53.990: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Projected
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected.go:858
[It] should provide container's memory limit [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
STEP: Creating a pod to test downward API volume plugin
Jul  9 14:41:54.124: INFO: Waiting up to 5m0s for pod "downwardapi-volume-390d9b25-8386-11e8-b2e3-0e8bc4815172" in namespace "e2e-tests-projected-7jnw4" to be "success or failure"
Jul  9 14:41:54.144: INFO: Pod "downwardapi-volume-390d9b25-8386-11e8-b2e3-0e8bc4815172": Phase="Pending", Reason="", readiness=false. Elapsed: 20.421545ms
Jul  9 14:41:56.161: INFO: Pod "downwardapi-volume-390d9b25-8386-11e8-b2e3-0e8bc4815172": Phase="Pending", Reason="", readiness=false. Elapsed: 2.03702514s
Jul  9 14:41:58.177: INFO: Pod "downwardapi-volume-390d9b25-8386-11e8-b2e3-0e8bc4815172": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.053271024s
STEP: Saw pod success
Jul  9 14:41:58.177: INFO: Pod "downwardapi-volume-390d9b25-8386-11e8-b2e3-0e8bc4815172" satisfied condition "success or failure"
Jul  9 14:41:58.193: INFO: Trying to get logs from node prtest-7d230a7-201-ig-n-bw59 pod downwardapi-volume-390d9b25-8386-11e8-b2e3-0e8bc4815172 container client-container: <nil>
STEP: delete the pod
Jul  9 14:41:58.246: INFO: Waiting for pod downwardapi-volume-390d9b25-8386-11e8-b2e3-0e8bc4815172 to disappear
Jul  9 14:41:58.262: INFO: Pod downwardapi-volume-390d9b25-8386-11e8-b2e3-0e8bc4815172 no longer exists
[AfterEach] [sig-storage] Projected
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:41:58.262: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-projected-7jnw4" for this suite.
Jul  9 14:42:04.341: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:42:05.037: INFO: namespace: e2e-tests-projected-7jnw4, resource: bindings, ignored listing per whitelist
Jul  9 14:42:05.797: INFO: namespace e2e-tests-projected-7jnw4 deletion completed in 7.506028951s

• [SLOW TEST:11.806 seconds]
[sig-storage] Projected
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected.go:34
  should provide container's memory limit [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
SS
------------------------------
[k8s.io] Probing container 
  should have monotonically increasing restart count  [Slow] [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [k8s.io] Probing container
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:42:05.797: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [k8s.io] Probing container
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/container_probe.go:48
[It] should have monotonically increasing restart count  [Slow] [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
STEP: Creating pod liveness-http in namespace e2e-tests-container-probe-d7gbl
Jul  9 14:42:09.973: INFO: Started pod liveness-http in namespace e2e-tests-container-probe-d7gbl
STEP: checking the pod's current state and verifying that restartCount is present
Jul  9 14:42:09.989: INFO: Initial restart count of pod liveness-http is 0
Jul  9 14:42:28.150: INFO: Restart count of pod e2e-tests-container-probe-d7gbl/liveness-http is now 1 (18.161214621s elapsed)
Jul  9 14:42:48.311: INFO: Restart count of pod e2e-tests-container-probe-d7gbl/liveness-http is now 2 (38.322062269s elapsed)
Jul  9 14:43:08.474: INFO: Restart count of pod e2e-tests-container-probe-d7gbl/liveness-http is now 3 (58.48519108s elapsed)
Jul  9 14:43:28.635: INFO: Restart count of pod e2e-tests-container-probe-d7gbl/liveness-http is now 4 (1m18.645707795s elapsed)
Jul  9 14:44:27.104: INFO: Restart count of pod e2e-tests-container-probe-d7gbl/liveness-http is now 5 (2m17.115280264s elapsed)
STEP: deleting the pod
[AfterEach] [k8s.io] Probing container
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:44:27.124: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-container-probe-d7gbl" for this suite.
Jul  9 14:44:33.203: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:44:34.052: INFO: namespace: e2e-tests-container-probe-d7gbl, resource: bindings, ignored listing per whitelist
Jul  9 14:44:34.661: INFO: namespace e2e-tests-container-probe-d7gbl deletion completed in 7.507290642s

• [SLOW TEST:148.864 seconds]
[k8s.io] Probing container
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:643
  should have monotonically increasing restart count  [Slow] [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
SSS
------------------------------
[k8s.io] Pods 
  should be submitted and removed  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [k8s.io] Pods
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:44:34.661: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [k8s.io] Pods
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/pods.go:127
[It] should be submitted and removed  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
STEP: creating the pod
STEP: setting up watch
STEP: submitting the pod to kubernetes
STEP: verifying the pod is in kubernetes
STEP: verifying pod creation was observed
Jul  9 14:44:38.931: INFO: running pod: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-submit-remove-98d8719d-8386-11e8-b2e3-0e8bc4815172", GenerateName:"", Namespace:"e2e-tests-pods-vjthz", SelfLink:"/api/v1/namespaces/e2e-tests-pods-vjthz/pods/pod-submit-remove-98d8719d-8386-11e8-b2e3-0e8bc4815172", UID:"98e0278f-8386-11e8-8047-42010a8e0002", ResourceVersion:"20847", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63666744274, loc:(*time.Location)(0x621c140)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"797818941"}, Annotations:map[string]string{"openshift.io/scc":"privileged"}, OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-vsnvn", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc4220d5b40), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil)}}}, InitContainers:[]v1.Container(nil), Containers:[]v1.Container{v1.Container{Name:"nginx", Image:"gcr.io/google-containers/nginx-slim-amd64:0.20", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-vsnvn", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc421b3e768), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string{"role":"app"}, ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"prtest-7d230a7-201-ig-n-jn9j", HostNetwork:false, HostPID:false, HostIPC:false, SecurityContext:(*v1.PodSecurityContext)(0xc4220d5d40), ImagePullSecrets:[]v1.LocalObjectReference{v1.LocalObjectReference{Name:"default-dockercfg-79cfq"}}, Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration(nil), HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(nil), DNSConfig:(*v1.PodDNSConfig)(nil)}, Status:v1.PodStatus{Phase:"Running", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63666744274, loc:(*time.Location)(0x621c140)}}, Reason:"", Message:""}, v1.PodCondition{Type:"Ready", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63666744278, loc:(*time.Location)(0x621c140)}}, Reason:"", Message:""}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63666744274, loc:(*time.Location)(0x621c140)}}, Reason:"", Message:""}}, Message:"", Reason:"", HostIP:"10.142.0.4", PodIP:"172.16.0.67", StartTime:(*v1.Time)(0xc421a09700), InitContainerStatuses:[]v1.ContainerStatus(nil), ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"nginx", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(0xc421a09720), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:true, RestartCount:0, Image:"gcr.io/google-containers/nginx-slim-amd64:0.20", ImageID:"docker-pullable://gcr.io/google-containers/nginx-slim-amd64@sha256:6654db6d4028756062edac466454ee5c9cf9b20ef79e35a81e3c840031eb1e2b", ContainerID:"docker://cb96530a9c22af5795cfffc1056a174d3e259e6a0bb0ce3ff24199430d979455"}}, QOSClass:"BestEffort"}}
STEP: deleting the pod gracefully
STEP: verifying the kubelet observed the termination notice
Jul  9 14:44:43.978: INFO: no pod exists with the name we were looking for, assuming the termination request was observed and completed
STEP: verifying pod deletion was observed
[AfterEach] [k8s.io] Pods
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:44:43.995: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-pods-vjthz" for this suite.
Jul  9 14:44:50.076: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:44:51.396: INFO: namespace: e2e-tests-pods-vjthz, resource: bindings, ignored listing per whitelist
Jul  9 14:44:51.537: INFO: namespace e2e-tests-pods-vjthz deletion completed in 7.512066288s

• [SLOW TEST:16.876 seconds]
[k8s.io] Pods
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:643
  should be submitted and removed  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
SSSSS
------------------------------
[sig-storage] Secrets 
  optional updates should be reflected in volume  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [sig-storage] Secrets
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:44:51.537: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[It] optional updates should be reflected in volume  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
Jul  9 14:44:51.685: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
STEP: Creating secret with name s-test-opt-del-a2e94f1a-8386-11e8-b2e3-0e8bc4815172
STEP: Creating secret with name s-test-opt-upd-a2e94f5b-8386-11e8-b2e3-0e8bc4815172
STEP: Creating the pod
STEP: Deleting secret s-test-opt-del-a2e94f1a-8386-11e8-b2e3-0e8bc4815172
STEP: Updating secret s-test-opt-upd-a2e94f5b-8386-11e8-b2e3-0e8bc4815172
STEP: Creating secret with name s-test-opt-create-a2e94f82-8386-11e8-b2e3-0e8bc4815172
STEP: waiting to observe update in volume
[AfterEach] [sig-storage] Secrets
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:46:04.793: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-secrets-6tvmq" for this suite.
Jul  9 14:46:26.873: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:46:28.058: INFO: namespace: e2e-tests-secrets-6tvmq, resource: bindings, ignored listing per whitelist
Jul  9 14:46:28.335: INFO: namespace e2e-tests-secrets-6tvmq deletion completed in 23.512328908s

• [SLOW TEST:96.798 seconds]
[sig-storage] Secrets
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33
  optional updates should be reflected in volume  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
SSSSSSSS
------------------------------
[sig-storage] Projected 
  should provide podname only [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [sig-storage] Projected
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:46:28.336: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Projected
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected.go:858
[It] should provide podname only [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
STEP: Creating a pod to test downward API volume plugin
Jul  9 14:46:28.474: INFO: Waiting up to 5m0s for pod "downwardapi-volume-dc96f075-8386-11e8-b2e3-0e8bc4815172" in namespace "e2e-tests-projected-pkx8k" to be "success or failure"
Jul  9 14:46:28.492: INFO: Pod "downwardapi-volume-dc96f075-8386-11e8-b2e3-0e8bc4815172": Phase="Pending", Reason="", readiness=false. Elapsed: 17.521759ms
Jul  9 14:46:30.509: INFO: Pod "downwardapi-volume-dc96f075-8386-11e8-b2e3-0e8bc4815172": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034394084s
Jul  9 14:46:32.525: INFO: Pod "downwardapi-volume-dc96f075-8386-11e8-b2e3-0e8bc4815172": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.050736134s
STEP: Saw pod success
Jul  9 14:46:32.525: INFO: Pod "downwardapi-volume-dc96f075-8386-11e8-b2e3-0e8bc4815172" satisfied condition "success or failure"
Jul  9 14:46:32.541: INFO: Trying to get logs from node prtest-7d230a7-201-ig-n-jsvv pod downwardapi-volume-dc96f075-8386-11e8-b2e3-0e8bc4815172 container client-container: <nil>
STEP: delete the pod
Jul  9 14:46:32.585: INFO: Waiting for pod downwardapi-volume-dc96f075-8386-11e8-b2e3-0e8bc4815172 to disappear
Jul  9 14:46:32.601: INFO: Pod downwardapi-volume-dc96f075-8386-11e8-b2e3-0e8bc4815172 no longer exists
[AfterEach] [sig-storage] Projected
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:46:32.601: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-projected-pkx8k" for this suite.
Jul  9 14:46:38.683: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:46:40.073: INFO: namespace: e2e-tests-projected-pkx8k, resource: bindings, ignored listing per whitelist
Jul  9 14:46:40.136: INFO: namespace e2e-tests-projected-pkx8k deletion completed in 7.503286794s

• [SLOW TEST:11.800 seconds]
[sig-storage] Projected
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/projected.go:34
  should provide podname only [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
SSSS
------------------------------
[sig-api-machinery] ConfigMap 
  should be consumable via the environment  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [sig-api-machinery] ConfigMap
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:46:40.136: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[It] should be consumable via the environment  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
STEP: Creating configMap e2e-tests-configmap-xh7h8/configmap-test-e3a1c22e-8386-11e8-b2e3-0e8bc4815172
STEP: Creating a pod to test consume configMaps
Jul  9 14:46:40.316: INFO: Waiting up to 5m0s for pod "pod-configmaps-e3a479fb-8386-11e8-b2e3-0e8bc4815172" in namespace "e2e-tests-configmap-xh7h8" to be "success or failure"
Jul  9 14:46:40.334: INFO: Pod "pod-configmaps-e3a479fb-8386-11e8-b2e3-0e8bc4815172": Phase="Pending", Reason="", readiness=false. Elapsed: 17.893607ms
Jul  9 14:46:42.350: INFO: Pod "pod-configmaps-e3a479fb-8386-11e8-b2e3-0e8bc4815172": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033823722s
Jul  9 14:46:44.366: INFO: Pod "pod-configmaps-e3a479fb-8386-11e8-b2e3-0e8bc4815172": Phase="Pending", Reason="", readiness=false. Elapsed: 4.050134508s
Jul  9 14:46:46.382: INFO: Pod "pod-configmaps-e3a479fb-8386-11e8-b2e3-0e8bc4815172": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.06585691s
STEP: Saw pod success
Jul  9 14:46:46.382: INFO: Pod "pod-configmaps-e3a479fb-8386-11e8-b2e3-0e8bc4815172" satisfied condition "success or failure"
Jul  9 14:46:46.397: INFO: Trying to get logs from node prtest-7d230a7-201-ig-n-jn9j pod pod-configmaps-e3a479fb-8386-11e8-b2e3-0e8bc4815172 container env-test: <nil>
STEP: delete the pod
Jul  9 14:46:46.446: INFO: Waiting for pod pod-configmaps-e3a479fb-8386-11e8-b2e3-0e8bc4815172 to disappear
Jul  9 14:46:46.466: INFO: Pod pod-configmaps-e3a479fb-8386-11e8-b2e3-0e8bc4815172 no longer exists
[AfterEach] [sig-api-machinery] ConfigMap
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:46:46.466: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-configmap-xh7h8" for this suite.
Jul  9 14:46:52.547: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:46:53.922: INFO: namespace: e2e-tests-configmap-xh7h8, resource: bindings, ignored listing per whitelist
Jul  9 14:46:53.983: INFO: namespace e2e-tests-configmap-xh7h8 deletion completed in 7.487821843s

• [SLOW TEST:13.847 seconds]
[sig-api-machinery] ConfigMap
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap.go:29
  should be consumable via the environment  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
[sig-apps] ReplicationController 
  should serve a basic image on each replica with a public image  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [sig-apps] ReplicationController
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:46:53.983: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[It] should serve a basic image on each replica with a public image  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
STEP: Creating replication controller my-hostname-basic-ebe3e427-8386-11e8-b2e3-0e8bc4815172
Jul  9 14:46:54.162: INFO: Pod name my-hostname-basic-ebe3e427-8386-11e8-b2e3-0e8bc4815172: Found 1 pods out of 1
Jul  9 14:46:54.162: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-ebe3e427-8386-11e8-b2e3-0e8bc4815172" are running
Jul  9 14:46:58.199: INFO: Pod "my-hostname-basic-ebe3e427-8386-11e8-b2e3-0e8bc4815172-krdtl" is running (conditions: [])
Jul  9 14:46:58.199: INFO: Trying to dial the pod
Jul  9 14:47:03.250: INFO: Controller my-hostname-basic-ebe3e427-8386-11e8-b2e3-0e8bc4815172: Got expected result from replica 1 [my-hostname-basic-ebe3e427-8386-11e8-b2e3-0e8bc4815172-krdtl]: "my-hostname-basic-ebe3e427-8386-11e8-b2e3-0e8bc4815172-krdtl", 1 of 1 required successes so far
[AfterEach] [sig-apps] ReplicationController
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:47:03.250: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-replication-controller-vwdvp" for this suite.
Jul  9 14:47:09.328: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:47:10.608: INFO: namespace: e2e-tests-replication-controller-vwdvp, resource: bindings, ignored listing per whitelist
Jul  9 14:47:10.780: INFO: namespace e2e-tests-replication-controller-vwdvp deletion completed in 7.500614402s

• [SLOW TEST:16.797 seconds]
[sig-apps] ReplicationController
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/apps/framework.go:22
  should serve a basic image on each replica with a public image  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
[sig-storage] ConfigMap 
  optional updates should be reflected in volume  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [sig-storage] ConfigMap
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:47:10.780: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[It] optional updates should be reflected in volume  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
Jul  9 14:47:10.935: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node
STEP: Creating configMap with name cm-test-opt-del-f5e933aa-8386-11e8-b2e3-0e8bc4815172
STEP: Creating configMap with name cm-test-opt-upd-f5e93406-8386-11e8-b2e3-0e8bc4815172
STEP: Creating the pod
STEP: Deleting configmap cm-test-opt-del-f5e933aa-8386-11e8-b2e3-0e8bc4815172
STEP: Updating configmap cm-test-opt-upd-f5e93406-8386-11e8-b2e3-0e8bc4815172
STEP: Creating configMap with name cm-test-opt-create-f5e93433-8386-11e8-b2e3-0e8bc4815172
STEP: waiting to observe update in volume
[AfterEach] [sig-storage] ConfigMap
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:48:26.037: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-configmap-6wmvl" for this suite.
Jul  9 14:48:48.116: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:48:49.208: INFO: namespace: e2e-tests-configmap-6wmvl, resource: bindings, ignored listing per whitelist
Jul  9 14:48:49.578: INFO: namespace e2e-tests-configmap-6wmvl deletion completed in 23.512218947s

• [SLOW TEST:98.798 seconds]
[sig-storage] ConfigMap
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32
  optional updates should be reflected in volume  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
SSSS
------------------------------
[sig-storage] Downward API volume 
  should provide node allocatable (cpu) as default cpu limit if the limit is not set  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] [sig-storage] Downward API volume
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:48:49.578: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] [sig-storage] Downward API volume
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:38
[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
STEP: Creating a pod to test downward API volume plugin
Jul  9 14:48:49.760: INFO: Waiting up to 5m0s for pod "downwardapi-volume-30cd58fb-8387-11e8-b2e3-0e8bc4815172" in namespace "e2e-tests-downward-api-dvrz2" to be "success or failure"
Jul  9 14:48:49.778: INFO: Pod "downwardapi-volume-30cd58fb-8387-11e8-b2e3-0e8bc4815172": Phase="Pending", Reason="", readiness=false. Elapsed: 18.38913ms
Jul  9 14:48:51.796: INFO: Pod "downwardapi-volume-30cd58fb-8387-11e8-b2e3-0e8bc4815172": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035871573s
Jul  9 14:48:53.812: INFO: Pod "downwardapi-volume-30cd58fb-8387-11e8-b2e3-0e8bc4815172": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.051895427s
STEP: Saw pod success
Jul  9 14:48:53.812: INFO: Pod "downwardapi-volume-30cd58fb-8387-11e8-b2e3-0e8bc4815172" satisfied condition "success or failure"
Jul  9 14:48:53.827: INFO: Trying to get logs from node prtest-7d230a7-201-ig-n-jn9j pod downwardapi-volume-30cd58fb-8387-11e8-b2e3-0e8bc4815172 container client-container: <nil>
STEP: delete the pod
Jul  9 14:48:53.940: INFO: Waiting for pod downwardapi-volume-30cd58fb-8387-11e8-b2e3-0e8bc4815172 to disappear
Jul  9 14:48:53.955: INFO: Pod downwardapi-volume-30cd58fb-8387-11e8-b2e3-0e8bc4815172 no longer exists
[AfterEach] [sig-storage] Downward API volume
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:48:53.955: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-downward-api-dvrz2" for this suite.
Jul  9 14:49:00.036: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:49:00.738: INFO: namespace: e2e-tests-downward-api-dvrz2, resource: bindings, ignored listing per whitelist
Jul  9 14:49:01.474: INFO: namespace e2e-tests-downward-api-dvrz2 deletion completed in 7.48847652s

• [SLOW TEST:11.896 seconds]
[sig-storage] Downward API volume
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:33
  should provide node allocatable (cpu) as default cpu limit if the limit is not set  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
SSSSS
------------------------------
[sig-network] Proxy version v1 
  should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
[BeforeEach] version v1
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:134
STEP: Creating a kubernetes client
Jul  9 14:49:01.474: INFO: >>> kubeConfig: /tmp/cluster-admin.kubeconfig
STEP: Building a namespace api object
STEP: Waiting for a default service account to be provisioned in namespace
[BeforeEach] version v1
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:68
[It] should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
Jul  9 14:49:01.674: INFO: (0) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 22.146303ms)
Jul  9 14:49:01.691: INFO: (1) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 16.996885ms)
Jul  9 14:49:01.708: INFO: (2) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 17.261723ms)
Jul  9 14:49:01.726: INFO: (3) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 17.475799ms)
Jul  9 14:49:01.743: INFO: (4) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 17.663148ms)
Jul  9 14:49:01.763: INFO: (5) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 19.91639ms)
Jul  9 14:49:01.781: INFO: (6) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 17.134238ms)
Jul  9 14:49:01.798: INFO: (7) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 17.154822ms)
Jul  9 14:49:01.816: INFO: (8) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 18.160551ms)
Jul  9 14:49:01.834: INFO: (9) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 18.208771ms)
Jul  9 14:49:01.854: INFO: (10) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 19.90107ms)
Jul  9 14:49:01.872: INFO: (11) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 17.572373ms)
Jul  9 14:49:01.889: INFO: (12) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 17.189985ms)
Jul  9 14:49:01.907: INFO: (13) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 18.128932ms)
Jul  9 14:49:01.924: INFO: (14) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 17.372922ms)
Jul  9 14:49:01.942: INFO: (15) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 17.59085ms)
Jul  9 14:49:01.960: INFO: (16) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 17.780756ms)
Jul  9 14:49:01.977: INFO: (17) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 17.089876ms)
Jul  9 14:49:01.995: INFO: (18) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 18.340494ms)
Jul  9 14:49:02.014: INFO: (19) /api/v1/nodes/prtest-7d230a7-201-ig-n-bw59:10250/proxy/logs/: <pre>
<a href="audit/">audit/</a>
<a href="boot.log">boot.log</a>
<a href="btmp">btmp</a>
<a href... (200; 18.404546ms)
[AfterEach] version v1
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:135
Jul  9 14:49:02.014: INFO: Waiting up to 3m0s for all (but 1) nodes to be ready
STEP: Destroying namespace "e2e-tests-proxy-zkdh9" for this suite.
Jul  9 14:49:08.093: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered
Jul  9 14:49:09.404: INFO: namespace: e2e-tests-proxy-zkdh9, resource: bindings, ignored listing per whitelist
Jul  9 14:49:09.527: INFO: namespace e2e-tests-proxy-zkdh9 deletion completed in 7.483883263s

• [SLOW TEST:8.054 seconds]
[sig-network] Proxy
/data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/framework.go:22
  version v1
  /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/network/proxy.go:60
    should proxy logs on node with explicit kubelet port using proxy subresource  [Conformance]
    /data/src/github.com/openshift/origin/_output/components/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/framework.go:648
------------------------------
SSSSSSSSSSJul  9 14:49:09.528: INFO: Running AfterSuite actions on all node
Jul  9 14:49:09.528: INFO: Running AfterSuite actions on node 1
Jul  9 14:49:09.528: INFO: Dumping logs locally to: /data/src/github.com/openshift/origin/_output/scripts/conformance-k8s/artifacts
Jul  9 14:49:09.528: INFO: Error running cluster/log-dump/log-dump.sh: fork/exec ../../cluster/log-dump/log-dump.sh: no such file or directory

Ran 147 of 852 Specs in 3921.945 seconds
SUCCESS! -- 147 Passed | 0 Failed | 0 Pending | 705 Skipped PASS

Run complete, results in /data/src/github.com/openshift/origin/_output/scripts/conformance-k8s/artifacts
+ gather
+ set +e
++ pwd
+ export PATH=/data/src/github.com/openshift/origin/_output/local/bin/linux/amd64:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/home/origin/.local/bin:/home/origin/bin
+ PATH=/data/src/github.com/openshift/origin/_output/local/bin/linux/amd64:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/home/origin/.local/bin:/home/origin/bin
+ oc get nodes --template '{{ range .items }}{{ .metadata.name }}{{ "\n" }}{{ end }}'
+ xargs -L 1 -I X bash -c 'oc get --raw /api/v1/nodes/X/proxy/metrics > /tmp/artifacts/X.metrics' ''
+ oc get --raw /metrics
+ set -e
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: RUN EXTENDED TESTS [01h 13m 12s] ##########
[workspace] $ /bin/bash /tmp/jenkins6564454191740592193.sh
########## STARTING STAGE: TAG THE LATEST CONFORMANCE RESULTS ##########
+ [[ -s /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ export PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
+ location=origin-ci-test/logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts
+ location_url=https://storage.googleapis.com/origin-ci-test/logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts
+ echo https://storage.googleapis.com/origin-ci-test/logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts
+ gsutil cp .latest-conformance gs://origin-ci-test/releases/openshift/origin/release-3.9/.latest-conformance
Copying file://.latest-conformance [Content-Type=application/octet-stream]...
/ [0 files][    0.0 B/  146.0 B]                                                
/ [1 files][  146.0 B/  146.0 B]                                                
Operation completed over 1 objects/146.0 B.                                      
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: TAG THE LATEST CONFORMANCE RESULTS [00h 00m 01s] ##########
[PostBuildScript] - Executing post build scripts.
[workspace] $ /bin/bash /tmp/jenkins5921006985036357739.sh
########## STARTING STAGE: DOWNLOAD ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ export PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/gathered
+ rm -rf /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/gathered
+ mkdir -p /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/gathered
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo stat /data/src/github.com/openshift/origin/_output/scripts
  File: ‘/data/src/github.com/openshift/origin/_output/scripts’
  Size: 87        	Blocks: 0          IO Block: 4096   directory
Device: ca02h/51714d	Inode: 142629843   Links: 6
Access: (2755/drwxr-sr-x)  Uid: ( 1001/  origin)   Gid: ( 1003/origin-git)
Context: unconfined_u:object_r:container_file_t:s0
Access: 2018-07-09 12:45:25.341951713 +0000
Modify: 2018-07-09 13:37:01.247365277 +0000
Change: 2018-07-09 13:37:01.247365277 +0000
 Birth: -
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod -R o+rX /data/src/github.com/openshift/origin/_output/scripts
+ scp -r -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel:/data/src/github.com/openshift/origin/_output/scripts /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/gathered
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo stat /tmp/artifacts
  File: ‘/tmp/artifacts’
  Size: 160       	Blocks: 0          IO Block: 4096   directory
Device: 27h/39d	Inode: 297006      Links: 3
Access: (0755/drwxr-xr-x)  Uid: ( 1001/  origin)   Gid: ( 1002/  docker)
Context: unconfined_u:object_r:user_tmp_t:s0
Access: 2018-07-09 13:36:00.717191856 +0000
Modify: 2018-07-09 14:49:11.618345806 +0000
Change: 2018-07-09 14:49:11.618345806 +0000
 Birth: -
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod -R o+rX /tmp/artifacts
+ scp -r -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel:/tmp/artifacts /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/gathered
+ tree /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/gathered
/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/gathered
├── artifacts
│   ├── junit
│   ├── master.metrics
│   ├── prtest-7d230a7-201-ig-m-khm7.metrics
│   ├── prtest-7d230a7-201-ig-n-bw59.metrics
│   ├── prtest-7d230a7-201-ig-n-jn9j.metrics
│   └── prtest-7d230a7-201-ig-n-jsvv.metrics
└── scripts
    ├── build-base-images
    │   ├── artifacts
    │   ├── logs
    │   └── openshift.local.home
    ├── conformance-k8s
    │   ├── artifacts
    │   │   ├── e2e.log
    │   │   ├── junit_01.xml
    │   │   ├── nethealth.txt
    │   │   ├── README.md
    │   │   └── version.txt
    │   ├── logs
    │   │   └── scripts.log
    │   └── openshift.local.home
    ├── push-release
    │   ├── artifacts
    │   ├── logs
    │   │   └── scripts.log
    │   └── openshift.local.home
    └── shell
        ├── artifacts
        ├── logs
        │   ├── 32be3508789e6074128eb701067480aca262e91d12ec421358d9fd13b4625bae.json
        │   ├── 9f59665dfbfa4468d8565f690237377cb7f6c9497166271b09d26fa34ea97d1f.json
        │   ├── d387558411bef750ec641ba7f13a523c0197a0bec458d353e0c7a778cfbb5951.json
        │   └── scripts.log
        └── openshift.local.home

19 directories, 16 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins5825059646556695753.sh
########## STARTING STAGE: GENERATE ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ export PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/generated
+ rm -rf /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/generated
+ mkdir /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/generated
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a 2>&1'
  WARNING: You're not using the default seccomp profile
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo cat /etc/sysconfig/docker /etc/sysconfig/docker-network /etc/sysconfig/docker-storage /etc/sysconfig/docker-storage-setup /etc/systemd/system/docker.service 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'oc get --raw /metrics --server=https://$( uname --nodename ):10250 --config=/etc/origin/master/admin.kubeconfig 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'oc get --raw /metrics --config=/etc/origin/master/admin.kubeconfig 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo df -T -h && sudo pvs && sudo vgs && sudo lvs && sudo findmnt --all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo yum list installed 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl --dmesg --no-pager --all --lines=all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl _PID=1 --no-pager --all --lines=all 2>&1'
+ tree /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/generated
/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/generated
├── avc_denials.log
├── containers.log
├── dmesg.log
├── docker.config
├── docker.info
├── filesystem.info
├── installed_packages.log
├── master-metrics.log
├── node-metrics.log
└── pid1.journal

0 directories, 10 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins8602187715856402463.sh
########## STARTING STAGE: FETCH SYSTEMD JOURNALS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ export PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/journals
+ rm -rf /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/journals
+ mkdir /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/journals
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit docker.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
+ tree /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/journals
/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/journals
├── dnsmasq.service
├── docker.service
└── systemd-journald.service

0 directories, 3 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins5022496423440759225.sh
########## STARTING STAGE: ASSEMBLE GCS OUTPUT ##########
+ [[ -s /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ export PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
+ trap 'exit 0' EXIT
+ mkdir -p gcs/artifacts gcs/artifacts/generated gcs/artifacts/journals gcs/artifacts/gathered
++ python -c 'import json; import urllib; print json.load(urllib.urlopen('\''https://ci.openshift.redhat.com/jenkins/job/test_branch_origin_extended_conformance_k8s_39/201/api/json'\''))['\''result'\'']'
+ result=SUCCESS
+ cat
++ date +%s
+ cat /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/builds/201/log
+ cp artifacts/generated/avc_denials.log artifacts/generated/containers.log artifacts/generated/dmesg.log artifacts/generated/docker.config artifacts/generated/docker.info artifacts/generated/filesystem.info artifacts/generated/installed_packages.log artifacts/generated/master-metrics.log artifacts/generated/node-metrics.log artifacts/generated/pid1.journal gcs/artifacts/generated/
+ cp artifacts/journals/dnsmasq.service artifacts/journals/docker.service artifacts/journals/systemd-journald.service gcs/artifacts/journals/
+ cp -r artifacts/gathered/artifacts artifacts/gathered/scripts gcs/artifacts/
++ pwd
+ scp -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config -r /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/gcs openshiftdevel:/data
+ scp -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.config/gcloud/gcs-publisher-credentials.json openshiftdevel:/data/credentials.json
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins5503295541486241561.sh
########## STARTING STAGE: PUSH THE ARTIFACTS AND METADATA ##########
+ [[ -s /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ export PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
++ mktemp
+ script=/tmp/tmp.EErHvwhllW
+ cat
+ chmod +x /tmp/tmp.EErHvwhllW
+ scp -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.EErHvwhllW openshiftdevel:/tmp/tmp.EErHvwhllW
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 300 /tmp/tmp.EErHvwhllW"'
+ cd /home/origin
+ trap 'exit 0' EXIT
+ [[ -n {"type":"periodic","job":"test_branch_origin_extended_conformance_k8s_39","buildid":"6ba167b2-8375-11e8-adf2-0a58ac10c23f","refs":{}} ]]
++ jq --compact-output .buildid
+ [[ "6ba167b2-8375-11e8-adf2-0a58ac10c23f" =~ ^"[0-9]+"$ ]]
+ echo 'Using BUILD_NUMBER'
Using BUILD_NUMBER
++ jq --compact-output '.buildid |= "201"'
+ JOB_SPEC='{"type":"periodic","job":"test_branch_origin_extended_conformance_k8s_39","buildid":"201","refs":{}}'
+ docker run -e 'JOB_SPEC={"type":"periodic","job":"test_branch_origin_extended_conformance_k8s_39","buildid":"201","refs":{}}' -v /data:/data:z registry.svc.ci.openshift.org/ci/gcsupload:latest --dry-run=false --gcs-path=gs://origin-ci-test --gcs-credentials-file=/data/credentials.json --path-strategy=single --default-org=openshift --default-repo=origin /data/gcs/artifacts /data/gcs/build-log.txt /data/gcs/finished.json
Unable to find image 'registry.svc.ci.openshift.org/ci/gcsupload:latest' locally
Trying to pull repository registry.svc.ci.openshift.org/ci/gcsupload ... 
latest: Pulling from registry.svc.ci.openshift.org/ci/gcsupload
605ce1bd3f31: Already exists
dc6346da9948: Already exists
40623dac2e3f: Pulling fs layer
40623dac2e3f: Verifying Checksum
40623dac2e3f: Download complete
40623dac2e3f: Pull complete
Digest: sha256:6daa31ec39dd93b14f14800567e13b9cf36b8daf9b12dc7ac53dcbf556163d30
Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/gcsupload:latest
{"component":"gcsupload","level":"info","msg":"Gathering artifacts from artifact directory: /data/gcs/artifacts","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/artifacts/master.metrics in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/master.metrics\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/artifacts/prtest-7d230a7-201-ig-m-khm7.metrics in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/prtest-7d230a7-201-ig-m-khm7.metrics\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/artifacts/prtest-7d230a7-201-ig-n-bw59.metrics in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/prtest-7d230a7-201-ig-n-bw59.metrics\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/artifacts/prtest-7d230a7-201-ig-n-jn9j.metrics in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/prtest-7d230a7-201-ig-n-jn9j.metrics\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/artifacts/prtest-7d230a7-201-ig-n-jsvv.metrics in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/prtest-7d230a7-201-ig-n-jsvv.metrics\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/avc_denials.log in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/avc_denials.log\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/containers.log in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/containers.log\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/dmesg.log in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/dmesg.log\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.config in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/docker.config\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.info in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/docker.info\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/filesystem.info in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/filesystem.info\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/installed_packages.log in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/installed_packages.log\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/master-metrics.log in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/master-metrics.log\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/node-metrics.log in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/node-metrics.log\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/pid1.journal in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/pid1.journal\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/dnsmasq.service in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/journals/dnsmasq.service\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/docker.service in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/journals/docker.service\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/systemd-journald.service in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/journals/systemd-journald.service\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/conformance-k8s/artifacts/README.md in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/README.md\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/conformance-k8s/artifacts/e2e.log in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/e2e.log\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/conformance-k8s/artifacts/junit_01.xml in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/junit_01.xml\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/conformance-k8s/artifacts/nethealth.txt in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/nethealth.txt\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/conformance-k8s/artifacts/version.txt in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/version.txt\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/conformance-k8s/logs/scripts.log in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/logs/scripts.log\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/push-release/logs/scripts.log in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/push-release/logs/scripts.log\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/shell/logs/32be3508789e6074128eb701067480aca262e91d12ec421358d9fd13b4625bae.json in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/shell/logs/32be3508789e6074128eb701067480aca262e91d12ec421358d9fd13b4625bae.json\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/shell/logs/9f59665dfbfa4468d8565f690237377cb7f6c9497166271b09d26fa34ea97d1f.json in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/shell/logs/9f59665dfbfa4468d8565f690237377cb7f6c9497166271b09d26fa34ea97d1f.json\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/shell/logs/d387558411bef750ec641ba7f13a523c0197a0bec458d353e0c7a778cfbb5951.json in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/shell/logs/d387558411bef750ec641ba7f13a523c0197a0bec458d353e0c7a778cfbb5951.json\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/scripts/shell/logs/scripts.log in artifact directory. Uploading as logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/shell/logs/scripts.log\n","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/journals/dnsmasq.service","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/README.md","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/junit_01.xml","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/build-log.txt","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/journals/systemd-journald.service","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/push-release/logs/scripts.log","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/shell/logs/9f59665dfbfa4468d8565f690237377cb7f6c9497166271b09d26fa34ea97d1f.json","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/shell/logs/scripts.log","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/prtest-7d230a7-201-ig-n-jn9j.metrics","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/avc_denials.log","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/docker.config","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/e2e.log","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/shell/logs/32be3508789e6074128eb701067480aca262e91d12ec421358d9fd13b4625bae.json","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/containers.log","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/shell/logs/d387558411bef750ec641ba7f13a523c0197a0bec458d353e0c7a778cfbb5951.json","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/latest-build.txt","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/prtest-7d230a7-201-ig-n-bw59.metrics","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/pid1.journal","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/logs/scripts.log","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/prtest-7d230a7-201-ig-m-khm7.metrics","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/dmesg.log","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/nethealth.txt","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/version.txt","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/finished.json","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/journals/docker.service","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/master.metrics","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/docker.info","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/installed_packages.log","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/master-metrics.log","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/node-metrics.log","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/prtest-7d230a7-201-ig-n-jsvv.metrics","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/filesystem.info","level":"info","msg":"Queued for upload","time":"2018-07-09T14:49:32Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/docker.config","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/version.txt","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/journals/dnsmasq.service","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/journals/systemd-journald.service","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/avc_denials.log","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/finished.json","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/shell/logs/9f59665dfbfa4468d8565f690237377cb7f6c9497166271b09d26fa34ea97d1f.json","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/installed_packages.log","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/latest-build.txt","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/logs/scripts.log","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/filesystem.info","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/docker.info","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/nethealth.txt","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/shell/logs/d387558411bef750ec641ba7f13a523c0197a0bec458d353e0c7a778cfbb5951.json","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/README.md","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/shell/logs/scripts.log","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/master-metrics.log","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/push-release/logs/scripts.log","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/containers.log","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/prtest-7d230a7-201-ig-n-jsvv.metrics","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/prtest-7d230a7-201-ig-m-khm7.metrics","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/node-metrics.log","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/shell/logs/32be3508789e6074128eb701067480aca262e91d12ec421358d9fd13b4625bae.json","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/pid1.journal","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/e2e.log","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/generated/dmesg.log","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/prtest-7d230a7-201-ig-n-jn9j.metrics","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/journals/docker.service","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/prtest-7d230a7-201-ig-n-bw59.metrics","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/build-log.txt","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/scripts/conformance-k8s/artifacts/junit_01.xml","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","dest":"logs/test_branch_origin_extended_conformance_k8s_39/201/artifacts/artifacts/master.metrics","level":"info","msg":"Finished upload","time":"2018-07-09T14:49:33Z"}
{"component":"gcsupload","level":"info","msg":"Finished upload to GCS","time":"2018-07-09T14:49:33Z"}
+ exit 0
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: PUSH THE ARTIFACTS AND METADATA [00h 00m 08s] ##########
[workspace] $ /bin/bash /tmp/jenkins7251369815457012108.sh
########## STARTING STAGE: GATHER ARTIFACTS FROM TEST CLUSTER ##########
+ [[ -s /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ export PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ base_artifact_dir=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts
+ source ./INSTANCE_PREFIX
++ INSTANCE_PREFIX=prtest-7d230a7-201
++ OS_TAG=71543b2
++ OS_PUSH_BASE_REPO=ci-pr-images/prtest-7d230a7-201-
++ gcloud compute instances list --regexp '.*prtest-7d230a7-201.*' --uri
+ for instance in '$( gcloud compute instances list --regexp ".*${INSTANCE_PREFIX}.*" --uri )'
++ mktemp
+ info=/tmp/tmp.bOJNfvh4OP
+ gcloud compute instances describe https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 --format json
++ jq .name --raw-output /tmp/tmp.bOJNfvh4OP
++ tail -c 5
+ name=khm7
+ jq '.tags.items | contains(["ocp-master"])' --exit-status /tmp/tmp.bOJNfvh4OP
true
+ artifact_dir=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/masters/khm7
+ mkdir -p /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/masters/khm7 /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/masters/khm7/generated /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/masters/khm7/journals
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo journalctl --unit origin-master.service --no-pager --all --lines=all
Warning: Permanently added 'compute.761873934373018687' (ECDSA) to the list of known hosts.
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo journalctl --unit origin-master-api.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo journalctl --unit origin-master-controllers.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo journalctl --unit etcd.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- oc get --raw /metrics --config=/etc/origin/master/admin.kubeconfig
error: Error loading config file "/etc/origin/master/admin.kubeconfig": open /etc/origin/master/admin.kubeconfig: permission denied
+ true
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo journalctl --unit origin-node.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo journalctl --unit openvswitch.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo journalctl --unit ovs-vswitchd.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo journalctl --unit ovsdb-server.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo journalctl --unit docker.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
++ uname --nodename
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- oc get --raw /metrics --server=https://ip-172-18-8-64.ec2.internal:10250
Unable to connect to the server: dial tcp: lookup ip-172-18-8-64.ec2.internal on 10.142.0.2:53: no such host
+ true
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a'
  WARNING: You're not using the default seccomp profile
WARNING: bridge-nf-call-iptables is disabled
WARNING: bridge-nf-call-ip6tables is disabled
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo yum history info origin
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- 'sudo df -h && sudo pvs && sudo vgs && sudo lvs'
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo yum list installed
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC
<no matches>
+ true
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- sudo journalctl _PID=1 --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-m-khm7 -- 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1'
+ for instance in '$( gcloud compute instances list --regexp ".*${INSTANCE_PREFIX}.*" --uri )'
++ mktemp
+ info=/tmp/tmp.r6oc1LYkkt
+ gcloud compute instances describe https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 --format json
++ jq .name --raw-output /tmp/tmp.r6oc1LYkkt
++ tail -c 5
+ name=bw59
+ jq '.tags.items | contains(["ocp-master"])' --exit-status /tmp/tmp.r6oc1LYkkt
false
+ jq '.tags.items | contains(["ocp-node"])' --exit-status /tmp/tmp.r6oc1LYkkt
true
+ artifact_dir=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/nodes/bw59
+ mkdir -p /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/nodes/bw59 /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/nodes/bw59/generated /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/nodes/bw59/journals
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- sudo journalctl --unit origin-node.service --no-pager --all --lines=all
Warning: Permanently added 'compute.9211230127901301822' (ECDSA) to the list of known hosts.
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- sudo journalctl --unit openvswitch.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- sudo journalctl --unit ovs-vswitchd.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- sudo journalctl --unit ovsdb-server.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- sudo journalctl --unit docker.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
++ uname --nodename
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- oc get --raw /metrics --server=https://ip-172-18-8-64.ec2.internal:10250
Unable to connect to the server: dial tcp: lookup ip-172-18-8-64.ec2.internal on 10.142.0.3:53: no such host
+ true
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a'
  WARNING: You're not using the default seccomp profile
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- sudo yum history info origin
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- 'sudo df -h && sudo pvs && sudo vgs && sudo lvs'
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- sudo yum list installed
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC
<no matches>
+ true
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- sudo journalctl _PID=1 --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-bw59 -- 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1'
+ for instance in '$( gcloud compute instances list --regexp ".*${INSTANCE_PREFIX}.*" --uri )'
++ mktemp
+ info=/tmp/tmp.F16sCD3Lzs
+ gcloud compute instances describe https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j --format json
++ jq .name --raw-output /tmp/tmp.F16sCD3Lzs
++ tail -c 5
+ name=jn9j
+ jq '.tags.items | contains(["ocp-master"])' --exit-status /tmp/tmp.F16sCD3Lzs
false
+ jq '.tags.items | contains(["ocp-node"])' --exit-status /tmp/tmp.F16sCD3Lzs
true
+ artifact_dir=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/nodes/jn9j
+ mkdir -p /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/nodes/jn9j /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/nodes/jn9j/generated /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/nodes/jn9j/journals
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- sudo journalctl --unit origin-node.service --no-pager --all --lines=all
Warning: Permanently added 'compute.5171888208638565438' (ECDSA) to the list of known hosts.
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- sudo journalctl --unit openvswitch.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- sudo journalctl --unit ovs-vswitchd.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- sudo journalctl --unit ovsdb-server.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- sudo journalctl --unit docker.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
++ uname --nodename
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- oc get --raw /metrics --server=https://ip-172-18-8-64.ec2.internal:10250
Unable to connect to the server: dial tcp: lookup ip-172-18-8-64.ec2.internal on 10.142.0.4:53: no such host
+ true
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a'
  WARNING: You're not using the default seccomp profile
WARNING: bridge-nf-call-iptables is disabled
WARNING: bridge-nf-call-ip6tables is disabled
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- sudo yum history info origin
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- 'sudo df -h && sudo pvs && sudo vgs && sudo lvs'
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- sudo yum list installed
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC
<no matches>
+ true
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- sudo journalctl _PID=1 --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jn9j -- 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1'
+ for instance in '$( gcloud compute instances list --regexp ".*${INSTANCE_PREFIX}.*" --uri )'
++ mktemp
+ info=/tmp/tmp.yzhtFF56S8
+ gcloud compute instances describe https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv --format json
++ jq .name --raw-output /tmp/tmp.yzhtFF56S8
++ tail -c 5
+ name=jsvv
+ jq '.tags.items | contains(["ocp-master"])' --exit-status /tmp/tmp.yzhtFF56S8
false
+ jq '.tags.items | contains(["ocp-node"])' --exit-status /tmp/tmp.yzhtFF56S8
true
+ artifact_dir=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/nodes/jsvv
+ mkdir -p /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/nodes/jsvv /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/nodes/jsvv/generated /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/artifacts/nodes/jsvv/journals
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- sudo journalctl --unit origin-node.service --no-pager --all --lines=all
Warning: Permanently added 'compute.7329302581236128830' (ECDSA) to the list of known hosts.
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- sudo journalctl --unit openvswitch.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- sudo journalctl --unit ovs-vswitchd.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- sudo journalctl --unit ovsdb-server.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- sudo journalctl --unit docker.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
++ uname --nodename
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- oc get --raw /metrics --server=https://ip-172-18-8-64.ec2.internal:10250
Unable to connect to the server: dial tcp: lookup ip-172-18-8-64.ec2.internal on 10.142.0.5:53: no such host
+ true
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a'
  WARNING: You're not using the default seccomp profile
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- sudo yum history info origin
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- 'sudo df -h && sudo pvs && sudo vgs && sudo lvs'
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- sudo yum list installed
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC
<no matches>
+ true
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- sudo journalctl _PID=1 --no-pager --all --lines=all
+ gcloud compute ssh https://www.googleapis.com/compute/v1/projects/openshift-gce-devel-ci/zones/us-east1-c/instances/prtest-7d230a7-201-ig-n-jsvv -- 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1'
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins2286516106014827161.sh
########## STARTING STAGE: DEPROVISION TEST CLUSTER ##########
+ [[ -s /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ export PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
++ mktemp
+ script=/tmp/tmp.U5FdBK2XwU
+ cat
+ chmod +x /tmp/tmp.U5FdBK2XwU
+ scp -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.U5FdBK2XwU openshiftdevel:/tmp/tmp.U5FdBK2XwU
+ ssh -F /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 900 /tmp/tmp.U5FdBK2XwU"'
+ cd /data/src/github.com/openshift/release
+ trap 'exit 0' EXIT
+ cd cluster/test-deploy/gcp
+ ../../bin/ansible.sh ansible-playbook playbooks/gcp/openshift-cluster/deprovision.yml
Activated service account credentials for: [jenkins-ci-provisioner@openshift-gce-devel.iam.gserviceaccount.com]

PLAY [Terminate running cluster and remove all supporting resources in GCE] ****

TASK [Gathering Facts] *********************************************************
Monday 09 July 2018  14:51:17 +0000 (0:00:00.065)       0:00:00.065 *********** 
ok: [localhost]

TASK [include_role] ************************************************************
Monday 09 July 2018  14:51:23 +0000 (0:00:05.993)       0:00:06.059 *********** 

TASK [openshift_gcp : Templatize DNS script] ***********************************
Monday 09 July 2018  14:51:23 +0000 (0:00:00.107)       0:00:06.166 *********** 
changed: [localhost]

TASK [openshift_gcp : Templatize provision script] *****************************
Monday 09 July 2018  14:51:24 +0000 (0:00:00.522)       0:00:06.689 *********** 
changed: [localhost]

TASK [openshift_gcp : Templatize de-provision script] **************************
Monday 09 July 2018  14:51:24 +0000 (0:00:00.332)       0:00:07.021 *********** 
changed: [localhost]

TASK [openshift_gcp : Provision GCP DNS domain] ********************************
Monday 09 July 2018  14:51:25 +0000 (0:00:00.311)       0:00:07.333 *********** 
skipping: [localhost]

TASK [openshift_gcp : Ensure that DNS resolves to the hosted zone] *************
Monday 09 July 2018  14:51:25 +0000 (0:00:00.024)       0:00:07.358 *********** 
skipping: [localhost]

TASK [openshift_gcp : Templatize SSH key provision script] *********************
Monday 09 July 2018  14:51:25 +0000 (0:00:00.023)       0:00:07.381 *********** 
changed: [localhost]

TASK [openshift_gcp : Provision GCP SSH key resources] *************************
Monday 09 July 2018  14:51:25 +0000 (0:00:00.275)       0:00:07.657 *********** 
skipping: [localhost]

TASK [openshift_gcp : Provision GCP resources] *********************************
Monday 09 July 2018  14:51:25 +0000 (0:00:00.024)       0:00:07.682 *********** 
skipping: [localhost]

TASK [openshift_gcp : De-provision GCP resources] ******************************
Monday 09 July 2018  14:51:25 +0000 (0:00:00.024)       0:00:07.707 *********** 
changed: [localhost]

PLAY RECAP *********************************************************************
localhost                  : ok=6    changed=5    unreachable=0    failed=0   

Monday 09 July 2018  14:55:49 +0000 (0:04:24.491)       0:04:32.198 *********** 
=============================================================================== 
openshift_gcp : De-provision GCP resources ---------------------------- 264.49s
Gathering Facts --------------------------------------------------------- 5.99s
openshift_gcp : Templatize DNS script ----------------------------------- 0.52s
openshift_gcp : Templatize provision script ----------------------------- 0.33s
openshift_gcp : Templatize de-provision script -------------------------- 0.31s
openshift_gcp : Templatize SSH key provision script --------------------- 0.28s
include_role ------------------------------------------------------------ 0.11s
openshift_gcp : Provision GCP resources --------------------------------- 0.03s
openshift_gcp : Provision GCP DNS domain -------------------------------- 0.02s
openshift_gcp : Provision GCP SSH key resources ------------------------- 0.02s
openshift_gcp : Ensure that DNS resolves to the hosted zone ------------- 0.02s
+ exit 0
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPROVISION TEST CLUSTER [00h 04m 39s] ##########
[workspace] $ /bin/bash /tmp/jenkins6459354874734353496.sh
########## STARTING STAGE: DELETE PR IMAGES ##########
+ [[ -s /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ export PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
+ trap 'exit 0' EXIT
+ source ./INSTANCE_PREFIX
++ INSTANCE_PREFIX=prtest-7d230a7-201
++ OS_TAG=71543b2
++ OS_PUSH_BASE_REPO=ci-pr-images/prtest-7d230a7-201-
+ export KUBECONFIG=/var/lib/jenkins/secrets/image-pr-push.kubeconfig
+ KUBECONFIG=/var/lib/jenkins/secrets/image-pr-push.kubeconfig
+ oc get is -o name -n ci-pr-images
+ grep prtest-7d230a7-201
+ xargs -r oc delete
imagestream "prtest-7d230a7-201-hello-openshift" deleted
imagestream "prtest-7d230a7-201-node" deleted
imagestream "prtest-7d230a7-201-openvswitch" deleted
imagestream "prtest-7d230a7-201-origin" deleted
imagestream "prtest-7d230a7-201-origin-base" deleted
imagestream "prtest-7d230a7-201-origin-cluster-capacity" deleted
imagestream "prtest-7d230a7-201-origin-deployer" deleted
imagestream "prtest-7d230a7-201-origin-docker-builder" deleted
imagestream "prtest-7d230a7-201-origin-docker-registry" deleted
imagestream "prtest-7d230a7-201-origin-egress-http-proxy" deleted
imagestream "prtest-7d230a7-201-origin-egress-router" deleted
imagestream "prtest-7d230a7-201-origin-f5-router" deleted
imagestream "prtest-7d230a7-201-origin-haproxy-router" deleted
imagestream "prtest-7d230a7-201-origin-keepalived-ipfailover" deleted
imagestream "prtest-7d230a7-201-origin-metrics-server" deleted
imagestream "prtest-7d230a7-201-origin-pod" deleted
imagestream "prtest-7d230a7-201-origin-recycler" deleted
imagestream "prtest-7d230a7-201-origin-service-catalog" deleted
imagestream "prtest-7d230a7-201-origin-sti-builder" deleted
imagestream "prtest-7d230a7-201-origin-template-service-broker" deleted
imagestream "prtest-7d230a7-201-origin-web-console" deleted
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins4916148129305265841.sh
########## STARTING STAGE: DEPROVISION CLOUD RESOURCES ##########
+ [[ -s /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e
++ export PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config
+ oct deprovision

PLAYBOOK: main.yml *************************************************************
4 plays in /var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml

PLAY [ensure we have the parameters necessary to deprovision virtual hosts] ****

TASK [ensure all required variables are set] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:9
skipping: [localhost] => (item=origin_ci_inventory_dir)  => {
    "changed": false, 
    "generated_timestamp": "2018-07-09 10:55:55.209240", 
    "item": "origin_ci_inventory_dir", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_region)  => {
    "changed": false, 
    "generated_timestamp": "2018-07-09 10:55:55.212317", 
    "item": "origin_ci_aws_region", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

PLAY [deprovision virtual hosts in EC2] ****************************************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

TASK [deprovision a virtual EC2 host] ******************************************
task path: /var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:28
included: /var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml for localhost

TASK [update the SSH configuration to remove AWS EC2 specifics] ****************
task path: /var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:2
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-07-09 10:55:55.970855", 
    "msg": ""
}

TASK [rename EC2 instance for termination reaper] ******************************
task path: /var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:8
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-07-09 10:55:56.513303", 
    "msg": "Tags {'Name': 'oct-terminate'} created for resource i-02f629264f06d6c91."
}

TASK [tear down the EC2 instance] **********************************************
task path: /var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:15
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-07-09 10:55:57.481050", 
    "instance_ids": [
        "i-02f629264f06d6c91"
    ], 
    "instances": [
        {
            "ami_launch_index": "0", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-0fdcb611c7ef7f0bd"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-0c54e16ef79083ff4"
                }
            }, 
            "dns_name": "ec2-34-207-171-153.compute-1.amazonaws.com", 
            "ebs_optimized": false, 
            "groups": {
                "sg-7e73221a": "default"
            }, 
            "hypervisor": "xen", 
            "id": "i-02f629264f06d6c91", 
            "image_id": "ami-0b77b87a37c3e662c", 
            "instance_type": "m4.xlarge", 
            "kernel": null, 
            "key_name": "libra", 
            "launch_time": "2018-07-09T12:42:13.000Z", 
            "placement": "us-east-1d", 
            "private_dns_name": "ip-172-18-8-178.ec2.internal", 
            "private_ip": "172.18.8.178", 
            "public_dns_name": "ec2-34-207-171-153.compute-1.amazonaws.com", 
            "public_ip": "34.207.171.153", 
            "ramdisk": null, 
            "region": "us-east-1", 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "running", 
            "state_code": 16, 
            "tags": {
                "Name": "oct-terminate", 
                "openshift_etcd": "", 
                "openshift_master": "", 
                "openshift_node": ""
            }, 
            "tenancy": "default", 
            "virtualization_type": "hvm"
        }
    ], 
    "tagged_instances": []
}

TASK [remove the serialized host variables] ************************************
task path: /var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:22
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-07-09 10:55:57.721528", 
    "path": "/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory/host_vars/172.18.8.178.yml", 
    "state": "absent"
}

PLAY [deprovision virtual hosts locally manged by Vagrant] *********************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

PLAY [clean up local configuration for deprovisioned instances] ****************

TASK [remove inventory configuration directory] ********************************
task path: /var/lib/jenkins/origin-ci-tool/4b405957477ba1b70cfacd1cf43c6d41a605fc8e/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:61
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-07-09 10:55:58.162570", 
    "path": "/var/lib/jenkins/jobs/test_branch_origin_extended_conformance_k8s_39/workspace/.config/origin-ci-tool/inventory", 
    "state": "absent"
}

PLAY RECAP *********************************************************************
localhost                  : ok=8    changed=4    unreachable=0    failed=0   

+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPROVISION CLOUD RESOURCES [00h 00m 04s] ##########
Archiving artifacts
Recording test results
[WS-CLEANUP] Deleting project workspace...[WS-CLEANUP] done
Finished: SUCCESS