SuccessConsole Output

Skipping 5,726 KB.. Full Log
    ], 
    "stdout": []
}

TASK [openshift_examples : Remove old xPaas template files] ********************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:109
ok: [localhost] => (item=/etc/origin/examples/xpaas-templates/sso70-basic.json) => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:08.680686", 
    "item": "/etc/origin/examples/xpaas-templates/sso70-basic.json", 
    "path": "/etc/origin/examples/xpaas-templates/sso70-basic.json", 
    "state": "absent"
}

TASK [openshift_examples : Remove old xPaas templates from openshift namespace] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:116
ok: [localhost] => (item=sso70-basic) => {
    "changed": false, 
    "cmd": [
        "/usr/local/bin/oc", 
        "--config=/etc/origin/master/admin.kubeconfig", 
        "-n", 
        "openshift", 
        "delete", 
        "templates/sso70-basic"
    ], 
    "delta": "0:00:00.237361", 
    "end": "2018-02-23 07:19:09.095940", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:09.121637", 
    "item": "sso70-basic", 
    "rc": 1, 
    "start": "2018-02-23 07:19:08.858579", 
    "stderr": [
        "Error from server (NotFound): templates.template.openshift.io \"sso70-basic\" not found"
    ], 
    "stdout": []
}

TASK [openshift_examples : Import xPaas image streams] *************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:124
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:09.174346", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_examples : Import xPaas templates] *****************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:132
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:09.222511", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] **************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/main.yml:2
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:09.287994", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Install firewalld packages] ********************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:2
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:09.334940", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Ensure iptables services are not enabled] ******************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:7
skipping: [localhost] => (item=iptables)  => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:09.384518", 
    "item": "iptables", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}
skipping: [localhost] => (item=ip6tables)  => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:09.397273", 
    "item": "ip6tables", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Wait 10 seconds after disabling iptables] ******************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:19
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:09.443854", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Start and enable firewalld service] ************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:24
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:09.493938", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] ***
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:33
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:09.540177", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Restart polkitd] *******************************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:37
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:09.589138", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Wait for polkit action to have been created] ***************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:44
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:09.638432", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Add firewalld allow rules] *********************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:53

TASK [os_firewall : Remove firewalld allow rules] ******************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:61

TASK [os_firewall : Ensure firewalld service is not enabled] *******************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:3
ok: [localhost] => {
    "changed": false, 
    "enabled": false, 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:09.930771", 
    "name": "firewalld", 
    "state": "stopped", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "yes", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "firewalld.service", 
        "DevicePolicy": "auto", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/dev/null", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "firewalld.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "65536", 
        "LimitNPROC": "63327", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "63327", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "masked", 
        "MainPID": "0", 
        "MemoryAccounting": "yes", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "firewalld.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "inherit", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "UMask": "0022", 
        "UnitFileState": "bad", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [os_firewall : Wait 10 seconds after disabling firewalld] *****************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:12
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:09.980710", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Install iptables packages] *********************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:17
ok: [localhost] => (item=iptables) => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:10.707815", 
    "item": "iptables", 
    "msg": "", 
    "rc": 0, 
    "results": [
        "iptables-1.4.21-18.2.el7_4.x86_64 providing iptables is already installed"
    ]
}
ok: [localhost] => (item=iptables-services) => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:11.397838", 
    "item": "iptables-services", 
    "msg": "", 
    "rc": 0, 
    "results": [
        "iptables-services-1.4.21-18.2.el7_4.x86_64 providing iptables-services is already installed"
    ]
}

TASK [os_firewall : Start and enable iptables service] *************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:24
ok: [localhost] => {
    "changed": false, 
    "enabled": true, 
    "generated_timestamp": "2018-02-23 07:19:11.676290", 
    "name": "iptables", 
    "state": "started", 
    "status": {
        "ActiveEnterTimestamp": "Fri 2018-02-23 06:58:37 UTC", 
        "ActiveEnterTimestampMonotonic": "2517306143", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "active", 
        "After": "basic.target systemd-journald.socket system.slice syslog.target", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "yes", 
        "AssertTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "AssertTimestampMonotonic": "2517257510", 
        "Before": "ip6tables.service docker.service shutdown.target network.service", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "yes", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "yes", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "ConditionResult": "yes", 
        "ConditionTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "ConditionTimestampMonotonic": "2517257508", 
        "Conflicts": "shutdown.target", 
        "ConsistsOf": "docker.service", 
        "ControlGroup": "/system.slice/iptables.service", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "IPv4 firewall with iptables", 
        "DevicePolicy": "auto", 
        "Environment": "BOOTUP=serial CONSOLETYPE=serial", 
        "ExecMainCode": "1", 
        "ExecMainExitTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "ExecMainExitTimestampMonotonic": "2517300535", 
        "ExecMainPID": "8008", 
        "ExecMainStartTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "ExecMainStartTimestampMonotonic": "2517260742", 
        "ExecMainStatus": "0", 
        "ExecReload": "{ path=/usr/libexec/iptables/iptables.init ; argv[]=/usr/libexec/iptables/iptables.init reload ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStart": "{ path=/usr/libexec/iptables/iptables.init ; argv[]=/usr/libexec/iptables/iptables.init start ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStop": "{ path=/usr/libexec/iptables/iptables.init ; argv[]=/usr/libexec/iptables/iptables.init stop ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/usr/lib/systemd/system/iptables.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "iptables.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "InactiveExitTimestampMonotonic": "2517260772", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "63327", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "63327", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "yes", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "iptables.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "yes", 
        "Requires": "basic.target", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system.slice", 
        "StandardError": "syslog", 
        "StandardInput": "null", 
        "StandardOutput": "syslog", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "exited", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "0", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "oneshot", 
        "UMask": "0022", 
        "UnitFilePreset": "disabled", 
        "UnitFileState": "enabled", 
        "WantedBy": "docker.service basic.target", 
        "Wants": "system.slice", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] ***
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:33
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:11.726285", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Add iptables allow rules] **********************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:37

TASK [os_firewall : Remove iptables rules] *************************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:45

TASK [openshift_hosted_templates : Create local temp dir for OpenShift hosted templates copy] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:2
changed: [localhost -> localhost] => {
    "changed": true, 
    "cmd": [
        "mktemp", 
        "-d", 
        "/tmp/openshift-ansible-XXXXXXX"
    ], 
    "delta": "0:00:01.003256", 
    "end": "2018-02-23 07:19:13.009839", 
    "generated_timestamp": "2018-02-23 07:19:13.026808", 
    "rc": 0, 
    "start": "2018-02-23 07:19:12.006583", 
    "stderr": [], 
    "stdout": [
        "/tmp/openshift-ansible-CxxsNn5"
    ]
}

TASK [openshift_hosted_templates : Create tar of OpenShift examples] ***********
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:10
changed: [localhost -> localhost] => {
    "changed": true, 
    "cmd": [
        "tar", 
        "-C", 
        "/usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/files/v3.6/origin", 
        "-cvf", 
        "/tmp/openshift-ansible-CxxsNn5/openshift-hosted-templates.tar", 
        "."
    ], 
    "delta": "0:00:00.003767", 
    "end": "2018-02-23 07:19:13.222983", 
    "generated_timestamp": "2018-02-23 07:19:13.239801", 
    "rc": 0, 
    "start": "2018-02-23 07:19:13.219216", 
    "stderr": [], 
    "stdout": [
        "./", 
        "./registry-console.yaml"
    ]
}

TASK [openshift_hosted_templates : Create remote OpenShift hosted templates directory] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:19
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-02-23 07:19:13.435824", 
    "gid": 0, 
    "group": "root", 
    "mode": "0755", 
    "owner": "root", 
    "path": "/etc/origin/hosted", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 35, 
    "state": "directory", 
    "uid": 0
}

TASK [openshift_hosted_templates : Unarchive the OpenShift hosted templates on the remote] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:25
changed: [localhost] => {
    "changed": true, 
    "dest": "/etc/origin/hosted/", 
    "extract_results": {
        "cmd": [
            "/bin/gtar", 
            "--extract", 
            "-C", 
            "/etc/origin/hosted/", 
            "-f", 
            "/home/origin/.ansible/tmp/ansible-tmp-1519370353.48-47815716738299/source"
        ], 
        "err": "", 
        "out": "", 
        "rc": 0
    }, 
    "generated_timestamp": "2018-02-23 07:19:13.988534", 
    "gid": 0, 
    "group": "root", 
    "handler": "TarArchive", 
    "mode": "02755", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 35, 
    "src": "/home/origin/.ansible/tmp/ansible-tmp-1519370353.48-47815716738299/source", 
    "state": "directory", 
    "uid": 0
}

TASK [openshift_hosted_templates : Cleanup the OpenShift hosted templates temp dir] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:30
changed: [localhost -> localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-02-23 07:19:14.194482", 
    "path": "/tmp/openshift-ansible-CxxsNn5", 
    "state": "absent"
}

TASK [openshift_hosted_templates : Modify registry paths if registry_url is not registry.access.redhat.com] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:34
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:14.241243", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_hosted_templates : Create temp directory for kubeconfig] *******
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:39
ok: [localhost] => {
    "changed": false, 
    "cmd": [
        "mktemp", 
        "-d", 
        "/tmp/openshift-ansible-XXXXXX"
    ], 
    "delta": "0:00:01.003587", 
    "end": "2018-02-23 07:19:15.425621", 
    "generated_timestamp": "2018-02-23 07:19:15.445269", 
    "rc": 0, 
    "start": "2018-02-23 07:19:14.422034", 
    "stderr": [], 
    "stdout": [
        "/tmp/openshift-ansible-5zQbYR"
    ]
}

TASK [openshift_hosted_templates : Record kubeconfig tmp dir] ******************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:44
ok: [localhost] => {
    "ansible_facts": {
        "openshift_hosted_templates_kubeconfig": "/tmp/openshift-ansible-5zQbYR/admin.kubeconfig"
    }, 
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:15.492213"
}

TASK [openshift_hosted_templates : Copy the admin client config(s)] ************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:48
ok: [localhost] => {
    "changed": false, 
    "cmd": [
        "cp", 
        "/etc/origin/master/admin.kubeconfig", 
        "/tmp/openshift-ansible-5zQbYR/admin.kubeconfig"
    ], 
    "delta": "0:00:00.003219", 
    "end": "2018-02-23 07:19:15.665920", 
    "generated_timestamp": "2018-02-23 07:19:15.682362", 
    "rc": 0, 
    "start": "2018-02-23 07:19:15.662701", 
    "stderr": [], 
    "stdout": []
}

TASK [openshift_hosted_templates : Create or update hosted templates] **********
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:53
ok: [localhost] => {
    "changed": false, 
    "cmd": [
        "/usr/local/bin/oc", 
        "create", 
        "-f", 
        "/etc/origin/hosted", 
        "--config=/tmp/openshift-ansible-5zQbYR/admin.kubeconfig", 
        "-n", 
        "openshift"
    ], 
    "delta": "0:00:00.236531", 
    "end": "2018-02-23 07:19:16.094097", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:16.120146", 
    "rc": 1, 
    "start": "2018-02-23 07:19:15.857566", 
    "stderr": [
        "Error from server (AlreadyExists): error when creating \"/etc/origin/hosted/registry-console.yaml\": templates.template.openshift.io \"registry-console\" already exists"
    ], 
    "stdout": []
}

TASK [openshift_hosted_templates : Delete temp directory] **********************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:63
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:16.310584", 
    "path": "/tmp/openshift-ansible-5zQbYR", 
    "state": "absent"
}

TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] **************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/main.yml:2
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:16.388004", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Install firewalld packages] ********************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:2
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:16.469930", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Ensure iptables services are not enabled] ******************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:7
skipping: [localhost] => (item=iptables)  => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:16.554781", 
    "item": "iptables", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}
skipping: [localhost] => (item=ip6tables)  => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:16.599224", 
    "item": "ip6tables", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Wait 10 seconds after disabling iptables] ******************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:19
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:16.679398", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Start and enable firewalld service] ************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:24
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:16.757037", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] ***
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:33
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:16.836954", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Restart polkitd] *******************************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:37
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:16.915585", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Wait for polkit action to have been created] ***************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:44
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:16.998575", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Add firewalld allow rules] *********************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:53

TASK [os_firewall : Remove firewalld allow rules] ******************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:61

TASK [os_firewall : Ensure firewalld service is not enabled] *******************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:3
ok: [localhost] => {
    "changed": false, 
    "enabled": false, 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:17.331455", 
    "name": "firewalld", 
    "state": "stopped", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "yes", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "firewalld.service", 
        "DevicePolicy": "auto", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/dev/null", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "firewalld.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "65536", 
        "LimitNPROC": "63327", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "63327", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "masked", 
        "MainPID": "0", 
        "MemoryAccounting": "yes", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "firewalld.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "inherit", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "UMask": "0022", 
        "UnitFileState": "bad", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [os_firewall : Wait 10 seconds after disabling firewalld] *****************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:12
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:17.415187", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Install iptables packages] *********************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:17
ok: [localhost] => (item=iptables) => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:18.234596", 
    "item": "iptables", 
    "msg": "", 
    "rc": 0, 
    "results": [
        "iptables-1.4.21-18.2.el7_4.x86_64 providing iptables is already installed"
    ]
}
ok: [localhost] => (item=iptables-services) => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:19.033246", 
    "item": "iptables-services", 
    "msg": "", 
    "rc": 0, 
    "results": [
        "iptables-services-1.4.21-18.2.el7_4.x86_64 providing iptables-services is already installed"
    ]
}

TASK [os_firewall : Start and enable iptables service] *************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:24
ok: [localhost] => {
    "changed": false, 
    "enabled": true, 
    "generated_timestamp": "2018-02-23 07:19:19.356851", 
    "name": "iptables", 
    "state": "started", 
    "status": {
        "ActiveEnterTimestamp": "Fri 2018-02-23 06:58:37 UTC", 
        "ActiveEnterTimestampMonotonic": "2517306143", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "active", 
        "After": "basic.target system.slice systemd-journald.socket syslog.target", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "yes", 
        "AssertTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "AssertTimestampMonotonic": "2517257510", 
        "Before": "shutdown.target ip6tables.service docker.service network.service", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "yes", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "yes", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "ConditionResult": "yes", 
        "ConditionTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "ConditionTimestampMonotonic": "2517257508", 
        "Conflicts": "shutdown.target", 
        "ConsistsOf": "docker.service", 
        "ControlGroup": "/system.slice/iptables.service", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "IPv4 firewall with iptables", 
        "DevicePolicy": "auto", 
        "Environment": "BOOTUP=serial CONSOLETYPE=serial", 
        "ExecMainCode": "1", 
        "ExecMainExitTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "ExecMainExitTimestampMonotonic": "2517300535", 
        "ExecMainPID": "8008", 
        "ExecMainStartTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "ExecMainStartTimestampMonotonic": "2517260742", 
        "ExecMainStatus": "0", 
        "ExecReload": "{ path=/usr/libexec/iptables/iptables.init ; argv[]=/usr/libexec/iptables/iptables.init reload ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStart": "{ path=/usr/libexec/iptables/iptables.init ; argv[]=/usr/libexec/iptables/iptables.init start ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStop": "{ path=/usr/libexec/iptables/iptables.init ; argv[]=/usr/libexec/iptables/iptables.init stop ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/usr/lib/systemd/system/iptables.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "iptables.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "InactiveExitTimestampMonotonic": "2517260772", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "63327", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "63327", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "yes", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "iptables.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "yes", 
        "Requires": "basic.target", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system.slice", 
        "StandardError": "syslog", 
        "StandardInput": "null", 
        "StandardOutput": "syslog", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "exited", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "0", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "oneshot", 
        "UMask": "0022", 
        "UnitFilePreset": "disabled", 
        "UnitFileState": "enabled", 
        "WantedBy": "basic.target docker.service", 
        "Wants": "system.slice", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] ***
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:33
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:19.441208", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Add iptables allow rules] **********************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:37

TASK [os_firewall : Remove iptables rules] *************************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:45

TASK [openshift_examples : Create local temp dir for OpenShift examples copy] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:14
changed: [localhost -> localhost] => {
    "changed": true, 
    "cmd": [
        "mktemp", 
        "-d", 
        "/tmp/openshift-ansible-XXXXXXX"
    ], 
    "delta": "0:00:00.002722", 
    "end": "2018-02-23 07:19:19.734335", 
    "generated_timestamp": "2018-02-23 07:19:19.750549", 
    "rc": 0, 
    "start": "2018-02-23 07:19:19.731613", 
    "stderr": [], 
    "stdout": [
        "/tmp/openshift-ansible-UnMI7FB"
    ]
}

TASK [openshift_examples : Create tar of OpenShift examples] *******************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:20
changed: [localhost -> localhost] => {
    "changed": true, 
    "cmd": [
        "tar", 
        "-C", 
        "/usr/share/ansible/openshift-ansible/roles/openshift_examples/files/examples/v3.6/", 
        "-cvf", 
        "/tmp/openshift-ansible-UnMI7FB/openshift-examples.tar", 
        "."
    ], 
    "delta": "0:00:00.010494", 
    "end": "2018-02-23 07:19:19.951773", 
    "generated_timestamp": "2018-02-23 07:19:19.976371", 
    "rc": 0, 
    "start": "2018-02-23 07:19:19.941279", 
    "stderr": [], 
    "stdout": [
        "./", 
        "./cfme-templates/", 
        "./cfme-templates/cfme-pv-db-example.yaml", 
        "./cfme-templates/cfme-pv-region-example.yaml", 
        "./cfme-templates/cfme-pv-server-example.yaml", 
        "./cfme-templates/cfme-template.yaml", 
        "./cfme-templates/jboss-middleware-manager-pv-example.yaml", 
        "./cfme-templates/jboss-middleware-manager-template.yaml", 
        "./db-templates/", 
        "./db-templates/OWNERS", 
        "./db-templates/README.md", 
        "./db-templates/mariadb-ephemeral-template.json", 
        "./db-templates/mariadb-persistent-template.json", 
        "./db-templates/mongodb-ephemeral-template.json", 
        "./db-templates/mongodb-persistent-template.json", 
        "./db-templates/mysql-ephemeral-template.json", 
        "./db-templates/mysql-persistent-template.json", 
        "./db-templates/postgresql-ephemeral-template.json", 
        "./db-templates/postgresql-persistent-template.json", 
        "./db-templates/redis-ephemeral-template.json", 
        "./db-templates/redis-persistent-template.json", 
        "./image-streams/", 
        "./image-streams/OWNERS", 
        "./image-streams/dotnet_imagestreams.json", 
        "./image-streams/image-streams-centos7.json", 
        "./image-streams/image-streams-rhel7.json", 
        "./quickstart-templates/", 
        "./quickstart-templates/OWNERS", 
        "./quickstart-templates/README.md", 
        "./quickstart-templates/amp.yml", 
        "./quickstart-templates/apicast-gateway-template.yml", 
        "./quickstart-templates/apicast.yml", 
        "./quickstart-templates/cakephp-mysql-persistent.json", 
        "./quickstart-templates/cakephp-mysql.json", 
        "./quickstart-templates/dancer-mysql-persistent.json", 
        "./quickstart-templates/dancer-mysql.json", 
        "./quickstart-templates/django-postgresql-persistent.json", 
        "./quickstart-templates/django-postgresql.json", 
        "./quickstart-templates/dotnet-example.json", 
        "./quickstart-templates/dotnet-pgsql-persistent.json", 
        "./quickstart-templates/dotnet-runtime-example.json", 
        "./quickstart-templates/httpd.json", 
        "./quickstart-templates/jenkins-ephemeral-template.json", 
        "./quickstart-templates/jenkins-persistent-template.json", 
        "./quickstart-templates/nodejs-mongodb-persistent.json", 
        "./quickstart-templates/nodejs-mongodb.json", 
        "./quickstart-templates/pvc.yml", 
        "./quickstart-templates/rails-postgresql-persistent.json", 
        "./quickstart-templates/rails-postgresql.json", 
        "./quickstart-templates/wildcard.yml", 
        "./xpaas-streams/", 
        "./xpaas-streams/fis-image-streams.json", 
        "./xpaas-streams/jboss-image-streams.json", 
        "./xpaas-templates/", 
        "./xpaas-templates/jws31-tomcat7-basic-s2i.json", 
        "./xpaas-templates/amq62-basic.json", 
        "./xpaas-templates/eap71-mongodb-persistent-s2i.json", 
        "./xpaas-templates/amq62-persistent-ssl.json", 
        "./xpaas-templates/eap71-mongodb-s2i.json", 
        "./xpaas-templates/amq62-persistent.json", 
        "./xpaas-templates/jws31-tomcat7-https-s2i.json", 
        "./xpaas-templates/amq62-ssl.json", 
        "./xpaas-templates/jws31-tomcat7-mysql-s2i.json", 
        "./xpaas-templates/amq63-basic.json", 
        "./xpaas-templates/eap71-mysql-persistent-s2i.json", 
        "./xpaas-templates/amq63-persistent-ssl.json", 
        "./xpaas-templates/eap71-mysql-s2i.json", 
        "./xpaas-templates/amq63-persistent.json", 
        "./xpaas-templates/jws31-tomcat8-basic-s2i.json", 
        "./xpaas-templates/amq63-ssl.json", 
        "./xpaas-templates/eap71-postgresql-persistent-s2i.json", 
        "./xpaas-templates/datagrid65-basic.json", 
        "./xpaas-templates/eap71-postgresql-s2i.json", 
        "./xpaas-templates/datagrid65-https.json", 
        "./xpaas-templates/eap64-postgresql-s2i.json", 
        "./xpaas-templates/datagrid65-mysql-persistent.json", 
        "./xpaas-templates/eap71-sso-s2i.json", 
        "./xpaas-templates/datagrid65-mysql.json", 
        "./xpaas-templates/eap64-mysql-persistent-s2i.json", 
        "./xpaas-templates/datagrid65-postgresql-persistent.json", 
        "./xpaas-templates/eap71-third-party-db-s2i.json", 
        "./xpaas-templates/datagrid65-postgresql.json", 
        "./xpaas-templates/jws30-tomcat7-basic-s2i.json", 
        "./xpaas-templates/datagrid71-basic.json", 
        "./xpaas-templates/jws30-tomcat7-https-s2i.json", 
        "./xpaas-templates/datagrid71-https.json", 
        "./xpaas-templates/eap64-sso-s2i.json", 
        "./xpaas-templates/datagrid71-mysql-persistent.json", 
        "./xpaas-templates/jws30-tomcat7-mongodb-s2i.json", 
        "./xpaas-templates/datagrid71-mysql.json", 
        "./xpaas-templates/eap64-mysql-s2i.json", 
        "./xpaas-templates/sso70-https.json", 
        "./xpaas-templates/datagrid71-postgresql-persistent.json", 
        "./xpaas-templates/jws30-tomcat7-mysql-s2i.json", 
        "./xpaas-templates/datagrid71-postgresql.json", 
        "./xpaas-templates/jws30-tomcat7-postgresql-s2i.json", 
        "./xpaas-templates/datavirt63-basic-s2i.json", 
        "./xpaas-templates/eap64-postgresql-persistent-s2i.json", 
        "./xpaas-templates/datavirt63-extensions-support-s2i.json", 
        "./xpaas-templates/jws30-tomcat8-basic-s2i.json", 
        "./xpaas-templates/datavirt63-secure-s2i.json", 
        "./xpaas-templates/eap64-third-party-db-s2i.json", 
        "./xpaas-templates/decisionserver62-amq-s2i.json", 
        "./xpaas-templates/eap70-amq-persistent-s2i.json", 
        "./xpaas-templates/decisionserver62-basic-s2i.json", 
        "./xpaas-templates/eap70-basic-s2i.json", 
        "./xpaas-templates/decisionserver62-https-s2i.json", 
        "./xpaas-templates/eap70-amq-s2i.json", 
        "./xpaas-templates/decisionserver63-amq-s2i.json", 
        "./xpaas-templates/eap70-https-s2i.json", 
        "./xpaas-templates/decisionserver63-basic-s2i.json", 
        "./xpaas-templates/eap70-mongodb-persistent-s2i.json", 
        "./xpaas-templates/decisionserver63-https-s2i.json", 
        "./xpaas-templates/eap70-mongodb-s2i.json", 
        "./xpaas-templates/decisionserver64-amq-s2i.json", 
        "./xpaas-templates/eap70-mysql-persistent-s2i.json", 
        "./xpaas-templates/decisionserver64-basic-s2i.json", 
        "./xpaas-templates/eap70-mysql-s2i.json", 
        "./xpaas-templates/decisionserver64-https-s2i.json", 
        "./xpaas-templates/eap71-amq-s2i.json", 
        "./xpaas-templates/eap64-amq-persistent-s2i.json", 
        "./xpaas-templates/jws31-tomcat8-https-s2i.json", 
        "./xpaas-templates/eap64-amq-s2i.json", 
        "./xpaas-templates/jws30-tomcat8-https-s2i.json", 
        "./xpaas-templates/eap64-basic-s2i.json", 
        "./xpaas-templates/jws30-tomcat8-mongodb-s2i.json", 
        "./xpaas-templates/eap64-https-s2i.json", 
        "./xpaas-templates/eap71-basic-s2i.json", 
        "./xpaas-templates/eap64-mongodb-persistent-s2i.json", 
        "./xpaas-templates/jws30-tomcat8-mysql-s2i.json", 
        "./xpaas-templates/eap64-mongodb-s2i.json", 
        "./xpaas-templates/eap71-amq-persistent-s2i.json", 
        "./xpaas-templates/eap70-postgresql-persistent-s2i.json", 
        "./xpaas-templates/jws30-tomcat8-postgresql-s2i.json", 
        "./xpaas-templates/eap70-postgresql-s2i.json", 
        "./xpaas-templates/jws31-tomcat8-mysql-s2i.json", 
        "./xpaas-templates/eap70-sso-s2i.json", 
        "./xpaas-templates/eap71-https-s2i.json", 
        "./xpaas-templates/eap70-third-party-db-s2i.json", 
        "./xpaas-templates/sso70-postgresql.json", 
        "./xpaas-templates/sso70-mysql.json", 
        "./xpaas-templates/jws30-tomcat7-mongodb-persistent-s2i.json", 
        "./xpaas-templates/jws30-tomcat7-mysql-persistent-s2i.json", 
        "./xpaas-templates/jws30-tomcat7-postgresql-persistent-s2i.json", 
        "./xpaas-templates/jws30-tomcat8-mongodb-persistent-s2i.json", 
        "./xpaas-templates/jws30-tomcat8-mysql-persistent-s2i.json", 
        "./xpaas-templates/jws30-tomcat8-postgresql-persistent-s2i.json", 
        "./xpaas-templates/jws31-tomcat7-mongodb-persistent-s2i.json", 
        "./xpaas-templates/jws31-tomcat7-mongodb-s2i.json", 
        "./xpaas-templates/jws31-tomcat7-mysql-persistent-s2i.json", 
        "./xpaas-templates/jws31-tomcat7-postgresql-persistent-s2i.json", 
        "./xpaas-templates/jws31-tomcat7-postgresql-s2i.json", 
        "./xpaas-templates/jws31-tomcat8-mongodb-persistent-s2i.json", 
        "./xpaas-templates/jws31-tomcat8-mongodb-s2i.json", 
        "./xpaas-templates/jws31-tomcat8-mysql-persistent-s2i.json", 
        "./xpaas-templates/jws31-tomcat8-postgresql-persistent-s2i.json", 
        "./xpaas-templates/jws31-tomcat8-postgresql-s2i.json", 
        "./xpaas-templates/karaf2-camel-amq-template.json", 
        "./xpaas-templates/karaf2-camel-log-template.json", 
        "./xpaas-templates/karaf2-camel-rest-sql-template.json", 
        "./xpaas-templates/karaf2-cxf-rest-template.json", 
        "./xpaas-templates/openjdk18-web-basic-s2i.json", 
        "./xpaas-templates/processserver63-amq-mysql-persistent-s2i.json", 
        "./xpaas-templates/processserver63-amq-mysql-s2i.json", 
        "./xpaas-templates/processserver63-amq-postgresql-persistent-s2i.json", 
        "./xpaas-templates/processserver63-amq-postgresql-s2i.json", 
        "./xpaas-templates/processserver63-basic-s2i.json", 
        "./xpaas-templates/processserver63-mysql-persistent-s2i.json", 
        "./xpaas-templates/processserver63-mysql-s2i.json", 
        "./xpaas-templates/processserver63-postgresql-persistent-s2i.json", 
        "./xpaas-templates/processserver63-postgresql-s2i.json", 
        "./xpaas-templates/processserver64-amq-mysql-persistent-s2i.json", 
        "./xpaas-templates/processserver64-amq-mysql-s2i.json", 
        "./xpaas-templates/processserver64-amq-postgresql-persistent-s2i.json", 
        "./xpaas-templates/processserver64-amq-postgresql-s2i.json", 
        "./xpaas-templates/processserver64-basic-s2i.json", 
        "./xpaas-templates/processserver64-mysql-persistent-s2i.json", 
        "./xpaas-templates/processserver64-mysql-s2i.json", 
        "./xpaas-templates/processserver64-postgresql-persistent-s2i.json", 
        "./xpaas-templates/processserver64-postgresql-s2i.json", 
        "./xpaas-templates/spring-boot-camel-amq-template.json", 
        "./xpaas-templates/spring-boot-camel-config-template.json", 
        "./xpaas-templates/spring-boot-camel-drools-template.json", 
        "./xpaas-templates/spring-boot-camel-infinispan-template.json", 
        "./xpaas-templates/spring-boot-camel-rest-sql-template.json", 
        "./xpaas-templates/spring-boot-camel-teiid-template.json", 
        "./xpaas-templates/spring-boot-camel-template.json", 
        "./xpaas-templates/spring-boot-camel-xml-template.json", 
        "./xpaas-templates/spring-boot-cxf-jaxrs-template.json", 
        "./xpaas-templates/spring-boot-cxf-jaxws-template.json", 
        "./xpaas-templates/sso70-mysql-persistent.json", 
        "./xpaas-templates/sso70-postgresql-persistent.json", 
        "./xpaas-templates/sso71-https.json", 
        "./xpaas-templates/sso71-mysql-persistent.json", 
        "./xpaas-templates/sso71-mysql.json", 
        "./xpaas-templates/sso71-postgresql-persistent.json", 
        "./xpaas-templates/sso71-postgresql.json"
    ]
}

TASK [openshift_examples : Create the remote OpenShift examples directory] *****
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:29
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-02-23 07:19:20.184232", 
    "gid": 0, 
    "group": "root", 
    "mode": "0755", 
    "owner": "root", 
    "path": "/etc/origin/examples", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 141, 
    "state": "directory", 
    "uid": 0
}

TASK [openshift_examples : Unarchive the OpenShift examples on the remote] *****
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:35
changed: [localhost] => {
    "changed": true, 
    "dest": "/etc/origin/examples/", 
    "extract_results": {
        "cmd": [
            "/bin/gtar", 
            "--extract", 
            "-C", 
            "/etc/origin/examples/", 
            "-f", 
            "/home/origin/.ansible/tmp/ansible-tmp-1519370360.24-55801809591229/source"
        ], 
        "err": "", 
        "out": "", 
        "rc": 0
    }, 
    "generated_timestamp": "2018-02-23 07:19:20.831220", 
    "gid": 0, 
    "group": "root", 
    "handler": "TarArchive", 
    "mode": "02755", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 141, 
    "src": "/home/origin/.ansible/tmp/ansible-tmp-1519370360.24-55801809591229/source", 
    "state": "directory", 
    "uid": 0
}

TASK [openshift_examples : Cleanup the OpenShift Examples temp dir] ************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:40
changed: [localhost -> localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-02-23 07:19:21.040381", 
    "path": "/tmp/openshift-ansible-UnMI7FB", 
    "state": "absent"
}

TASK [openshift_examples : Modify registry paths if registry_url is not registry.access.redhat.com] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:48
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:21.091697", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_examples : Import RHEL streams] ********************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:54
skipping: [localhost] => (item=/etc/origin/examples/image-streams/image-streams-rhel7.json)  => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:21.154705", 
    "item": "/etc/origin/examples/image-streams/image-streams-rhel7.json", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}
skipping: [localhost] => (item=/etc/origin/examples/image-streams/dotnet_imagestreams.json)  => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:21.165593", 
    "item": "/etc/origin/examples/image-streams/dotnet_imagestreams.json", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_examples : Import Centos Image streams] ************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:64
ok: [localhost] => {
    "changed": false, 
    "cmd": [
        "/usr/local/bin/oc", 
        "replace", 
        "--config=/etc/origin/master/admin.kubeconfig", 
        "-n", 
        "openshift", 
        "-f", 
        "/etc/origin/examples/image-streams/image-streams-centos7.json"
    ], 
    "delta": "0:00:00.393703", 
    "end": "2018-02-23 07:19:21.752310", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:21.779618", 
    "rc": 0, 
    "start": "2018-02-23 07:19:21.358607", 
    "stderr": [], 
    "stdout": [
        "imagestream \"httpd\" replaced", 
        "imagestream \"ruby\" replaced", 
        "imagestream \"nodejs\" replaced", 
        "imagestream \"perl\" replaced", 
        "imagestream \"php\" replaced", 
        "imagestream \"python\" replaced", 
        "imagestream \"wildfly\" replaced", 
        "imagestream \"mysql\" replaced", 
        "imagestream \"mariadb\" replaced", 
        "imagestream \"postgresql\" replaced", 
        "imagestream \"mongodb\" replaced", 
        "imagestream \"redis\" replaced", 
        "imagestream \"jenkins\" replaced"
    ]
}

TASK [openshift_examples : Import db templates] ********************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:72
ok: [localhost] => {
    "changed": false, 
    "cmd": [
        "/usr/local/bin/oc", 
        "replace", 
        "--config=/etc/origin/master/admin.kubeconfig", 
        "-n", 
        "openshift", 
        "-f", 
        "/etc/origin/examples/db-templates"
    ], 
    "delta": "0:00:00.507164", 
    "end": "2018-02-23 07:19:22.494192", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:22.524583", 
    "rc": 0, 
    "start": "2018-02-23 07:19:21.987028", 
    "stderr": [], 
    "stdout": [
        "template \"mariadb-ephemeral\" replaced", 
        "template \"mariadb-persistent\" replaced", 
        "template \"mongodb-ephemeral\" replaced", 
        "template \"mongodb-persistent\" replaced", 
        "template \"mysql-ephemeral\" replaced", 
        "template \"mysql-persistent\" replaced", 
        "template \"postgresql-ephemeral\" replaced", 
        "template \"postgresql-persistent\" replaced", 
        "template \"redis-ephemeral\" replaced", 
        "template \"redis-persistent\" replaced"
    ]
}

TASK [openshift_examples : Remove defunct quickstart template files] ***********
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:80
ok: [localhost] => (item=/etc/origin/examples/quickstart-templates/nodejs.json) => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:22.754115", 
    "item": "/etc/origin/examples/quickstart-templates/nodejs.json", 
    "path": "/etc/origin/examples/quickstart-templates/nodejs.json", 
    "state": "absent"
}
ok: [localhost] => (item=/etc/origin/examples/quickstart-templates/cakephp.json) => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:22.912027", 
    "item": "/etc/origin/examples/quickstart-templates/cakephp.json", 
    "path": "/etc/origin/examples/quickstart-templates/cakephp.json", 
    "state": "absent"
}
ok: [localhost] => (item=/etc/origin/examples/quickstart-templates/dancer.json) => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:23.062966", 
    "item": "/etc/origin/examples/quickstart-templates/dancer.json", 
    "path": "/etc/origin/examples/quickstart-templates/dancer.json", 
    "state": "absent"
}
ok: [localhost] => (item=/etc/origin/examples/quickstart-templates/django.json) => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:23.224597", 
    "item": "/etc/origin/examples/quickstart-templates/django.json", 
    "path": "/etc/origin/examples/quickstart-templates/django.json", 
    "state": "absent"
}

TASK [openshift_examples : Remove defunct quickstart templates from openshift namespace] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:90
ok: [localhost] => (item=nodejs-example) => {
    "changed": false, 
    "cmd": [
        "/usr/local/bin/oc", 
        "--config=/etc/origin/master/admin.kubeconfig", 
        "-n", 
        "openshift", 
        "delete", 
        "templates/nodejs-example"
    ], 
    "delta": "0:00:00.231499", 
    "end": "2018-02-23 07:19:23.634524", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:23.658052", 
    "item": "nodejs-example", 
    "rc": 1, 
    "start": "2018-02-23 07:19:23.403025", 
    "stderr": [
        "Error from server (NotFound): templates.template.openshift.io \"nodejs-example\" not found"
    ], 
    "stdout": []
}
ok: [localhost] => (item=cakephp-example) => {
    "changed": false, 
    "cmd": [
        "/usr/local/bin/oc", 
        "--config=/etc/origin/master/admin.kubeconfig", 
        "-n", 
        "openshift", 
        "delete", 
        "templates/cakephp-example"
    ], 
    "delta": "0:00:00.231832", 
    "end": "2018-02-23 07:19:24.020672", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:24.041943", 
    "item": "cakephp-example", 
    "rc": 1, 
    "start": "2018-02-23 07:19:23.788840", 
    "stderr": [
        "Error from server (NotFound): templates.template.openshift.io \"cakephp-example\" not found"
    ], 
    "stdout": []
}
ok: [localhost] => (item=dancer-example) => {
    "changed": false, 
    "cmd": [
        "/usr/local/bin/oc", 
        "--config=/etc/origin/master/admin.kubeconfig", 
        "-n", 
        "openshift", 
        "delete", 
        "templates/dancer-example"
    ], 
    "delta": "0:00:00.232165", 
    "end": "2018-02-23 07:19:24.408831", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:24.439390", 
    "item": "dancer-example", 
    "rc": 1, 
    "start": "2018-02-23 07:19:24.176666", 
    "stderr": [
        "Error from server (NotFound): templates.template.openshift.io \"dancer-example\" not found"
    ], 
    "stdout": []
}
ok: [localhost] => (item=django-example) => {
    "changed": false, 
    "cmd": [
        "/usr/local/bin/oc", 
        "--config=/etc/origin/master/admin.kubeconfig", 
        "-n", 
        "openshift", 
        "delete", 
        "templates/django-example"
    ], 
    "delta": "0:00:00.232225", 
    "end": "2018-02-23 07:19:24.802013", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:24.823084", 
    "item": "django-example", 
    "rc": 1, 
    "start": "2018-02-23 07:19:24.569788", 
    "stderr": [
        "Error from server (NotFound): templates.template.openshift.io \"django-example\" not found"
    ], 
    "stdout": []
}

TASK [openshift_examples : Import quickstart-templates] ************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:101
ok: [localhost] => {
    "changed": false, 
    "cmd": [
        "/usr/local/bin/oc", 
        "replace", 
        "--config=/etc/origin/master/admin.kubeconfig", 
        "-n", 
        "openshift", 
        "-f", 
        "/etc/origin/examples/quickstart-templates"
    ], 
    "delta": "0:00:00.670448", 
    "end": "2018-02-23 07:19:25.678891", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:25.703404", 
    "rc": 0, 
    "start": "2018-02-23 07:19:25.008443", 
    "stderr": [], 
    "stdout": [
        "template \"system\" replaced", 
        "template \"3scale-gateway\" replaced", 
        "template \"cakephp-mysql-persistent\" replaced", 
        "template \"cakephp-mysql-example\" replaced", 
        "template \"dancer-mysql-persistent\" replaced", 
        "template \"dancer-mysql-example\" replaced", 
        "template \"django-psql-persistent\" replaced", 
        "template \"django-psql-example\" replaced", 
        "template \"dotnet-example\" replaced", 
        "template \"dotnet-pgsql-persistent\" replaced", 
        "template \"dotnet-runtime-example\" replaced", 
        "template \"httpd-example\" replaced", 
        "template \"jenkins-ephemeral\" replaced", 
        "template \"jenkins-persistent\" replaced", 
        "template \"nodejs-mongo-persistent\" replaced", 
        "template \"nodejs-mongodb-example\" replaced", 
        "template \"amp-pvc\" replaced", 
        "template \"rails-pgsql-persistent\" replaced", 
        "template \"rails-postgresql-example\" replaced", 
        "template \"amp-apicast-wildcard-router\" replaced"
    ]
}

TASK [openshift_examples : Remove old xPaas template files] ********************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:109
ok: [localhost] => (item=/etc/origin/examples/xpaas-templates/sso70-basic.json) => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:25.907282", 
    "item": "/etc/origin/examples/xpaas-templates/sso70-basic.json", 
    "path": "/etc/origin/examples/xpaas-templates/sso70-basic.json", 
    "state": "absent"
}

TASK [openshift_examples : Remove old xPaas templates from openshift namespace] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:116
ok: [localhost] => (item=sso70-basic) => {
    "changed": false, 
    "cmd": [
        "/usr/local/bin/oc", 
        "--config=/etc/origin/master/admin.kubeconfig", 
        "-n", 
        "openshift", 
        "delete", 
        "templates/sso70-basic"
    ], 
    "delta": "0:00:00.235067", 
    "end": "2018-02-23 07:19:26.319879", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:26.343681", 
    "item": "sso70-basic", 
    "rc": 1, 
    "start": "2018-02-23 07:19:26.084812", 
    "stderr": [
        "Error from server (NotFound): templates.template.openshift.io \"sso70-basic\" not found"
    ], 
    "stdout": []
}

TASK [openshift_examples : Import xPaas image streams] *************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:124
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:26.391052", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_examples : Import xPaas templates] *****************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_examples/tasks/main.yml:132
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:26.440193", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Assert - Do not use firewalld on Atomic Host] **************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/main.yml:2
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:26.507192", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Install firewalld packages] ********************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:2
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:26.555348", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Ensure iptables services are not enabled] ******************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:7
skipping: [localhost] => (item=iptables)  => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:26.607530", 
    "item": "iptables", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}
skipping: [localhost] => (item=ip6tables)  => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:26.620707", 
    "item": "ip6tables", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Wait 10 seconds after disabling iptables] ******************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:19
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:26.666849", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Start and enable firewalld service] ************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:24
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:26.713345", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] ***
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:33
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:26.774923", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Restart polkitd] *******************************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:37
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:26.821156", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Wait for polkit action to have been created] ***************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:44
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:26.868913", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Add firewalld allow rules] *********************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:53

TASK [os_firewall : Remove firewalld allow rules] ******************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/firewalld.yml:61

TASK [os_firewall : Ensure firewalld service is not enabled] *******************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:3
ok: [localhost] => {
    "changed": false, 
    "enabled": false, 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:27.176905", 
    "name": "firewalld", 
    "state": "stopped", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "yes", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "no", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "firewalld.service", 
        "DevicePolicy": "auto", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/dev/null", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "firewalld.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "65536", 
        "LimitNPROC": "63327", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "63327", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "masked", 
        "MainPID": "0", 
        "MemoryAccounting": "yes", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "firewalld.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "inherit", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "UMask": "0022", 
        "UnitFileState": "bad", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [os_firewall : Wait 10 seconds after disabling firewalld] *****************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:12
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:27.228395", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Install iptables packages] *********************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:17
ok: [localhost] => (item=iptables) => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:28.028583", 
    "item": "iptables", 
    "msg": "", 
    "rc": 0, 
    "results": [
        "iptables-1.4.21-18.2.el7_4.x86_64 providing iptables is already installed"
    ]
}
ok: [localhost] => (item=iptables-services) => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:28.716176", 
    "item": "iptables-services", 
    "msg": "", 
    "rc": 0, 
    "results": [
        "iptables-services-1.4.21-18.2.el7_4.x86_64 providing iptables-services is already installed"
    ]
}

TASK [os_firewall : Start and enable iptables service] *************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:24
ok: [localhost] => {
    "changed": false, 
    "enabled": true, 
    "generated_timestamp": "2018-02-23 07:19:28.988711", 
    "name": "iptables", 
    "state": "started", 
    "status": {
        "ActiveEnterTimestamp": "Fri 2018-02-23 06:58:37 UTC", 
        "ActiveEnterTimestampMonotonic": "2517306143", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "active", 
        "After": "system.slice systemd-journald.socket basic.target syslog.target", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "yes", 
        "AssertTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "AssertTimestampMonotonic": "2517257510", 
        "Before": "docker.service network.service ip6tables.service shutdown.target", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "yes", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "yes", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "ConditionResult": "yes", 
        "ConditionTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "ConditionTimestampMonotonic": "2517257508", 
        "Conflicts": "shutdown.target", 
        "ConsistsOf": "docker.service", 
        "ControlGroup": "/system.slice/iptables.service", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "IPv4 firewall with iptables", 
        "DevicePolicy": "auto", 
        "Environment": "BOOTUP=serial CONSOLETYPE=serial", 
        "ExecMainCode": "1", 
        "ExecMainExitTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "ExecMainExitTimestampMonotonic": "2517300535", 
        "ExecMainPID": "8008", 
        "ExecMainStartTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "ExecMainStartTimestampMonotonic": "2517260742", 
        "ExecMainStatus": "0", 
        "ExecReload": "{ path=/usr/libexec/iptables/iptables.init ; argv[]=/usr/libexec/iptables/iptables.init reload ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStart": "{ path=/usr/libexec/iptables/iptables.init ; argv[]=/usr/libexec/iptables/iptables.init start ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStop": "{ path=/usr/libexec/iptables/iptables.init ; argv[]=/usr/libexec/iptables/iptables.init stop ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/usr/lib/systemd/system/iptables.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "iptables.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestamp": "Fri 2018-02-23 06:58:36 UTC", 
        "InactiveExitTimestampMonotonic": "2517260772", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "63327", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "63327", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "yes", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "iptables.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "yes", 
        "Requires": "basic.target", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system.slice", 
        "StandardError": "syslog", 
        "StandardInput": "null", 
        "StandardOutput": "syslog", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "exited", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "0", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "oneshot", 
        "UMask": "0022", 
        "UnitFilePreset": "disabled", 
        "UnitFileState": "enabled", 
        "WantedBy": "basic.target docker.service", 
        "Wants": "system.slice", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] ***
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:33
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:29.040482", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [os_firewall : Add iptables allow rules] **********************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:37

TASK [os_firewall : Remove iptables rules] *************************************
task path: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/firewall/iptables.yml:45

TASK [openshift_hosted_templates : Create local temp dir for OpenShift hosted templates copy] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:2
changed: [localhost -> localhost] => {
    "changed": true, 
    "cmd": [
        "mktemp", 
        "-d", 
        "/tmp/openshift-ansible-XXXXXXX"
    ], 
    "delta": "0:00:00.002942", 
    "end": "2018-02-23 07:19:29.325803", 
    "generated_timestamp": "2018-02-23 07:19:29.342699", 
    "rc": 0, 
    "start": "2018-02-23 07:19:29.322861", 
    "stderr": [], 
    "stdout": [
        "/tmp/openshift-ansible-CVq2clu"
    ]
}

TASK [openshift_hosted_templates : Create tar of OpenShift examples] ***********
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:10
changed: [localhost -> localhost] => {
    "changed": true, 
    "cmd": [
        "tar", 
        "-C", 
        "/usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/files/v3.6/origin", 
        "-cvf", 
        "/tmp/openshift-ansible-CVq2clu/openshift-hosted-templates.tar", 
        "."
    ], 
    "delta": "0:00:00.004047", 
    "end": "2018-02-23 07:19:29.532989", 
    "generated_timestamp": "2018-02-23 07:19:29.548573", 
    "rc": 0, 
    "start": "2018-02-23 07:19:29.528942", 
    "stderr": [], 
    "stdout": [
        "./", 
        "./registry-console.yaml"
    ]
}

TASK [openshift_hosted_templates : Create remote OpenShift hosted templates directory] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:19
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-02-23 07:19:29.739500", 
    "gid": 0, 
    "group": "root", 
    "mode": "0755", 
    "owner": "root", 
    "path": "/etc/origin/hosted", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 35, 
    "state": "directory", 
    "uid": 0
}

TASK [openshift_hosted_templates : Unarchive the OpenShift hosted templates on the remote] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:25
changed: [localhost] => {
    "changed": true, 
    "dest": "/etc/origin/hosted/", 
    "extract_results": {
        "cmd": [
            "/bin/gtar", 
            "--extract", 
            "-C", 
            "/etc/origin/hosted/", 
            "-f", 
            "/home/origin/.ansible/tmp/ansible-tmp-1519370369.79-6356245976251/source"
        ], 
        "err": "", 
        "out": "", 
        "rc": 0
    }, 
    "generated_timestamp": "2018-02-23 07:19:30.291772", 
    "gid": 0, 
    "group": "root", 
    "handler": "TarArchive", 
    "mode": "02755", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 35, 
    "src": "/home/origin/.ansible/tmp/ansible-tmp-1519370369.79-6356245976251/source", 
    "state": "directory", 
    "uid": 0
}

TASK [openshift_hosted_templates : Cleanup the OpenShift hosted templates temp dir] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:30
changed: [localhost -> localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-02-23 07:19:30.498191", 
    "path": "/tmp/openshift-ansible-CVq2clu", 
    "state": "absent"
}

TASK [openshift_hosted_templates : Modify registry paths if registry_url is not registry.access.redhat.com] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:34
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:30.546545", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_hosted_templates : Create temp directory for kubeconfig] *******
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:39
ok: [localhost] => {
    "changed": false, 
    "cmd": [
        "mktemp", 
        "-d", 
        "/tmp/openshift-ansible-XXXXXX"
    ], 
    "delta": "0:00:00.002808", 
    "end": "2018-02-23 07:19:30.715483", 
    "generated_timestamp": "2018-02-23 07:19:30.732179", 
    "rc": 0, 
    "start": "2018-02-23 07:19:30.712675", 
    "stderr": [], 
    "stdout": [
        "/tmp/openshift-ansible-KW9lb7"
    ]
}

TASK [openshift_hosted_templates : Record kubeconfig tmp dir] ******************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:44
ok: [localhost] => {
    "ansible_facts": {
        "openshift_hosted_templates_kubeconfig": "/tmp/openshift-ansible-KW9lb7/admin.kubeconfig"
    }, 
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:30.779074"
}

TASK [openshift_hosted_templates : Copy the admin client config(s)] ************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:48
ok: [localhost] => {
    "changed": false, 
    "cmd": [
        "cp", 
        "/etc/origin/master/admin.kubeconfig", 
        "/tmp/openshift-ansible-KW9lb7/admin.kubeconfig"
    ], 
    "delta": "0:00:00.004046", 
    "end": "2018-02-23 07:19:30.962614", 
    "generated_timestamp": "2018-02-23 07:19:30.983260", 
    "rc": 0, 
    "start": "2018-02-23 07:19:30.958568", 
    "stderr": [], 
    "stdout": []
}

TASK [openshift_hosted_templates : Create or update hosted templates] **********
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:53
ok: [localhost] => {
    "changed": false, 
    "cmd": [
        "/usr/local/bin/oc", 
        "replace", 
        "-f", 
        "/etc/origin/hosted", 
        "--config=/tmp/openshift-ansible-KW9lb7/admin.kubeconfig", 
        "-n", 
        "openshift"
    ], 
    "delta": "0:00:00.271227", 
    "end": "2018-02-23 07:19:31.435564", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:31.461524", 
    "rc": 0, 
    "start": "2018-02-23 07:19:31.164337", 
    "stderr": [], 
    "stdout": [
        "template \"registry-console\" replaced"
    ]
}

TASK [openshift_hosted_templates : Delete temp directory] **********************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_hosted_templates/tasks/main.yml:63
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:31.664778", 
    "path": "/tmp/openshift-ansible-KW9lb7", 
    "state": "absent"
}
META: ran handlers
META: ran handlers

PLAY [Clean up and display warnings] *******************************************
META: ran handlers

TASK [openshift_excluder : Detecting Atomic Host Operating System] *************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/main.yml:2
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:31.845965", 
    "stat": {
        "exists": false
    }
}

TASK [openshift_excluder : Debug r_openshift_excluder_enable_docker_excluder] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/main.yml:9
ok: [localhost] => {
    "generated_timestamp": "2018-02-23 07:19:31.884298", 
    "r_openshift_excluder_enable_docker_excluder": "false"
}

TASK [openshift_excluder : Debug r_openshift_excluder_enable_openshift_excluder] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/main.yml:13
ok: [localhost] => {
    "generated_timestamp": "2018-02-23 07:19:31.924011", 
    "r_openshift_excluder_enable_openshift_excluder": "false"
}

TASK [openshift_excluder : Fail if invalid openshift_excluder_action provided] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/main.yml:17
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:31.959564", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_excluder : Fail if r_openshift_excluder_service_type is not defined] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/main.yml:22
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:31.995508", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_excluder : Fail if r_openshift_excluder_upgrade_target is not defined] ***
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/main.yml:27
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:32.028033", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_excluder : Include main action task file] **********************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/main.yml:34
included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/enable.yml for localhost

TASK [openshift_excluder : Install excluders] **********************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/enable.yml:2
included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/install.yml for localhost

TASK [openshift_excluder : Install docker excluder] ****************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/install.yml:9
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:32.169864", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_excluder : Install openshift excluder] *************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/install.yml:16
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:32.212373", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_excluder : set_fact] *******************************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/install.yml:23
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:32.258012", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_excluder : Enable excluders] ***********************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/enable.yml:5
included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for localhost

TASK [openshift_excluder : Check for docker-excluder] **************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml:2
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:32.502859", 
    "stat": {
        "exists": false
    }
}

TASK [openshift_excluder : Enable docker excluder] *****************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml:7
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:32.537150", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}

TASK [openshift_excluder : Check for openshift excluder] ***********************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml:13
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:32.718944", 
    "stat": {
        "exists": false
    }
}

TASK [openshift_excluder : Enable openshift excluder] **************************
task path: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml:18
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:32.752732", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}
META: ran handlers

TASK [grep pluginOrderOverride] ************************************************
task path: /usr/share/ansible/openshift-ansible/playbooks/common/openshift-cluster/upgrades/post_control_plane.yml:121
ok: [localhost] => {
    "changed": false, 
    "cmd": [
        "grep", 
        "pluginOrderOverride", 
        "/etc/origin/master/master-config.yaml"
    ], 
    "delta": "0:00:00.002938", 
    "end": "2018-02-23 07:19:32.920732", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2018-02-23 07:19:32.938008", 
    "rc": 0, 
    "start": "2018-02-23 07:19:32.917794", 
    "stderr": [], 
    "stdout": [
        "  pluginOrderOverride: null"
    ]
}

TASK [Warn if pluginOrderOverride is in use in master-config.yaml] *************
task path: /usr/share/ansible/openshift-ansible/playbooks/common/openshift-cluster/upgrades/post_control_plane.yml:128
ok: [localhost] => {
    "generated_timestamp": "2018-02-23 07:19:32.976158", 
    "msg": "WARNING pluginOrderOverride is being deprecated in master-config.yaml, please see https://docs.openshift.com/enterprise/latest/architecture/additional_concepts/admission_controllers.html for more information."
}

TASK [Warn if shared-resource-viewer could not be updated] *********************
task path: /usr/share/ansible/openshift-ansible/playbooks/common/openshift-cluster/upgrades/post_control_plane.yml:135
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 07:19:33.004695", 
    "skip_reason": "Conditional result was False", 
    "skipped": true
}
META: ran handlers

PLAY RECAP *********************************************************************
localhost                  : ok=649  changed=55   unreachable=0    failed=0   

+ set +o xtrace
########## FINISHED STAGE: SUCCESS: UPDATE ORIGIN TO RELEASE [00h 08m 24s] ##########
[workspace] $ /bin/bash /tmp/jenkins5700352079461902563.sh
########## STARTING STAGE: EXPOSE THE KUBECONFIG ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ export PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ mktemp
+ script=/tmp/tmp.xB3hPjE7X4
+ cat
+ chmod +x /tmp/tmp.xB3hPjE7X4
+ scp -F ./.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.xB3hPjE7X4 openshiftdevel:/tmp/tmp.xB3hPjE7X4
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.xB3hPjE7X4"'
+ cd /home/origin
+ sudo chmod a+x /etc/ /etc/origin/ /etc/origin/master/
+ sudo chmod a+rw /etc/origin/master/admin.kubeconfig
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: EXPOSE THE KUBECONFIG [00h 00m 01s] ##########
[workspace] $ /bin/bash /tmp/jenkins223524860090935198.sh
########## STARTING STAGE: ENSURE BUILT VERSION OF ORIGIN IS UPDATED ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ export PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ mktemp
+ script=/tmp/tmp.6PZPZMuJ8M
+ cat
+ chmod +x /tmp/tmp.6PZPZMuJ8M
+ scp -F ./.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.6PZPZMuJ8M openshiftdevel:/tmp/tmp.6PZPZMuJ8M
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 600 /tmp/tmp.6PZPZMuJ8M"'
+ cd /data/src/github.com/openshift/aos-cd-jobs
++ cat ./ORIGIN_TAG
+ docker_image_tag=v3.6.1
+ docker_repository=openshift/origin
+ docker inspect openshift/origin:v3.6.1
[
    {
        "Id": "sha256:8561271a7bae6772415a6936aadff87f5edb1c9c0aa6bd77b0c884d15783439e",
        "RepoTags": [
            "openshift/origin:latest",
            "openshift/origin:v3.6.1"
        ],
        "RepoDigests": [],
        "Parent": "sha256:802c2fff01b27b8f08fc03107cda5254f1ced71d145f89f73f0edd1674108503",
        "Comment": "",
        "Created": "2018-02-23T06:50:17.305892872Z",
        "Container": "bd4df1e173e8e6553f92c795d3fbaf3155d2b7a41474015d047f9dcf4f413f7e",
        "ContainerConfig": {
            "Hostname": "bd4df1e173e8",
            "Domainname": "",
            "User": "",
            "AttachStdin": false,
            "AttachStdout": false,
            "AttachStderr": false,
            "Tty": false,
            "OpenStdin": false,
            "StdinOnce": false,
            "Env": [
                "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
            ],
            "Cmd": [
                "sleep 86400"
            ],
            "Image": "openshift/origin-base",
            "Volumes": null,
            "WorkingDir": "",
            "Entrypoint": [
                "/bin/sh",
                "-c"
            ],
            "OnBuild": null,
            "Labels": {
                "build-date": "20180107",
                "io.k8s.description": "This is the base image from which all OpenShift Origin images inherit.",
                "io.k8s.display-name": "OpenShift Origin CentOS 7 Base",
                "io.openshift.tags": "openshift,base",
                "license": "GPLv2",
                "name": "CentOS Base Image",
                "vendor": "CentOS"
            }
        },
        "DockerVersion": "1.13.1",
        "Author": "",
        "Config": {
            "Hostname": "",
            "Domainname": "",
            "User": "",
            "AttachStdin": false,
            "AttachStdout": false,
            "AttachStderr": false,
            "ExposedPorts": {
                "53/tcp": {},
                "8443/tcp": {}
            },
            "Tty": false,
            "OpenStdin": false,
            "StdinOnce": false,
            "Env": [
                "HOME=/root",
                "OPENSHIFT_CONTAINERIZED=true",
                "KUBECONFIG=/var/lib/origin/openshift.local.config/master/admin.kubeconfig",
                "PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
            ],
            "Cmd": null,
            "ArgsEscaped": true,
            "Image": "",
            "Volumes": null,
            "WorkingDir": "/var/lib/origin",
            "Entrypoint": [
                "/usr/bin/openshift"
            ],
            "OnBuild": null,
            "Labels": {
                "build-date": "20180107",
                "io.k8s.description": "OpenShift Origin is a platform for developing, building, and deploying containerized applications.",
                "io.k8s.display-name": "OpenShift Origin Application Platform",
                "io.openshift.tags": "openshift,core",
                "license": "GPLv2",
                "name": "CentOS Base Image",
                "vendor": "CentOS"
            }
        },
        "Architecture": "amd64",
        "Os": "linux",
        "Size": 1018023944,
        "VirtualSize": 1018023944,
        "GraphDriver": {
            "Name": "devicemapper",
            "Data": {
                "DeviceId": "49",
                "DeviceName": "docker-202:2-8678629-636055e929de95c07e258262486e37b74d6c3f34ee7edcdfa1563a7eecde7125",
                "DeviceSize": "10737418240"
            }
        },
        "RootFS": {
            "Type": "layers",
            "Layers": [
                "sha256:e15afa4858b655f8a5da4c4a41e05b908229f6fab8543434db79207478511ff7",
                "sha256:f4a186608ab3201a8f5c8fce7e2cb0ee193d5d9c0ffe67b0836aab34f239f611",
                "sha256:7c837afd8f130992e48d9fe8f68ea48da6ea8d52e4b526f9d8e8b5e075f8cebd",
                "sha256:cf6056646ac4fe846a883920af2ec0a8af7dba0796fc3d62bd7d9c0b8b37fea4"
            ]
        }
    }
]
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: ENSURE BUILT VERSION OF ORIGIN IS UPDATED [00h 00m 00s] ##########
[workspace] $ /bin/bash /tmp/jenkins4008805149997883566.sh
########## STARTING STAGE: RUN EXTENDED TESTS ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ export PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ mktemp
+ script=/tmp/tmp.PI8Nh318Mv
+ cat
+ chmod +x /tmp/tmp.PI8Nh318Mv
+ scp -F ./.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.PI8Nh318Mv openshiftdevel:/tmp/tmp.PI8Nh318Mv
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.PI8Nh318Mv"'
+ cd /data/src/github.com/openshift/origin
+ OS_BUILD_ENV_PULL_IMAGE=true
+ OS_BUILD_ENV_PRESERVE=_output/local/bin/linux/amd64/extended.test
+ hack/env make build-extended-test
[INFO] Pulling the openshift/origin-release:golang-1.7 image to update it...
Trying to pull repository registry.access.redhat.com/openshift/origin-release ... 
Trying to pull repository docker.io/openshift/origin-release ... 
sha256:b82c4cd9dc5b1bd947017ef0c232951b13d19730543326ba218ce6491d7d60ad: Pulling from docker.io/openshift/origin-release
Digest: sha256:b82c4cd9dc5b1bd947017ef0c232951b13d19730543326ba218ce6491d7d60ad
Status: Image is up to date for docker.io/openshift/origin-release:golang-1.7
hack/build-go.sh test/extended/extended.test
++ Building go targets for linux/amd64: test/extended/extended.test
hack/build-go.sh took 379 seconds
+ cat
+ sudo cp ci-dnsmasq.conf /etc/dnsmasq.d/ci-dnsmasq.conf
+ sudo systemctl restart dnsmasq
+ sudo systemctl status dnsmasq
● dnsmasq.service - DNS caching server.
   Loaded: loaded (/usr/lib/systemd/system/dnsmasq.service; enabled; vendor preset: disabled)
   Active: active (running) since Fri 2018-02-23 07:26:03 UTC; 18ms ago
 Main PID: 31595 (dnsmasq)
   Memory: 220.0K
   CGroup: /system.slice/dnsmasq.service
           └─31595 /usr/sbin/dnsmasq -k

Feb 23 07:26:03 ip-172-18-2-110.ec2.internal systemd[1]: Started DNS caching server..
Feb 23 07:26:03 ip-172-18-2-110.ec2.internal systemd[1]: Starting DNS caching server....
Feb 23 07:26:03 ip-172-18-2-110.ec2.internal dnsmasq[31595]: started, version 2.76 cachesize 150
Feb 23 07:26:03 ip-172-18-2-110.ec2.internal dnsmasq[31595]: compile time options: IPv6 GNU-getopt DBus no-i18n IDN DHCP DHCPv6 no-Lua TFTP no-conntrack ipset auth no-DNSSEC loop-detect inotify
Feb 23 07:26:03 ip-172-18-2-110.ec2.internal dnsmasq[31595]: DBus support enabled: connected to system bus
Feb 23 07:26:03 ip-172-18-2-110.ec2.internal dnsmasq[31595]: using nameserver 127.0.0.1#8053 for domain 17.30.172.in-addr.arpa
Feb 23 07:26:03 ip-172-18-2-110.ec2.internal dnsmasq[31595]: using nameserver 127.0.0.1#8053 for domain local
Feb 23 07:26:03 ip-172-18-2-110.ec2.internal dnsmasq[31595]: using nameserver 172.18.0.2#53
Feb 23 07:26:03 ip-172-18-2-110.ec2.internal dnsmasq[31595]: read /etc/hosts - 2 addresses
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: RUN EXTENDED TESTS [00h 06m 29s] ##########
[PostBuildScript] - Execution post build scripts.
[workspace] $ /bin/bash /tmp/jenkins2798077570972592396.sh
########## STARTING STAGE: DOWNLOAD ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ export PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/gathered
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/gathered
+ mkdir -p /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/gathered
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo stat /data/src/github.com/openshift/origin/_output/scripts
  File: ‘/data/src/github.com/openshift/origin/_output/scripts’
  Size: 127       	Blocks: 0          IO Block: 4096   directory
Device: ca02h/51714d	Inode: 58928693    Links: 8
Access: (2755/drwxr-sr-x)  Uid: ( 1001/  origin)   Gid: ( 1003/origin-git)
Context: unconfined_u:object_r:default_t:s0
Access: 2018-02-23 06:22:05.840407044 +0000
Modify: 2018-02-23 06:57:34.695029634 +0000
Change: 2018-02-23 06:57:34.695029634 +0000
 Birth: -
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod -R o+rX /data/src/github.com/openshift/origin/_output/scripts
+ scp -r -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel:/data/src/github.com/openshift/origin/_output/scripts /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/gathered
+ tree /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/gathered
/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/gathered
└── scripts
    ├── ansible_junit
    │   ├── AyVpYzXGEs.xml
    │   ├── CGDTbDbGGW.xml
    │   ├── DkhbHiCTgN.xml
    │   └── UnMsduFeZZ.xml
    ├── build-base-images
    │   ├── artifacts
    │   ├── logs
    │   └── openshift.local.home
    ├── build-images
    │   ├── artifacts
    │   ├── logs
    │   │   └── scripts.log
    │   └── openshift.local.home
    ├── env
    │   ├── artifacts
    │   ├── logs
    │   │   └── scripts.log
    │   └── openshift.local.home
    ├── origin_version
    │   ├── artifacts
    │   ├── logs
    │   └── openshift.local.home
    └── tmp.Trm1JdaCQa
        ├── artifacts
        ├── logs
        └── openshift.local.home

22 directories, 6 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins2485516537417204224.sh
########## STARTING STAGE: GENERATE ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ export PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/generated
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/generated
+ mkdir /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/generated
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a 2>&1'
  WARNING: You're not using the default seccomp profile
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo cat /etc/sysconfig/docker /etc/sysconfig/docker-network /etc/sysconfig/docker-storage /etc/sysconfig/docker-storage-setup /etc/systemd/system/docker.service 2>&1'
+ true
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1 2>&1'
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'oc get --raw /metrics --server=https://$( uname --nodename ):10250 --config=/etc/origin/master/admin.kubeconfig 2>&1'
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC 2>&1'
+ true
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'oc get --raw /metrics --config=/etc/origin/master/admin.kubeconfig 2>&1'
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo df -T -h && sudo pvs && sudo vgs && sudo lvs && sudo findmnt --all 2>&1'
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo yum list installed 2>&1'
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl --dmesg --no-pager --all --lines=all 2>&1'
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl _PID=1 --no-pager --all --lines=all 2>&1'
+ tree /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/generated
/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/generated
├── avc_denials.log
├── containers.log
├── dmesg.log
├── docker.config
├── docker.info
├── filesystem.info
├── installed_packages.log
├── master-metrics.log
├── node-metrics.log
└── pid1.journal

0 directories, 10 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins7716862806355933854.sh
########## STARTING STAGE: FETCH SYSTEMD JOURNALS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ export PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/journals
+ rm -rf /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/journals
+ mkdir /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/journals
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit docker.service --no-pager --all --lines=all
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit origin-master.service --no-pager --all --lines=all
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit origin-master-api.service --no-pager --all --lines=all
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit origin-master-controllers.service --no-pager --all --lines=all
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit origin-node.service --no-pager --all --lines=all
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit openvswitch.service --no-pager --all --lines=all
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit ovs-vswitchd.service --no-pager --all --lines=all
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit ovsdb-server.service --no-pager --all --lines=all
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit etcd.service --no-pager --all --lines=all
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
+ tree /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/journals
/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/artifacts/journals
├── dnsmasq.service
├── docker.service
├── etcd.service
├── openvswitch.service
├── origin-master-api.service
├── origin-master-controllers.service
├── origin-master.service
├── origin-node.service
├── ovsdb-server.service
├── ovs-vswitchd.service
└── systemd-journald.service

0 directories, 11 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins967684454370125186.sh
########## STARTING STAGE: FORWARD PARAMETERS TO THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ export PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod o+rw /etc/environment
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''BUILD_URL=https://ci.openshift.redhat.com/jenkins/job/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/'\'' >> /etc/environment'
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: FORWARD PARAMETERS TO THE REMOTE HOST [00h 00m 01s] ##########
[workspace] $ /bin/bash /tmp/jenkins2219791930544431008.sh
########## STARTING STAGE: RECORD THE ENDING METADATA ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ export PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ mktemp
+ script=/tmp/tmp.0Z4XYs7x24
+ cat
+ chmod +x /tmp/tmp.0Z4XYs7x24
+ scp -F ./.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.0Z4XYs7x24 openshiftdevel:/tmp/tmp.0Z4XYs7x24
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 300 /tmp/tmp.0Z4XYs7x24"'
+ cd /data/src/github.com/openshift/aos-cd-jobs
+ trap 'exit 0' EXIT
+ sjb/gcs/finished.py
+ exit 0
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: RECORD THE ENDING METADATA [00h 00m 01s] ##########
[workspace] $ /bin/bash /tmp/jenkins2121110374770003075.sh
########## STARTING STAGE: ASSEMBLE GCS OUTPUT ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ export PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
+ trap 'exit 0' EXIT
+ mkdir -p gcs/artifacts gcs/artifacts/generated gcs/artifacts/journals gcs/artifacts/gathered
+ scp -F ./.config/origin-ci-tool/inventory/.ssh_config openshiftdevel:/data/finished.json gcs/
+ cat /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/builds/1476/log
+ cp artifacts/generated/avc_denials.log artifacts/generated/containers.log artifacts/generated/dmesg.log artifacts/generated/docker.config artifacts/generated/docker.info artifacts/generated/filesystem.info artifacts/generated/installed_packages.log artifacts/generated/master-metrics.log artifacts/generated/node-metrics.log artifacts/generated/pid1.journal gcs/artifacts/generated/
+ cp artifacts/journals/dnsmasq.service artifacts/journals/docker.service artifacts/journals/etcd.service artifacts/journals/openvswitch.service artifacts/journals/origin-master-api.service artifacts/journals/origin-master-controllers.service artifacts/journals/origin-master.service artifacts/journals/origin-node.service artifacts/journals/ovsdb-server.service artifacts/journals/ovs-vswitchd.service artifacts/journals/systemd-journald.service gcs/artifacts/journals/
+ cp -r artifacts/gathered/scripts gcs/artifacts/
++ pwd
+ scp -F ./.config/origin-ci-tool/inventory/.ssh_config -r /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/gcs openshiftdevel:/data
+ scp -F ./.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.config/gcloud/gcs-publisher-credentials.json openshiftdevel:/data/credentials.json
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins1599982812852971354.sh
########## STARTING STAGE: PUSH THE ARTIFACTS AND METADATA ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ export PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ mktemp
+ script=/tmp/tmp.pvGgH6dTtY
+ cat
+ chmod +x /tmp/tmp.pvGgH6dTtY
+ scp -F ./.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.pvGgH6dTtY openshiftdevel:/tmp/tmp.pvGgH6dTtY
+ ssh -F ./.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 300 /tmp/tmp.pvGgH6dTtY"'
+ cd /home/origin
+ trap 'exit 0' EXIT
+ [[ -n {"type":"presubmit","job":"test_pull_request_openshift_ansible_extended_conformance_install_update_containerized","buildid":"f0e9c7e0-1860-11e8-9ee3-0a58ac100ecc","refs":{"org":"openshift","repo":"openshift-ansible","base_ref":"release-3.6","base_sha":"5542b7de8fb6f896266dcb9f820fcf500409bca3","pulls":[{"number":7232,"author":"kwoodson","sha":"11198c0fcc614c4df56241e505df2d09e0710fcd"}]}} ]]
++ jq --compact-output .buildid
+ [[ "f0e9c7e0-1860-11e8-9ee3-0a58ac100ecc" =~ ^"[0-9]+"$ ]]
+ echo 'Using BUILD_NUMBER'
Using BUILD_NUMBER
++ jq --compact-output '.buildid |= "1476"'
+ JOB_SPEC='{"type":"presubmit","job":"test_pull_request_openshift_ansible_extended_conformance_install_update_containerized","buildid":"1476","refs":{"org":"openshift","repo":"openshift-ansible","base_ref":"release-3.6","base_sha":"5542b7de8fb6f896266dcb9f820fcf500409bca3","pulls":[{"number":7232,"author":"kwoodson","sha":"11198c0fcc614c4df56241e505df2d09e0710fcd"}]}}'
+ docker run -e 'JOB_SPEC={"type":"presubmit","job":"test_pull_request_openshift_ansible_extended_conformance_install_update_containerized","buildid":"1476","refs":{"org":"openshift","repo":"openshift-ansible","base_ref":"release-3.6","base_sha":"5542b7de8fb6f896266dcb9f820fcf500409bca3","pulls":[{"number":7232,"author":"kwoodson","sha":"11198c0fcc614c4df56241e505df2d09e0710fcd"}]}}' -v /data:/data:z registry.svc.ci.openshift.org/ci/gcsupload:latest --dry-run=false --gcs-bucket=origin-ci-test --gcs-credentials-file=/data/credentials.json --path-strategy=single --default-org=openshift --default-repo=origin /data/gcs/artifacts /data/gcs/build-log.txt /data/gcs/finished.json /data/gcs/started.json
Unable to find image 'registry.svc.ci.openshift.org/ci/gcsupload:latest' locally
Trying to pull repository registry.svc.ci.openshift.org/ci/gcsupload ... 
sha256:8cde44133fe8168724f74a3e7f5417580fc11c06a74ebef31ed6a4119537e08c: Pulling from registry.svc.ci.openshift.org/ci/gcsupload
6d987f6f4279: Pulling fs layer
186fc760c1fb: Pulling fs layer
03330e9843bb: Pulling fs layer
186fc760c1fb: Verifying Checksum
186fc760c1fb: Download complete
6d987f6f4279: Verifying Checksum
6d987f6f4279: Download complete
03330e9843bb: Verifying Checksum
03330e9843bb: Download complete
6d987f6f4279: Pull complete
186fc760c1fb: Pull complete
03330e9843bb: Pull complete
Digest: sha256:8cde44133fe8168724f74a3e7f5417580fc11c06a74ebef31ed6a4119537e08c
Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/gcsupload:latest
time="2018-02-23T07:26:40Z" level=info msg="Gathering artifacts from artifact directory: /data/gcs/artifacts" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/generated/avc_denials.log in artifact directory. Uploading as artifacts/generated/avc_denials.log\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/generated/containers.log in artifact directory. Uploading as artifacts/generated/containers.log\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/generated/dmesg.log in artifact directory. Uploading as artifacts/generated/dmesg.log\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/generated/docker.config in artifact directory. Uploading as artifacts/generated/docker.config\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/generated/docker.info in artifact directory. Uploading as artifacts/generated/docker.info\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/generated/filesystem.info in artifact directory. Uploading as artifacts/generated/filesystem.info\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/generated/installed_packages.log in artifact directory. Uploading as artifacts/generated/installed_packages.log\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/generated/master-metrics.log in artifact directory. Uploading as artifacts/generated/master-metrics.log\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/generated/node-metrics.log in artifact directory. Uploading as artifacts/generated/node-metrics.log\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/generated/pid1.journal in artifact directory. Uploading as artifacts/generated/pid1.journal\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/journals/dnsmasq.service in artifact directory. Uploading as artifacts/journals/dnsmasq.service\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/journals/docker.service in artifact directory. Uploading as artifacts/journals/docker.service\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/journals/etcd.service in artifact directory. Uploading as artifacts/journals/etcd.service\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/journals/openvswitch.service in artifact directory. Uploading as artifacts/journals/openvswitch.service\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/journals/origin-master-api.service in artifact directory. Uploading as artifacts/journals/origin-master-api.service\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/journals/origin-master-controllers.service in artifact directory. Uploading as artifacts/journals/origin-master-controllers.service\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/journals/origin-master.service in artifact directory. Uploading as artifacts/journals/origin-master.service\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/journals/origin-node.service in artifact directory. Uploading as artifacts/journals/origin-node.service\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/journals/ovs-vswitchd.service in artifact directory. Uploading as artifacts/journals/ovs-vswitchd.service\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/journals/ovsdb-server.service in artifact directory. Uploading as artifacts/journals/ovsdb-server.service\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/journals/systemd-journald.service in artifact directory. Uploading as artifacts/journals/systemd-journald.service\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/scripts/ansible_junit/AyVpYzXGEs.xml in artifact directory. Uploading as artifacts/scripts/ansible_junit/AyVpYzXGEs.xml\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/scripts/ansible_junit/CGDTbDbGGW.xml in artifact directory. Uploading as artifacts/scripts/ansible_junit/CGDTbDbGGW.xml\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/scripts/ansible_junit/DkhbHiCTgN.xml in artifact directory. Uploading as artifacts/scripts/ansible_junit/DkhbHiCTgN.xml\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/scripts/ansible_junit/UnMsduFeZZ.xml in artifact directory. Uploading as artifacts/scripts/ansible_junit/UnMsduFeZZ.xml\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/scripts/build-images/logs/scripts.log in artifact directory. Uploading as artifacts/scripts/build-images/logs/scripts.log\n" 
time="2018-02-23T07:26:40Z" level=info msg="Found /data/gcs/artifacts/scripts/env/logs/scripts.log in artifact directory. Uploading as artifacts/scripts/env/logs/scripts.log\n" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/scripts/ansible_junit/AyVpYzXGEs.xml" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/scripts/ansible_junit/DkhbHiCTgN.xml" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/build-log.txt" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/generated/dmesg.log" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/journals/systemd-journald.service" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/journals/ovsdb-server.service" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/scripts/ansible_junit/CGDTbDbGGW.xml" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/scripts/env/logs/scripts.log" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/started.json" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/directory/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476.txt" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/journals/openvswitch.service" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/generated/master-metrics.log" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/journals/origin-master-api.service" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/scripts/build-images/logs/scripts.log" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/generated/avc_denials.log" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/generated/docker.config" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/scripts/ansible_junit/UnMsduFeZZ.xml" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/generated/filesystem.info" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/generated/pid1.journal" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/journals/origin-node.service" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/finished.json" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/journals/docker.service" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/journals/origin-master.service" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/generated/containers.log" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/generated/docker.info" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/generated/node-metrics.log" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/journals/ovs-vswitchd.service" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/journals/etcd.service" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/journals/origin-master-controllers.service" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/generated/installed_packages.log" 
time="2018-02-23T07:26:40Z" level=info msg="Queued for upload" dest="pr-logs/pull/openshift_openshift-ansible/7232/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/1476/artifacts/journals/dnsmasq.service" 
+ exit 0
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: PUSH THE ARTIFACTS AND METADATA [00h 00m 14s] ##########
[workspace] $ /bin/bash /tmp/jenkins1045100368901728129.sh
########## STARTING STAGE: DEPROVISION CLOUD RESOURCES ##########
+ [[ -s /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate ]]
+ source /var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704
++ export PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config
+ oct deprovision

PLAYBOOK: main.yml *************************************************************
4 plays in /var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml

PLAY [ensure we have the parameters necessary to deprovision virtual hosts] ****

TASK [ensure all required variables are set] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:9
skipping: [localhost] => (item=origin_ci_inventory_dir)  => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 02:26:42.643877", 
    "item": "origin_ci_inventory_dir", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_region)  => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 02:26:42.646351", 
    "item": "origin_ci_aws_region", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

PLAY [deprovision virtual hosts in EC2] ****************************************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

TASK [deprovision a virtual EC2 host] ******************************************
task path: /var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:28
included: /var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml for localhost

TASK [update the SSH configuration to remove AWS EC2 specifics] ****************
task path: /var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:2
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2018-02-23 02:26:43.454385", 
    "msg": ""
}

TASK [rename EC2 instance for termination reaper] ******************************
task path: /var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:8
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-02-23 02:26:44.025111", 
    "msg": "Tags {'Name': 'oct-terminate'} created for resource i-0c163ccadbacd5266."
}

TASK [tear down the EC2 instance] **********************************************
task path: /var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:15
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-02-23 02:26:44.942540", 
    "instance_ids": [
        "i-0c163ccadbacd5266"
    ], 
    "instances": [
        {
            "ami_launch_index": "0", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-0cecc80618bdd9d11"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-08385c45672e27891"
                }
            }, 
            "dns_name": "ec2-54-209-5-234.compute-1.amazonaws.com", 
            "ebs_optimized": false, 
            "groups": {
                "sg-7e73221a": "default"
            }, 
            "hypervisor": "xen", 
            "id": "i-0c163ccadbacd5266", 
            "image_id": "ami-0ab8233482bac3d81", 
            "instance_type": "m4.xlarge", 
            "kernel": null, 
            "key_name": "libra", 
            "launch_time": "2018-02-23T06:16:14.000Z", 
            "placement": "us-east-1d", 
            "private_dns_name": "ip-172-18-2-110.ec2.internal", 
            "private_ip": "172.18.2.110", 
            "public_dns_name": "ec2-54-209-5-234.compute-1.amazonaws.com", 
            "public_ip": "54.209.5.234", 
            "ramdisk": null, 
            "region": "us-east-1", 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "running", 
            "state_code": 16, 
            "tags": {
                "Name": "oct-terminate", 
                "openshift_etcd": "", 
                "openshift_master": "", 
                "openshift_node": ""
            }, 
            "tenancy": "default", 
            "virtualization_type": "hvm"
        }
    ], 
    "tagged_instances": []
}

TASK [remove the serialized host variables] ************************************
task path: /var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:22
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-02-23 02:26:45.202186", 
    "path": "/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config/origin-ci-tool/inventory/host_vars/172.18.2.110.yml", 
    "state": "absent"
}

PLAY [deprovision virtual hosts locally manged by Vagrant] *********************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

PLAY [clean up local configuration for deprovisioned instances] ****************

TASK [remove inventory configuration directory] ********************************
task path: /var/lib/jenkins/origin-ci-tool/7fa507a4bc591e690754be8f19277886f5151704/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:61
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2018-02-23 02:26:45.654344", 
    "path": "/var/lib/jenkins/jobs/test_pull_request_openshift_ansible_extended_conformance_install_update_containerized/workspace/.config/origin-ci-tool/inventory", 
    "state": "absent"
}

PLAY RECAP *********************************************************************
localhost                  : ok=8    changed=4    unreachable=0    failed=0   

+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPROVISION CLOUD RESOURCES [00h 00m 04s] ##########
Archiving artifacts
Recording test results
[WS-CLEANUP] Deleting project workspace...[WS-CLEANUP] done
Finished: SUCCESS