Console Output

Started by user OpenShift CI Robot
[EnvInject] - Loading node environment variables.
Building in workspace /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace
[WS-CLEANUP] Deleting project workspace...
[workspace] $ /bin/bash /tmp/jenkins7996773479211595355.sh
########## STARTING STAGE: INSTALL THE ORIGIN-CI-TOOL ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
++ readlink /var/lib/jenkins/origin-ci-tool/latest
+ latest=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
+ touch /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
+ cp /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin/activate /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
+ cat
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
+ mkdir -p /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
+ rm -rf /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool
+ oct configure ansible-client verbosity 2
Option verbosity updated to be 2.
+ oct configure aws-client keypair_name libra
Option keypair_name updated to be libra.
+ oct configure aws-client private_key_path /var/lib/jenkins/.ssh/devenv.pem
Option private_key_path updated to be /var/lib/jenkins/.ssh/devenv.pem.
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: INSTALL THE ORIGIN-CI-TOOL [00h 00m 01s] ##########
[workspace] $ /bin/bash /tmp/jenkins6157372002618388085.sh
########## STARTING STAGE: PROVISION CLOUD RESOURCES ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
+ oct provision remote all-in-one --os rhel --stage base --provider aws --discrete-ssh-config --name ci-kubernetes-machine-api-operator_683

PLAYBOOK: aws-up.yml ***********************************************************
2 plays in /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/aws-up.yml

PLAY [ensure we have the parameters necessary to bring up the AWS EC2 instance] ***

TASK [ensure all required variables are set] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/aws-up.yml:9
skipping: [localhost] => (item=origin_ci_inventory_dir)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.621748", 
    "item": "origin_ci_inventory_dir", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_keypair_name)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.624299", 
    "item": "origin_ci_aws_keypair_name", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_private_key_path)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.627473", 
    "item": "origin_ci_aws_private_key_path", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_region)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.631788", 
    "item": "origin_ci_aws_region", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_ami_tags)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.636218", 
    "item": "origin_ci_aws_ami_tags", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_instance_name)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.640500", 
    "item": "origin_ci_aws_instance_name", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_master_instance_type)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.643577", 
    "item": "origin_ci_aws_master_instance_type", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_identifying_tag_key)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.646663", 
    "item": "origin_ci_aws_identifying_tag_key", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_hostname)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.650951", 
    "item": "origin_ci_aws_hostname", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_ssh_config_strategy)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.654019", 
    "item": "origin_ci_ssh_config_strategy", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=openshift_schedulable)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.657149", 
    "item": "openshift_schedulable", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=openshift_node_labels)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.660271", 
    "item": "openshift_node_labels", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

TASK [ensure all required variables are set] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/aws-up.yml:27
skipping: [localhost] => (item=origin_ci_aws_master_subnet)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.696828", 
    "item": "origin_ci_aws_master_subnet", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_etcd_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.701857", 
    "item": "origin_ci_aws_etcd_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_node_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.707286", 
    "item": "origin_ci_aws_node_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_master_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.711569", 
    "item": "origin_ci_aws_master_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_master_external_elb_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.716984", 
    "item": "origin_ci_aws_master_external_elb_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_master_internal_elb_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.722437", 
    "item": "origin_ci_aws_master_internal_elb_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_router_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.726718", 
    "item": "origin_ci_aws_router_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_router_elb_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:28.732733", 
    "item": "origin_ci_aws_router_elb_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

PLAY [provision an AWS EC2 instance] *******************************************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

TASK [inventory : initialize the inventory directory] **************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:2
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:29.467958", 
    "gid": 995, 
    "group": "jenkins", 
    "mode": "0755", 
    "owner": "jenkins", 
    "path": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 6, 
    "state": "directory", 
    "uid": 997
}

TASK [inventory : add the nested group mapping] ********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:7
changed: [localhost] => {
    "changed": true, 
    "checksum": "18aaee00994df38cc3a63b635893175235331a9c", 
    "dest": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/nested_group_mappings", 
    "generated_timestamp": "2019-03-01 06:23:29.922274", 
    "gid": 995, 
    "group": "jenkins", 
    "md5sum": "b30c3226ea63efa3ff9c5e346c14a16e", 
    "mode": "0644", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 93, 
    "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1551439409.71-259511336220944/source", 
    "state": "file", 
    "uid": 997
}

TASK [inventory : initialize the OSEv3 group variables directory] **************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:12
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-03-01 06:23:30.085828", 
    "gid": 995, 
    "group": "jenkins", 
    "mode": "0755", 
    "owner": "jenkins", 
    "path": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/group_vars/OSEv3", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 6, 
    "state": "directory", 
    "uid": 997
}

TASK [inventory : initialize the host variables directory] *********************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:17
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-03-01 06:23:30.246667", 
    "gid": 995, 
    "group": "jenkins", 
    "mode": "0755", 
    "owner": "jenkins", 
    "path": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/host_vars", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 6, 
    "state": "directory", 
    "uid": 997
}

TASK [inventory : add the default Origin installation configuration] ***********
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:22
changed: [localhost] => {
    "changed": true, 
    "checksum": "4c06ba508f055c20f13426e8587342e8765a7b66", 
    "dest": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/group_vars/OSEv3/general.yml", 
    "generated_timestamp": "2019-03-01 06:23:30.535461", 
    "gid": 995, 
    "group": "jenkins", 
    "md5sum": "8aec71c75f7d512b278ae7c6f2959b12", 
    "mode": "0644", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 331, 
    "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1551439410.4-66583748456754/source", 
    "state": "file", 
    "uid": 997
}

TASK [aws-up : determine if we are inside AWS EC2] *****************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:2
changed: [localhost] => {
    "changed": true, 
    "cmd": [
        "curl", 
        "-s", 
        "http://instance-data.ec2.internal"
    ], 
    "delta": "0:00:00.011322", 
    "end": "2019-03-01 06:23:30.756792", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2019-03-01 06:23:30.772309", 
    "rc": 0, 
    "start": "2019-03-01 06:23:30.745470", 
    "stderr": [], 
    "stdout": [
        "1.0", 
        "2007-01-19", 
        "2007-03-01", 
        "2007-08-29", 
        "2007-10-10", 
        "2007-12-15", 
        "2008-02-01", 
        "2008-09-01", 
        "2009-04-04", 
        "2011-01-01", 
        "2011-05-01", 
        "2012-01-12", 
        "2014-02-25", 
        "2014-11-05", 
        "2015-10-20", 
        "2016-04-19", 
        "2016-06-30", 
        "2016-09-02", 
        "2018-03-28", 
        "2018-08-17", 
        "2018-09-24", 
        "latest"
    ], 
    "warnings": [
        "Consider using get_url or uri module rather than running curl"
    ]
}
 [WARNING]: Consider using get_url or uri module rather than running curl

TASK [aws-up : configure EC2 parameters for inventory when controlling from inside EC2] ***
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:7
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_destination_variable": "private_dns_name", 
        "origin_ci_aws_host_address_variable": "private_ip", 
        "origin_ci_aws_vpc_destination_variable": "private_ip_address"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:30.813226"
}

TASK [aws-up : determine where to put the AWS API cache] ***********************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:14
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_cache_dir": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ec2_cache"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:30.849704"
}

TASK [aws-up : ensure we have a place to put the AWS API cache] ****************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:18
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-03-01 06:23:31.012897", 
    "gid": 995, 
    "group": "jenkins", 
    "mode": "0755", 
    "owner": "jenkins", 
    "path": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ec2_cache", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 6, 
    "state": "directory", 
    "uid": 997
}

TASK [aws-up : place the EC2 dynamic inventory script] *************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:23
changed: [localhost] => {
    "changed": true, 
    "checksum": "625b8af723189db3b96ba0026d0f997a0025bc47", 
    "dest": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/ec2.py", 
    "generated_timestamp": "2019-03-01 06:23:31.300862", 
    "gid": 995, 
    "group": "jenkins", 
    "md5sum": "cac06c14065dac74904232b89d4ba24c", 
    "mode": "0755", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 63725, 
    "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1551439411.17-235843402155016/source", 
    "state": "file", 
    "uid": 997
}

TASK [aws-up : place the EC2 dynamic inventory configuration] ******************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:29
changed: [localhost] => {
    "changed": true, 
    "checksum": "31b75f62f1df61f6ccc63cd1d94118b2eba5f05f", 
    "dest": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/ec2.ini", 
    "generated_timestamp": "2019-03-01 06:23:31.587688", 
    "gid": 995, 
    "group": "jenkins", 
    "md5sum": "1f7cb329f0ee40069f5165570b434f46", 
    "mode": "0644", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 391, 
    "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1551439411.34-12208809648681/source", 
    "state": "file", 
    "uid": 997
}

TASK [aws-up : place the EC2 tag to group mappings] ****************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:34
changed: [localhost] => {
    "changed": true, 
    "checksum": "b4205a33dc73f62bd4f77f35d045cf8e09ae62b0", 
    "dest": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/tag_to_group_mappings", 
    "generated_timestamp": "2019-03-01 06:23:31.876358", 
    "gid": 995, 
    "group": "jenkins", 
    "md5sum": "bc3a567a1b6f342e1005182efc1b66be", 
    "mode": "0644", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 287, 
    "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1551439411.74-188274662564091/source", 
    "state": "file", 
    "uid": 997
}

TASK [aws-up : list available AMIs] ********************************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:40
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:35.589537", 
    "results": [
        {
            "ami_id": "ami-04f9b88b6b0571f20", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "encrypted": false, 
                    "size": 75, 
                    "snapshot_id": "snap-0655d2d962c590c8c", 
                    "volume_type": "gp2"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "encrypted": false, 
                    "size": 50, 
                    "snapshot_id": "snap-0d86ae865b17f4def", 
                    "volume_type": "gp2"
                }
            }, 
            "creationDate": "2018-06-26T12:22:31.000Z", 
            "description": "OpenShift Origin development AMI on rhel at the base stage.", 
            "hypervisor": "xen", 
            "is_public": false, 
            "location": "531415883065/ami_build_origin_int_rhel_base_758", 
            "name": "ami_build_origin_int_rhel_base_758", 
            "owner_id": "531415883065", 
            "platform": null, 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "available", 
            "tags": {
                "Name": "ami_build_origin_int_rhel_base_758", 
                "image_stage": "base", 
                "operating_system": "rhel", 
                "ready": "yes"
            }, 
            "virtualization_type": "hvm"
        }, 
        {
            "ami_id": "ami-0b77b87a37c3e662c", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "encrypted": false, 
                    "size": 75, 
                    "snapshot_id": "snap-02ec23d4818f2747e", 
                    "volume_type": "gp2"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "encrypted": false, 
                    "size": 50, 
                    "snapshot_id": "snap-0d8726e441d4ca329", 
                    "volume_type": "gp2"
                }
            }, 
            "creationDate": "2018-06-26T22:18:53.000Z", 
            "description": "OpenShift Origin development AMI on rhel at the base stage.", 
            "hypervisor": "xen", 
            "is_public": false, 
            "location": "531415883065/ami_build_origin_int_rhel_base_760", 
            "name": "ami_build_origin_int_rhel_base_760", 
            "owner_id": "531415883065", 
            "platform": null, 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "available", 
            "tags": {
                "Name": "ami_build_origin_int_rhel_base_760", 
                "image_stage": "base", 
                "operating_system": "rhel", 
                "ready": "yes"
            }, 
            "virtualization_type": "hvm"
        }
    ]
}

TASK [aws-up : choose appropriate AMIs for use] ********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:49
ok: [localhost] => (item={u'ami_id': u'ami-04f9b88b6b0571f20', u'root_device_type': u'ebs', u'description': u'OpenShift Origin development AMI on rhel at the base stage.', u'tags': {u'ready': u'yes', u'image_stage': u'base', u'Name': u'ami_build_origin_int_rhel_base_758', u'operating_system': u'rhel'}, u'hypervisor': u'xen', u'block_device_mapping': {u'/dev/sdb': {u'encrypted': False, u'snapshot_id': u'snap-0d86ae865b17f4def', u'delete_on_termination': True, u'volume_type': u'gp2', u'size': 50}, u'/dev/sda1': {u'encrypted': False, u'snapshot_id': u'snap-0655d2d962c590c8c', u'delete_on_termination': True, u'volume_type': u'gp2', u'size': 75}}, u'architecture': u'x86_64', u'owner_id': u'531415883065', u'platform': None, u'state': u'available', u'location': u'531415883065/ami_build_origin_int_rhel_base_758', u'is_public': False, u'creationDate': u'2018-06-26T12:22:31.000Z', u'root_device_name': u'/dev/sda1', u'virtualization_type': u'hvm', u'name': u'ami_build_origin_int_rhel_base_758'}) => {
    "ansible_facts": {
        "origin_ci_aws_ami_id_candidate": "ami-04f9b88b6b0571f20"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:35.638947", 
    "item": {
        "ami_id": "ami-04f9b88b6b0571f20", 
        "architecture": "x86_64", 
        "block_device_mapping": {
            "/dev/sda1": {
                "delete_on_termination": true, 
                "encrypted": false, 
                "size": 75, 
                "snapshot_id": "snap-0655d2d962c590c8c", 
                "volume_type": "gp2"
            }, 
            "/dev/sdb": {
                "delete_on_termination": true, 
                "encrypted": false, 
                "size": 50, 
                "snapshot_id": "snap-0d86ae865b17f4def", 
                "volume_type": "gp2"
            }
        }, 
        "creationDate": "2018-06-26T12:22:31.000Z", 
        "description": "OpenShift Origin development AMI on rhel at the base stage.", 
        "hypervisor": "xen", 
        "is_public": false, 
        "location": "531415883065/ami_build_origin_int_rhel_base_758", 
        "name": "ami_build_origin_int_rhel_base_758", 
        "owner_id": "531415883065", 
        "platform": null, 
        "root_device_name": "/dev/sda1", 
        "root_device_type": "ebs", 
        "state": "available", 
        "tags": {
            "Name": "ami_build_origin_int_rhel_base_758", 
            "image_stage": "base", 
            "operating_system": "rhel", 
            "ready": "yes"
        }, 
        "virtualization_type": "hvm"
    }
}
ok: [localhost] => (item={u'ami_id': u'ami-0b77b87a37c3e662c', u'root_device_type': u'ebs', u'description': u'OpenShift Origin development AMI on rhel at the base stage.', u'tags': {u'ready': u'yes', u'image_stage': u'base', u'Name': u'ami_build_origin_int_rhel_base_760', u'operating_system': u'rhel'}, u'hypervisor': u'xen', u'block_device_mapping': {u'/dev/sdb': {u'encrypted': False, u'snapshot_id': u'snap-0d8726e441d4ca329', u'delete_on_termination': True, u'volume_type': u'gp2', u'size': 50}, u'/dev/sda1': {u'encrypted': False, u'snapshot_id': u'snap-02ec23d4818f2747e', u'delete_on_termination': True, u'volume_type': u'gp2', u'size': 75}}, u'architecture': u'x86_64', u'owner_id': u'531415883065', u'platform': None, u'state': u'available', u'location': u'531415883065/ami_build_origin_int_rhel_base_760', u'is_public': False, u'creationDate': u'2018-06-26T22:18:53.000Z', u'root_device_name': u'/dev/sda1', u'virtualization_type': u'hvm', u'name': u'ami_build_origin_int_rhel_base_760'}) => {
    "ansible_facts": {
        "origin_ci_aws_ami_id_candidate": "ami-0b77b87a37c3e662c"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:35.646141", 
    "item": {
        "ami_id": "ami-0b77b87a37c3e662c", 
        "architecture": "x86_64", 
        "block_device_mapping": {
            "/dev/sda1": {
                "delete_on_termination": true, 
                "encrypted": false, 
                "size": 75, 
                "snapshot_id": "snap-02ec23d4818f2747e", 
                "volume_type": "gp2"
            }, 
            "/dev/sdb": {
                "delete_on_termination": true, 
                "encrypted": false, 
                "size": 50, 
                "snapshot_id": "snap-0d8726e441d4ca329", 
                "volume_type": "gp2"
            }
        }, 
        "creationDate": "2018-06-26T22:18:53.000Z", 
        "description": "OpenShift Origin development AMI on rhel at the base stage.", 
        "hypervisor": "xen", 
        "is_public": false, 
        "location": "531415883065/ami_build_origin_int_rhel_base_760", 
        "name": "ami_build_origin_int_rhel_base_760", 
        "owner_id": "531415883065", 
        "platform": null, 
        "root_device_name": "/dev/sda1", 
        "root_device_type": "ebs", 
        "state": "available", 
        "tags": {
            "Name": "ami_build_origin_int_rhel_base_760", 
            "image_stage": "base", 
            "operating_system": "rhel", 
            "ready": "yes"
        }, 
        "virtualization_type": "hvm"
    }
}

TASK [aws-up : determine which AMI to use] *************************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:55
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_ami_id": "ami-0b77b87a37c3e662c"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:35.685662"
}

TASK [aws-up : determine which subnets are available] **************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:60
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:36.704519", 
    "subnets": [
        {
            "availability_zone": "us-east-1c", 
            "available_ip_address_count": 4055, 
            "cidr_block": "172.18.16.0/20", 
            "default_for_az": "false", 
            "id": "subnet-8bdb5ac2", 
            "map_public_ip_on_launch": "true", 
            "state": "available", 
            "tags": {
                "Name": "devenv-subnet-2", 
                "origin_ci_aws_cluster_component": "master_subnet"
            }, 
            "vpc_id": "vpc-69705d0c"
        }, 
        {
            "availability_zone": "us-east-1d", 
            "available_ip_address_count": 4054, 
            "cidr_block": "172.18.0.0/20", 
            "default_for_az": "false", 
            "id": "subnet-cf57c596", 
            "map_public_ip_on_launch": "true", 
            "state": "available", 
            "tags": {
                "Name": "devenv-subnet-1", 
                "origin_ci_aws_cluster_component": "master_subnet"
            }, 
            "vpc_id": "vpc-69705d0c"
        }
    ]
}

TASK [aws-up : determine which subnets to use for the master] ******************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:67
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_master_subnet_ids": [
            "subnet-8bdb5ac2", 
            "subnet-cf57c596"
        ]
    }, 
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:36.748915"
}

TASK [aws-up : determine which security groups are available] ******************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:72
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:37.829932", 
    "security_groups": [
        {
            "description": "default VPC security group", 
            "group_id": "sg-7e73221a", 
            "group_name": "default", 
            "ip_permissions": [
                {
                    "ip_protocol": "-1", 
                    "ip_ranges": [], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "user_id_group_pairs": [
                        {
                            "group_id": "sg-7e73221a", 
                            "user_id": "531415883065"
                        }
                    ]
                }, 
                {
                    "from_port": 80, 
                    "ip_protocol": "tcp", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "54.241.19.245/32"
                        }, 
                        {
                            "cidr_ip": "97.65.119.184/29"
                        }, 
                        {
                            "cidr_ip": "107.20.219.35/32"
                        }, 
                        {
                            "cidr_ip": "108.166.48.153/32"
                        }, 
                        {
                            "cidr_ip": "212.199.177.64/27"
                        }, 
                        {
                            "cidr_ip": "212.72.208.162/32"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": 443, 
                    "user_id_group_pairs": []
                }, 
                {
                    "from_port": 53, 
                    "ip_protocol": "tcp", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "119.254.120.64/26"
                        }, 
                        {
                            "cidr_ip": "209.132.176.0/20"
                        }, 
                        {
                            "cidr_ip": "209.132.186.34/32"
                        }, 
                        {
                            "cidr_ip": "213.175.37.10/32"
                        }, 
                        {
                            "cidr_ip": "62.40.79.66/32"
                        }, 
                        {
                            "cidr_ip": "66.187.224.0/20"
                        }, 
                        {
                            "cidr_ip": "66.187.239.0/24"
                        }, 
                        {
                            "cidr_ip": "38.140.108.0/24"
                        }, 
                        {
                            "cidr_ip": "213.175.37.9/32"
                        }, 
                        {
                            "cidr_ip": "38.99.12.232/29"
                        }, 
                        {
                            "cidr_ip": "4.14.33.72/30"
                        }, 
                        {
                            "cidr_ip": "4.14.35.88/29"
                        }, 
                        {
                            "cidr_ip": "50.227.40.96/29"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": 8444, 
                    "user_id_group_pairs": []
                }, 
                {
                    "from_port": 22, 
                    "ip_protocol": "tcp", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "0.0.0.0/0"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": 22, 
                    "user_id_group_pairs": []
                }, 
                {
                    "from_port": 53, 
                    "ip_protocol": "udp", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "209.132.176.0/20"
                        }, 
                        {
                            "cidr_ip": "66.187.224.0/20"
                        }, 
                        {
                            "cidr_ip": "66.187.239.0/24"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": 53, 
                    "user_id_group_pairs": []
                }, 
                {
                    "from_port": 0, 
                    "ip_protocol": "udp", 
                    "ip_ranges": [], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": 65535, 
                    "user_id_group_pairs": [
                        {
                            "group_id": "sg-0d1a5377", 
                            "user_id": "531415883065"
                        }, 
                        {
                            "group_id": "sg-5875023f", 
                            "user_id": "531415883065"
                        }, 
                        {
                            "group_id": "sg-7e73221a", 
                            "user_id": "531415883065"
                        }, 
                        {
                            "group_id": "sg-e1760186", 
                            "user_id": "531415883065"
                        }
                    ]
                }, 
                {
                    "from_port": 3389, 
                    "ip_protocol": "tcp", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "0.0.0.0/0"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": 3389, 
                    "user_id_group_pairs": []
                }, 
                {
                    "from_port": -1, 
                    "ip_protocol": "icmp", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "0.0.0.0/0"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": -1, 
                    "user_id_group_pairs": []
                }
            ], 
            "ip_permissions_egress": [
                {
                    "ip_protocol": "-1", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "0.0.0.0/0"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "user_id_group_pairs": []
                }
            ], 
            "owner_id": "531415883065", 
            "tags": {
                "Name": "devenv-vpc", 
                "openshift_infra": "true", 
                "origin_ci_aws_cluster_component": "master_security_group"
            }, 
            "vpc_id": "vpc-69705d0c"
        }
    ]
}

TASK [aws-up : determine which security group to use] **************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:79
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_master_security_group_ids": [
            "sg-7e73221a"
        ]
    }, 
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:37.873885"
}

TASK [aws-up : provision an AWS EC2 instance] **********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:84
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-03-01 06:23:55.663342", 
    "instance_ids": [
        "i-015e1b4f6f2a05e43"
    ], 
    "instances": [
        {
            "ami_launch_index": "0", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-02bd0a734a29daa54"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-08bf4550ecb1eb195"
                }
            }, 
            "dns_name": "ec2-35-172-227-13.compute-1.amazonaws.com", 
            "ebs_optimized": false, 
            "groups": {
                "sg-7e73221a": "default"
            }, 
            "hypervisor": "xen", 
            "id": "i-015e1b4f6f2a05e43", 
            "image_id": "ami-0b77b87a37c3e662c", 
            "instance_type": "m4.xlarge", 
            "kernel": null, 
            "key_name": "libra", 
            "launch_time": "2019-03-01T11:23:39.000Z", 
            "placement": "us-east-1c", 
            "private_dns_name": "ip-172-18-25-241.ec2.internal", 
            "private_ip": "172.18.25.241", 
            "public_dns_name": "ec2-35-172-227-13.compute-1.amazonaws.com", 
            "public_ip": "35.172.227.13", 
            "ramdisk": null, 
            "region": "us-east-1", 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "running", 
            "state_code": 16, 
            "tags": {
                "Name": "ci-kubernetes-machine-api-operator_683", 
                "openshift_etcd": "", 
                "openshift_master": "", 
                "openshift_node": ""
            }, 
            "tenancy": "default", 
            "virtualization_type": "hvm"
        }
    ], 
    "tagged_instances": []
}

TASK [aws-up : determine the host address] *************************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:110
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_host": "172.18.25.241"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:55.704485"
}

TASK [aws-up : determine the default user to use for SSH] **********************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:114
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:55.737699", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

TASK [aws-up : determine the default user to use for SSH] **********************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:119
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_ssh_user": "origin"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:55.776111"
}

TASK [aws-up : update variables for the host] **********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:124
changed: [localhost] => {
    "changed": true, 
    "checksum": "e9390312fe66a0295233bbae9cf05dc16534f8b2", 
    "dest": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/host_vars/172.18.25.241.yml", 
    "generated_timestamp": "2019-03-01 06:23:56.076315", 
    "gid": 995, 
    "group": "jenkins", 
    "md5sum": "b88cad3eb1b1e9d1f461c19729e93da0", 
    "mode": "0644", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 754, 
    "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1551439435.94-11680551590323/source", 
    "state": "file", 
    "uid": 997
}

TASK [aws-up : determine where updated SSH configuration should go] ************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:141
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_ssh_config_files": [
            "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config"
        ]
    }, 
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:56.118636"
}

TASK [aws-up : determine where updated SSH configuration should go] ************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:146
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:23:56.185597", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

TASK [aws-up : ensure the targeted SSH configuration file exists] **************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:151
changed: [localhost] => (item=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config) => {
    "changed": true, 
    "dest": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config", 
    "generated_timestamp": "2019-03-01 06:23:56.383687", 
    "gid": 995, 
    "group": "jenkins", 
    "item": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config", 
    "mode": "0644", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 0, 
    "state": "file", 
    "uid": 997
}

TASK [aws-up : update the SSH configuration] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:157
changed: [localhost] => (item=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config) => {
    "changed": true, 
    "generated_timestamp": "2019-03-01 06:23:56.671042", 
    "item": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config", 
    "msg": "Block inserted"
}

TASK [aws-up : wait for SSH to be available] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:175
ok: [localhost] => {
    "changed": false, 
    "elapsed": 71, 
    "generated_timestamp": "2019-03-01 06:25:08.042949", 
    "path": null, 
    "port": 22, 
    "search_regex": null, 
    "state": "started"
}

PLAY RECAP *********************************************************************
localhost                  : ok=28   changed=13   unreachable=0    failed=0   

+ set +o xtrace
########## FINISHED STAGE: SUCCESS: PROVISION CLOUD RESOURCES [00h 01m 41s] ##########
[workspace] $ /bin/bash /tmp/jenkins5008358354793255484.sh
########## STARTING STAGE: FORWARD GCS CREDENTIALS TO REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
+ (( i = 0 ))
+ (( i < 10 ))
+ scp -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.config/gcloud/gcs-publisher-credentials.json openshiftdevel:/data/credentials.json
+ break
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: FORWARD GCS CREDENTIALS TO REMOTE HOST [00h 00m 02s] ##########
[workspace] $ /bin/bash /tmp/jenkins8449937278318899561.sh
########## STARTING STAGE: FORWARD PARAMETERS TO THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod o+rw /etc/environment
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''JOB_SPEC={"type":"presubmit","job":"ci-kubernetes-machine-api-operator","buildid":"1101442818523533312","prowjobid":"64e3c8c9-3c14-11e9-b6ae-0a58ac10f509","refs":{"org":"openshift","repo":"machine-api-operator","base_ref":"master","base_sha":"9b9b9f9446267e97478002dd222851919e53c58d","pulls":[{"number":220,"author":"ingvagabund","sha":"86992f9cb25c5b6df83c6c79bfcad3d60253350f"}]}}'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''buildId='\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''BUILD_ID=1101442818523533312'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''REPO_OWNER=openshift'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''REPO_NAME=machine-api-operator'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_BASE_REF=master'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_BASE_SHA=9b9b9f9446267e97478002dd222851919e53c58d'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_REFS=master:9b9b9f9446267e97478002dd222851919e53c58d,220:86992f9cb25c5b6df83c6c79bfcad3d60253350f'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_NUMBER=220'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_PULL_SHA=86992f9cb25c5b6df83c6c79bfcad3d60253350f'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''JOB_SPEC={"type":"presubmit","job":"ci-kubernetes-machine-api-operator","buildid":"1101442818523533312","prowjobid":"64e3c8c9-3c14-11e9-b6ae-0a58ac10f509","refs":{"org":"openshift","repo":"machine-api-operator","base_ref":"master","base_sha":"9b9b9f9446267e97478002dd222851919e53c58d","pulls":[{"number":220,"author":"ingvagabund","sha":"86992f9cb25c5b6df83c6c79bfcad3d60253350f"}]}}'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''BUILD_NUMBER=683'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''CLONEREFS_ARGS='\'' >> /etc/environment'
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: FORWARD PARAMETERS TO THE REMOTE HOST [00h 00m 05s] ##########
[workspace] $ /bin/bash /tmp/jenkins10959618891523853.sh
########## STARTING STAGE: SYNC REPOSITORIES ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ mktemp
+ script=/tmp/tmp.Fn8mg0jSYs
+ cat
+ chmod +x /tmp/tmp.Fn8mg0jSYs
+ scp -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.Fn8mg0jSYs openshiftdevel:/tmp/tmp.Fn8mg0jSYs
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.Fn8mg0jSYs"'
+ cd /home/origin
++ jq --compact-output '.buildid |= "683"'
+ JOB_SPEC='{"type":"presubmit","job":"ci-kubernetes-machine-api-operator","buildid":"683","prowjobid":"64e3c8c9-3c14-11e9-b6ae-0a58ac10f509","refs":{"org":"openshift","repo":"machine-api-operator","base_ref":"master","base_sha":"9b9b9f9446267e97478002dd222851919e53c58d","pulls":[{"number":220,"author":"ingvagabund","sha":"86992f9cb25c5b6df83c6c79bfcad3d60253350f"}]}}'
+ for image in ''\''registry.svc.ci.openshift.org/ci/clonerefs:latest'\''' ''\''registry.svc.ci.openshift.org/ci/initupload:latest'\'''
+ (( i = 0 ))
+ (( i < 5 ))
+ docker pull registry.svc.ci.openshift.org/ci/clonerefs:latest
Trying to pull repository registry.svc.ci.openshift.org/ci/clonerefs ... 
latest: Pulling from registry.svc.ci.openshift.org/ci/clonerefs
1160f4abea84: Pulling fs layer
be60dbe7622d: Pulling fs layer
d26b76701841: Pulling fs layer
4de8d7060c7b: Pulling fs layer
3a00cbb24bdb: Pulling fs layer
4de8d7060c7b: Waiting
3a00cbb24bdb: Waiting
1160f4abea84: Download complete
be60dbe7622d: Verifying Checksum
be60dbe7622d: Download complete
3a00cbb24bdb: Verifying Checksum
3a00cbb24bdb: Download complete
4de8d7060c7b: Verifying Checksum
4de8d7060c7b: Download complete
d26b76701841: Verifying Checksum
d26b76701841: Download complete
1160f4abea84: Pull complete
be60dbe7622d: Pull complete
d26b76701841: Pull complete
4de8d7060c7b: Pull complete
3a00cbb24bdb: Pull complete
Digest: sha256:a6cd3068468d4891147c1c6cb386a6e485aa0aea1c5fd8848c94afb5756000cb
Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/clonerefs:latest
+ break
+ for image in ''\''registry.svc.ci.openshift.org/ci/clonerefs:latest'\''' ''\''registry.svc.ci.openshift.org/ci/initupload:latest'\'''
+ (( i = 0 ))
+ (( i < 5 ))
+ docker pull registry.svc.ci.openshift.org/ci/initupload:latest
Trying to pull repository registry.svc.ci.openshift.org/ci/initupload ... 
latest: Pulling from registry.svc.ci.openshift.org/ci/initupload
a073c86ecf9e: Pulling fs layer
cc3fc741b1a9: Pulling fs layer
e2be10d718b3: Pulling fs layer
8e5b170ec95b: Pulling fs layer
8e5b170ec95b: Waiting
cc3fc741b1a9: Verifying Checksum
cc3fc741b1a9: Download complete
a073c86ecf9e: Verifying Checksum
a073c86ecf9e: Download complete
e2be10d718b3: Verifying Checksum
e2be10d718b3: Download complete
8e5b170ec95b: Verifying Checksum
8e5b170ec95b: Download complete
a073c86ecf9e: Pull complete
cc3fc741b1a9: Pull complete
e2be10d718b3: Pull complete
8e5b170ec95b: Pull complete
Digest: sha256:78272c2b96f467047821879ea05400fc30f7bc2981102be3e58361dc0c78b493
Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/initupload:latest
+ break
+ clonerefs_args=
+ docker run -v /data:/data:z registry.svc.ci.openshift.org/ci/clonerefs:latest --src-root=/data --log=/data/clone.json --repo=openshift,machine-api-operator=master:9b9b9f9446267e97478002dd222851919e53c58d,220:86992f9cb25c5b6df83c6c79bfcad3d60253350f
{"component":"clonerefs","level":"info","msg":"Cloning refs","refs":{"org":"openshift","repo":"machine-api-operator","base_ref":"master","base_sha":"9b9b9f9446267e97478002dd222851919e53c58d","pulls":[{"number":220,"author":"","sha":"86992f9cb25c5b6df83c6c79bfcad3d60253350f"}]},"time":"2019-03-01T11:26:18Z"}
{"command":"mkdir -p /data/src/github.com/openshift/machine-api-operator","component":"clonerefs","error":null,"level":"info","msg":"Ran command","output":"","time":"2019-03-01T11:26:18Z"}
{"command":"git init","component":"clonerefs","error":null,"level":"info","msg":"Ran command","output":"Initialized empty Git repository in /data/src/github.com/openshift/machine-api-operator/.git/\n","time":"2019-03-01T11:26:18Z"}
{"command":"git config user.name ci-robot","component":"clonerefs","error":null,"level":"info","msg":"Ran command","output":"","time":"2019-03-01T11:26:18Z"}
{"command":"git config user.email ci-robot@k8s.io","component":"clonerefs","error":null,"level":"info","msg":"Ran command","output":"","time":"2019-03-01T11:26:18Z"}
{"command":"git fetch https://github.com/openshift/machine-api-operator.git --tags --prune","component":"clonerefs","error":null,"level":"info","msg":"Ran command","output":"From https://github.com/openshift/machine-api-operator\n * branch            HEAD       -\u003e FETCH_HEAD\n * [new tag]         v0.1.0     -\u003e v0.1.0\n * [new tag]         v0.2.0     -\u003e v0.2.0\n","time":"2019-03-01T11:26:19Z"}
{"command":"git fetch https://github.com/openshift/machine-api-operator.git master","component":"clonerefs","error":null,"level":"info","msg":"Ran command","output":"From https://github.com/openshift/machine-api-operator\n * branch            master     -\u003e FETCH_HEAD\n","time":"2019-03-01T11:26:19Z"}
{"command":"git checkout 9b9b9f9446267e97478002dd222851919e53c58d","component":"clonerefs","error":null,"level":"info","msg":"Ran command","output":"Note: checking out '9b9b9f9446267e97478002dd222851919e53c58d'.\n\nYou are in 'detached HEAD' state. You can look around, make experimental\nchanges and commit them, and you can discard any commits you make in this\nstate without impacting any branches by performing another checkout.\n\nIf you want to create a new branch to retain commits you create, you may\ndo so (now or later) by using -b with the checkout command again. Example:\n\n  git checkout -b \u003cnew-branch-name\u003e\n\nHEAD is now at 9b9b9f9... Merge pull request #237 from frobware/ensure-node-annotations-is-non-nil\n","time":"2019-03-01T11:26:20Z"}
{"command":"git branch --force master 9b9b9f9446267e97478002dd222851919e53c58d","component":"clonerefs","error":null,"level":"info","msg":"Ran command","output":"","time":"2019-03-01T11:26:20Z"}
{"command":"git checkout master","component":"clonerefs","error":null,"level":"info","msg":"Ran command","output":"Switched to branch 'master'\n","time":"2019-03-01T11:26:20Z"}
{"command":"git fetch https://github.com/openshift/machine-api-operator.git pull/220/head","component":"clonerefs","error":null,"level":"info","msg":"Ran command","output":"From https://github.com/openshift/machine-api-operator\n * branch            refs/pull/220/head -\u003e FETCH_HEAD\n","time":"2019-03-01T11:26:20Z"}
{"command":"git merge --no-ff 86992f9cb25c5b6df83c6c79bfcad3d60253350f","component":"clonerefs","error":null,"level":"info","msg":"Ran command","output":"Merge made by the 'recursive' strategy.\n Gopkg.lock                                         |  33 +++--\n Makefile                                           |   7 +\n README.md                                          |  32 ++++\n cmd/machine-api-operator/main.go                   |   9 +-\n ...er-version-operator_01_clusteroperator.crd.yaml |  42 ++++++\n config/kubemark-install-config.yaml                |   9 ++\n config/kubemark.yaml                               |  80 ++++++++++\n config/kubemark_rbac.yaml                          |  69 +++++++++\n config/kustomization.yaml                          |   6 +\n ...0_machine-api-operator_01_images.configmap.yaml |   2 +-\n ...0000_30_machine-api-operator_09_deployment.yaml |   4 +\n kustomization.yaml                                 |  24 +++\n pkg/operator/config.go                             |  11 ++\n vendor/github.com/aws/aws-sdk-go/aws/version.go    |   2 +-\n .../openshift/api/config/v1/types_features.go      |  45 +++---\n .../api/config/v1/zz_generated.deepcopy.go         |  26 ++++\n .../pkg/e2e/autoscaler/autoscaler.go               |   1 +\n .../e2e/operators/cluster-autoscaler-operator.go   |  16 +-\n .../x/crypto/internal/chacha20/chacha_s390x.go     |  11 +-\n .../x/crypto/internal/chacha20/chacha_s390x.s      |  23 ---\n vendor/golang.org/x/crypto/poly1305/sum_s390x.go   |  17 +--\n vendor/golang.org/x/crypto/poly1305/sum_s390x.s    |  22 ---\n .../golang.org/x/crypto/poly1305/sum_vmsl_s390x.s  |  22 ---\n vendor/golang.org/x/sys/cpu/byteorder.go           |  30 ++++\n vendor/golang.org/x/sys/cpu/cpu.go                 | 126 ++++++++++++++++\n vendor/golang.org/x/sys/cpu/cpu_arm.go             |   9 ++\n vendor/golang.org/x/sys/cpu/cpu_gc_s390x.go        |  21 +++\n vendor/golang.org/x/sys/cpu/cpu_gc_x86.go          |  16 ++\n vendor/golang.org/x/sys/cpu/cpu_gccgo.c            |  43 ++++++\n vendor/golang.org/x/sys/cpu/cpu_gccgo.go           |  26 ++++\n vendor/golang.org/x/sys/cpu/cpu_gccgo_s390x.go     |  22 +++\n vendor/golang.org/x/sys/cpu/cpu_linux.go           |  59 ++++++++\n vendor/golang.org/x/sys/cpu/cpu_linux_arm64.go     |  67 +++++++++\n vendor/golang.org/x/sys/cpu/cpu_linux_ppc64x.go    |  33 +++++\n vendor/golang.org/x/sys/cpu/cpu_linux_s390x.go     | 161 +++++++++++++++++++++\n vendor/golang.org/x/sys/cpu/cpu_mips64x.go         |  11 ++\n vendor/golang.org/x/sys/cpu/cpu_mipsx.go           |  11 ++\n vendor/golang.org/x/sys/cpu/cpu_other_arm64.go     |  11 ++\n vendor/golang.org/x/sys/cpu/cpu_other_ppc64x.go    |  12 ++\n vendor/golang.org/x/sys/cpu/cpu_s390x.s            |  57 ++++++++\n vendor/golang.org/x/sys/cpu/cpu_x86.go             |  59 ++++++++\n vendor/golang.org/x/sys/cpu/cpu_x86.s              |  27 ++++\n 42 files changed, 1188 insertions(+), 126 deletions(-)\n create mode 100644 config/0000_00_cluster-version-operator_01_clusteroperator.crd.yaml\n create mode 100644 config/kubemark-install-config.yaml\n create mode 100644 config/kubemark.yaml\n create mode 100644 config/kubemark_rbac.yaml\n create mode 100644 config/kustomization.yaml\n create mode 100644 kustomization.yaml\n create mode 100644 vendor/golang.org/x/sys/cpu/byteorder.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_arm.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_gc_s390x.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_gc_x86.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_gccgo.c\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_gccgo.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_gccgo_s390x.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_linux.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_linux_arm64.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_linux_ppc64x.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_linux_s390x.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_mips64x.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_mipsx.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_other_arm64.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_other_ppc64x.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_s390x.s\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_x86.go\n create mode 100644 vendor/golang.org/x/sys/cpu/cpu_x86.s\n","time":"2019-03-01T11:26:20Z"}
{"command":"git submodule update --init --recursive","component":"clonerefs","error":null,"level":"info","msg":"Ran command","output":"","time":"2019-03-01T11:26:20Z"}
{"component":"clonerefs","level":"info","msg":"Finished cloning refs","time":"2019-03-01T11:26:20Z"}
+ docker run -e 'JOB_SPEC={"type":"presubmit","job":"ci-kubernetes-machine-api-operator","buildid":"683","prowjobid":"64e3c8c9-3c14-11e9-b6ae-0a58ac10f509","refs":{"org":"openshift","repo":"machine-api-operator","base_ref":"master","base_sha":"9b9b9f9446267e97478002dd222851919e53c58d","pulls":[{"number":220,"author":"ingvagabund","sha":"86992f9cb25c5b6df83c6c79bfcad3d60253350f"}]}}' -v /data:/data:z registry.svc.ci.openshift.org/ci/initupload:latest --clone-log=/data/clone.json --dry-run=false --gcs-path=gs://origin-ci-test --gcs-credentials-file=/data/credentials.json --path-strategy=single --default-org=openshift --default-repo=origin
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/started.json","level":"info","msg":"Queued for upload","time":"2019-03-01T11:26:22Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/clone-log.txt","level":"info","msg":"Queued for upload","time":"2019-03-01T11:26:22Z"}
{"component":"initupload","dest":"pr-logs/directory/ci-kubernetes-machine-api-operator/683.txt","level":"info","msg":"Queued for upload","time":"2019-03-01T11:26:22Z"}
{"component":"initupload","dest":"pr-logs/directory/ci-kubernetes-machine-api-operator/latest-build.txt","level":"info","msg":"Queued for upload","time":"2019-03-01T11:26:22Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/latest-build.txt","level":"info","msg":"Queued for upload","time":"2019-03-01T11:26:22Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/clone-records.json","level":"info","msg":"Queued for upload","time":"2019-03-01T11:26:22Z"}
{"component":"initupload","dest":"pr-logs/directory/ci-kubernetes-machine-api-operator/latest-build.txt","level":"info","msg":"Finished upload","time":"2019-03-01T11:26:23Z"}
{"component":"initupload","dest":"pr-logs/directory/ci-kubernetes-machine-api-operator/683.txt","level":"info","msg":"Finished upload","time":"2019-03-01T11:26:23Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/started.json","level":"info","msg":"Finished upload","time":"2019-03-01T11:26:23Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/clone-records.json","level":"info","msg":"Finished upload","time":"2019-03-01T11:26:23Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/clone-log.txt","level":"info","msg":"Finished upload","time":"2019-03-01T11:26:23Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/latest-build.txt","level":"info","msg":"Finished upload","time":"2019-03-01T11:26:23Z"}
{"component":"initupload","level":"info","msg":"Finished upload to GCS","time":"2019-03-01T11:26:23Z"}
+ sudo chmod -R a+rwX /data
+ sudo chown -R origin:origin-git /data
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: SYNC REPOSITORIES [00h 01m 10s] ##########
[workspace] $ /bin/bash /tmp/jenkins4234416549632705646.sh
########## STARTING STAGE: FORWARD PARAMETERS TO THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod o+rw /etc/environment
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''JOB_NAME=ci-kubernetes-machine-api-operator'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''BUILD_NUMBER=683'\'' >> /etc/environment'
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: FORWARD PARAMETERS TO THE REMOTE HOST [00h 00m 01s] ##########
[workspace] $ /bin/bash /tmp/jenkins4392028412178196173.sh
########## STARTING STAGE: UPLOAD THE DEFAULT AWS CREDENTIASL ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'mkdir ~/.aws'
+ scp -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.aws/credentials 'openshiftdevel:~/.aws'
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: UPLOAD THE DEFAULT AWS CREDENTIASL [00h 00m 01s] ##########
[workspace] $ /bin/bash /tmp/jenkins5107883664498133579.sh
########## STARTING STAGE: INSTALL MINIKUBE ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ mktemp
+ script=/tmp/tmp.VqoYIf2YRx
+ cat
+ chmod +x /tmp/tmp.VqoYIf2YRx
+ scp -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.VqoYIf2YRx openshiftdevel:/tmp/tmp.VqoYIf2YRx
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.VqoYIf2YRx"'
+ cd /home/origin
+ curl -Lo minikube https://storage.googleapis.com/minikube/releases/v0.30.0/minikube-linux-amd64
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100 40.3M  100 40.3M    0     0  36.6M      0  0:00:01  0:00:01 --:--:-- 36.7M
+ chmod +x minikube
+ sudo mv minikube /usr/bin/
+ curl -Lo kubectl https://storage.googleapis.com/kubernetes-release/release/v1.10.0/bin/linux/amd64/kubectl
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100 51.7M  100 51.7M    0     0  63.6M      0 --:--:-- --:--:-- --:--:-- 63.6M
+ chmod +x kubectl
+ sudo mv kubectl /usr/bin/
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: INSTALL MINIKUBE [00h 00m 03s] ##########
[workspace] $ /bin/bash /tmp/jenkins6976017368153065992.sh
########## STARTING STAGE: DEPLOY KUBERNETES ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ mktemp
+ script=/tmp/tmp.0uL2PsLEwx
+ cat
+ chmod +x /tmp/tmp.0uL2PsLEwx
+ scp -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.0uL2PsLEwx openshiftdevel:/tmp/tmp.0uL2PsLEwx
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.0uL2PsLEwx"'
+ cd /home/origin
+ sudo setenforce 0
+ sudo minikube start --vm-driver=none --extra-config=kubelet.cgroup-driver=systemd --kubernetes-version v1.12.0 --v 5
There is a newer version of minikube available (v0.34.1).  Download it here:
https://github.com/kubernetes/minikube/releases/tag/v0.34.1

To disable this notification, run the following:
minikube config set WantUpdateNotification false
Starting local Kubernetes v1.12.0 cluster...
Starting VM...
Creating CA: /root/.minikube/certs/ca.pem
Creating client certificate: /root/.minikube/certs/cert.pem
Getting VM IP address...
Moving files into cluster...
Downloading kubeadm v1.12.0
Downloading kubelet v1.12.0
Finished Downloading kubeadm v1.12.0
Finished Downloading kubelet v1.12.0
Setting up certs...
Connecting to cluster...
Setting up kubeconfig...
Starting cluster components...
Kubectl is now configured to use the cluster.
===================
WARNING: IT IS RECOMMENDED NOT TO RUN THE NONE DRIVER ON PERSONAL WORKSTATIONS
	The 'none' driver will run an insecure kubernetes apiserver as root that may leave the host vulnerable to CSRF attacks

When using the none driver, the kubectl config and credentials generated will be root owned and will appear in the root home directory.
You will need to move the files to the appropriate location and then set the correct permissions.  An example of this is below:

	sudo mv /root/.kube $HOME/.kube # this will write over any previous configuration
	sudo chown -R $USER $HOME/.kube
	sudo chgrp -R $USER $HOME/.kube

	sudo mv /root/.minikube $HOME/.minikube # this will write over any previous configuration
	sudo chown -R $USER $HOME/.minikube
	sudo chgrp -R $USER $HOME/.minikube

This can also be done automatically by setting the env var CHANGE_MINIKUBE_NONE_USER=true
Loading cached images from config file.
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPLOY KUBERNETES [00h 01m 04s] ##########
[workspace] $ /bin/bash /tmp/jenkins5215576142987496976.sh
########## STARTING STAGE: BUILD THE MACHINE API OPERATOR ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ mktemp
+ script=/tmp/tmp.I0d9BRhDle
+ cat
+ chmod +x /tmp/tmp.I0d9BRhDle
+ scp -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.I0d9BRhDle openshiftdevel:/tmp/tmp.I0d9BRhDle
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.I0d9BRhDle"'
+ cd /home/origin
+ export GOPATH=/data
+ GOPATH=/data
+ cd /data/src/github.com/openshift/machine-api-operator
+ make build
docker run --rm -v "/data/src/github.com/openshift/machine-api-operator":/go/src/github.com/openshift/machine-api-operator:Z -w /go/src/github.com/openshift/machine-api-operator golang:1.10 ./hack/go-build.sh machine-api-operator
Unable to find image 'golang:1.10' locally
Trying to pull repository registry.access.redhat.com/golang ... 
Trying to pull repository docker.io/library/golang ... 
1.10: Pulling from docker.io/library/golang
741437d97401: Pulling fs layer
34d8874714d7: Pulling fs layer
0a108aa26679: Pulling fs layer
7f0334c36886: Pulling fs layer
d35724ed4672: Pulling fs layer
c0eaf021aeaf: Pulling fs layer
d3d9c96611f1: Pulling fs layer
d35724ed4672: Waiting
c0eaf021aeaf: Waiting
d3d9c96611f1: Waiting
7f0334c36886: Waiting
0a108aa26679: Verifying Checksum
0a108aa26679: Download complete
34d8874714d7: Verifying Checksum
34d8874714d7: Download complete
741437d97401: Verifying Checksum
741437d97401: Download complete
7f0334c36886: Verifying Checksum
7f0334c36886: Download complete
d35724ed4672: Verifying Checksum
d35724ed4672: Download complete
d3d9c96611f1: Verifying Checksum
d3d9c96611f1: Download complete
c0eaf021aeaf: Verifying Checksum
c0eaf021aeaf: Download complete
741437d97401: Pull complete
34d8874714d7: Pull complete
0a108aa26679: Pull complete
7f0334c36886: Pull complete
d35724ed4672: Pull complete
c0eaf021aeaf: Pull complete
d3d9c96611f1: Pull complete
Digest: sha256:6d5e79878a3e4f1b30b7aa4d24fb6ee6184e905a9b172fc72593935633be4c46
Status: Downloaded newer image for docker.io/golang:1.10
Using version from git...
Building github.com/openshift/machine-api-operator/cmd/machine-api-operator (v0.1.0-235-gaff61672)
docker run --rm -v "/data/src/github.com/openshift/machine-api-operator":/go/src/github.com/openshift/machine-api-operator:Z -w /go/src/github.com/openshift/machine-api-operator golang:1.10 ./hack/go-build.sh nodelink-controller
Using version from git...
Building github.com/openshift/machine-api-operator/cmd/nodelink-controller (v0.1.0-235-gaff61672)
docker run --rm -v "/data/src/github.com/openshift/machine-api-operator":/go/src/github.com/openshift/machine-api-operator:Z -w /go/src/github.com/openshift/machine-api-operator golang:1.10 ./hack/go-build.sh machine-healthcheck
Using version from git...
Building github.com/openshift/machine-api-operator/cmd/machine-healthcheck (v0.1.0-235-gaff61672)
+ go get -u github.com/openshift/imagebuilder/cmd/imagebuilder
+ sudo mv /data/bin/imagebuilder /usr/bin
++ git describe --always --abbrev=7
+ sudo imagebuilder -t docker.io/machine-api-operator:v0.1.0-235-gaff6167 .
--> Image registry.svc.ci.openshift.org/openshift/release:golang-1.10 was not found, pulling ...
--> Pulled 0/2 layers, 12% complete
--> Pulled 1/2 layers, 64% complete
--> Pulled 2/2 layers, 100% complete
--> Extracting
--> FROM registry.svc.ci.openshift.org/openshift/release:golang-1.10 as builder
--> WORKDIR /go/src/github.com/openshift/machine-api-operator
--> COPY . .
--> RUN NO_DOCKER=1 make build
./hack/go-build.sh machine-api-operator
Using version from git...
Building github.com/openshift/machine-api-operator/cmd/machine-api-operator (v0.1.0-235-gaff61672)
./hack/go-build.sh nodelink-controller
Using version from git...
Building github.com/openshift/machine-api-operator/cmd/nodelink-controller (v0.1.0-235-gaff61672)
./hack/go-build.sh machine-healthcheck
Using version from git...
Building github.com/openshift/machine-api-operator/cmd/machine-healthcheck (v0.1.0-235-gaff61672)
--> Image registry.svc.ci.openshift.org/openshift/origin-v4.0:base was not found, pulling ...
--> Pulled 2/4 layers, 50% complete
--> Pulled 3/4 layers, 75% complete
--> Pulled 4/4 layers, 100% complete
--> Extracting
--> FROM registry.svc.ci.openshift.org/openshift/origin-v4.0:base as 1
--> COPY --from=builder /go/src/github.com/openshift/machine-api-operator/owned-manifests owned-manifests
--> COPY --from=builder /go/src/github.com/openshift/machine-api-operator/install manifests
--> COPY --from=builder /go/src/github.com/openshift/machine-api-operator/bin/machine-api-operator .
--> COPY --from=builder /go/src/github.com/openshift/machine-api-operator/bin/nodelink-controller .
--> COPY --from=builder /go/src/github.com/openshift/machine-api-operator/bin/machine-healthcheck .
--> LABEL io.openshift.release.operator true
--> Committing changes to docker.io/machine-api-operator:v0.1.0-235-gaff6167 ...
--> Done
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: BUILD THE MACHINE API OPERATOR [00h 04m 39s] ##########
[workspace] $ /bin/bash /tmp/jenkins5008406683596265610.sh
########## STARTING STAGE: DEPLOY AND TEST THE MACHINE API OPERATOR ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ mktemp
+ script=/tmp/tmp.UlKECsva5i
+ cat
+ chmod +x /tmp/tmp.UlKECsva5i
+ scp -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.UlKECsva5i openshiftdevel:/tmp/tmp.UlKECsva5i
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.UlKECsva5i"'
+ cd /home/origin
+ export GOPATH=/data
+ GOPATH=/data
+ cd /data/src/github.com/openshift/machine-api-operator
+ make build-integration
Building integration test binary...
mkdir -p bin
docker run --rm -v "/data/src/github.com/openshift/machine-api-operator":/go/src/github.com/openshift/machine-api-operator:Z -w /go/src/github.com/openshift/machine-api-operator golang:1.10 go build  -o bin/integration github.com/openshift/machine-api-operator/test/integration
+ sudo pip install awscli
Collecting awscli
  Downloading https://files.pythonhosted.org/packages/aa/ea/cb62728e9b38f9d8c620d60815f8dd54ca015f6b9af8f5a3d03d9b2e3c64/awscli-1.16.115-py2.py3-none-any.whl (1.4MB)
Collecting botocore==1.12.105 (from awscli)
  Downloading https://files.pythonhosted.org/packages/cf/ce/acc9013dee20fc94c9b9ae121f5b7b342a206f0d577be1e5c6129811194a/botocore-1.12.105-py2.py3-none-any.whl (5.3MB)
Collecting colorama<=0.3.9,>=0.2.5 (from awscli)
  Downloading https://files.pythonhosted.org/packages/db/c8/7dcf9dbcb22429512708fe3a547f8b6101c0d02137acbd892505aee57adf/colorama-0.3.9-py2.py3-none-any.whl
Collecting rsa<=3.5.0,>=3.1.2 (from awscli)
  Downloading https://files.pythonhosted.org/packages/e1/ae/baedc9cb175552e95f3395c43055a6a5e125ae4d48a1d7a924baca83e92e/rsa-3.4.2-py2.py3-none-any.whl (46kB)
Collecting docutils>=0.10 (from awscli)
  Downloading https://files.pythonhosted.org/packages/50/09/c53398e0005b11f7ffb27b7aa720c617aba53be4fb4f4f3f06b9b5c60f28/docutils-0.14-py2-none-any.whl (543kB)
Collecting s3transfer<0.3.0,>=0.2.0 (from awscli)
  Downloading https://files.pythonhosted.org/packages/d7/de/5737f602e22073ecbded7a0c590707085e154e32b68d86545dcc31004c02/s3transfer-0.2.0-py2.py3-none-any.whl (69kB)
Requirement already satisfied (use --upgrade to upgrade): PyYAML<=3.13,>=3.10 in /usr/lib64/python2.7/site-packages (from awscli)
Requirement already satisfied (use --upgrade to upgrade): jmespath<1.0.0,>=0.7.1 in /usr/lib/python2.7/site-packages (from botocore==1.12.105->awscli)
Collecting python-dateutil<3.0.0,>=2.1; python_version >= "2.7" (from botocore==1.12.105->awscli)
  Downloading https://files.pythonhosted.org/packages/41/17/c62faccbfbd163c7f57f3844689e3a78bae1f403648a6afb1d0866d87fbb/python_dateutil-2.8.0-py2.py3-none-any.whl (226kB)
Collecting urllib3<1.25,>=1.20; python_version == "2.7" (from botocore==1.12.105->awscli)
  Downloading https://files.pythonhosted.org/packages/62/00/ee1d7de624db8ba7090d1226aebefab96a2c71cd5cfa7629d6ad3f61b79e/urllib3-1.24.1-py2.py3-none-any.whl (118kB)
Requirement already satisfied (use --upgrade to upgrade): pyasn1>=0.1.3 in /usr/lib/python2.7/site-packages (from rsa<=3.5.0,>=3.1.2->awscli)
Collecting futures<4.0.0,>=2.2.0; python_version == "2.6" or python_version == "2.7" (from s3transfer<0.3.0,>=0.2.0->awscli)
  Downloading https://files.pythonhosted.org/packages/2d/99/b2c4e9d5a30f6471e410a146232b4118e697fa3ffc06d6a65efde84debd0/futures-3.2.0-py2-none-any.whl
Requirement already satisfied (use --upgrade to upgrade): six>=1.5 in /usr/lib/python2.7/site-packages (from python-dateutil<3.0.0,>=2.1; python_version >= "2.7"->botocore==1.12.105->awscli)
Installing collected packages: docutils, python-dateutil, urllib3, botocore, colorama, rsa, futures, s3transfer, awscli
  Found existing installation: python-dateutil 1.5
    Uninstalling python-dateutil-1.5:
      Successfully uninstalled python-dateutil-1.5
  Found existing installation: urllib3 1.10.2
    Uninstalling urllib3-1.10.2:
      Successfully uninstalled urllib3-1.10.2
Successfully installed awscli-1.16.115 botocore-1.12.105 colorama-0.3.9 docutils-0.14 futures-3.2.0 python-dateutil-2.8.0 rsa-3.4.2 s3transfer-0.2.0 urllib3-1.24.1
You are using pip version 8.1.2, however version 19.0.3 is available.
You should consider upgrading via the 'pip install --upgrade pip' command.
+ curl https://releases.hashicorp.com/terraform/0.11.8/terraform_0.11.8_linux_amd64.zip -o terraform_0.11.8_linux_amd64.zip
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100 17.0M  100 17.0M    0     0  55.9M      0 --:--:-- --:--:-- --:--:-- 56.0M
+ unzip terraform_0.11.8_linux_amd64.zip
Archive:  terraform_0.11.8_linux_amd64.zip
  inflating: terraform               
+ sudo cp ./terraform /usr/bin/.
+ set +x
+ sudo cp -r /home/origin/.aws /root/.
++ git describe --always --abbrev=7
+ sudo -E ./bin/integration --kubeconfig /root/.kube/config --mao-image machine-api-operator:v0.1.0-235-gaff6167 --cluster-id mao-683-pr-220
Initializing modules...
- module.vpc
  Found version 1.58.0 of terraform-aws-modules/vpc/aws on registry.terraform.io
  Getting source "terraform-aws-modules/vpc/aws"

Initializing provider plugins...
- Checking for available provider plugins on https://releases.hashicorp.com...
- Downloading plugin for provider "aws" (2.0.0)...

The following providers do not have any version constraints in configuration,
so the latest version was installed.

To prevent automatic upgrades to new major versions that may contain breaking
changes, it is recommended to add version = "..." constraints to the
corresponding provider blocks in configuration, with the constraint strings
suggested below.

* provider.aws: version = "~> 2.0"

Terraform has been successfully initialized!

You may now begin working with Terraform. Try running "terraform plan" to see
any changes that are required for your infrastructure. All Terraform commands
should now work.

If you ever set or change modules or backend configuration for Terraform,
rerun this command to reinitialize your working directory. If you forget, other
commands will detect it and remind you to do so if necessary.
aws_iam_role.role: Creating...
  arn:                   "" => "<computed>"
  assume_role_policy:    "" => "{\n    \"Version\": \"2012-10-17\",\n    \"Statement\": [\n        {\n            \"Action\": \"sts:AssumeRole\",\n            \"Principal\": {\n               \"Service\": \"ec2.amazonaws.com\"\n            },\n            \"Effect\": \"Allow\",\n            \"Sid\": \"\"\n        }\n    ]\n}\n"
  create_date:           "" => "<computed>"
  force_detach_policies: "" => "false"
  max_session_duration:  "" => "3600"
  name:                  "" => "mao-683-pr-220-role"
  path:                  "" => "/"
  unique_id:             "" => "<computed>"
module.vpc.aws_vpc.this: Creating...
  arn:                              "" => "<computed>"
  assign_generated_ipv6_cidr_block: "" => "false"
  cidr_block:                       "" => "10.0.0.0/16"
  default_network_acl_id:           "" => "<computed>"
  default_route_table_id:           "" => "<computed>"
  default_security_group_id:        "" => "<computed>"
  dhcp_options_id:                  "" => "<computed>"
  enable_classiclink:               "" => "<computed>"
  enable_classiclink_dns_support:   "" => "<computed>"
  enable_dns_hostnames:             "" => "false"
  enable_dns_support:               "" => "true"
  instance_tenancy:                 "" => "default"
  ipv6_association_id:              "" => "<computed>"
  ipv6_cidr_block:                  "" => "<computed>"
  main_route_table_id:              "" => "<computed>"
  owner_id:                         "" => "<computed>"
  tags.%:                           "" => "3"
  tags.Environment:                 "" => "dev"
  tags.Name:                        "" => "mao-683-pr-220"
  tags.Owner:                       "" => "user"
aws_iam_role.role: Creation complete after 0s (ID: mao-683-pr-220-role)
aws_iam_instance_profile.test_profile: Creating...
  arn:         "" => "<computed>"
  create_date: "" => "<computed>"
  name:        "" => "mao-683-pr-220-worker-profile"
  path:        "" => "/"
  role:        "" => "mao-683-pr-220-role"
  roles.#:     "" => "<computed>"
  unique_id:   "" => "<computed>"
aws_iam_instance_profile.test_profile: Creation complete after 0s (ID: mao-683-pr-220-worker-profile)
module.vpc.aws_vpc.this: Creation complete after 1s (ID: vpc-006255ee085f7e6a4)
module.vpc.aws_subnet.public[2]: Creating...
  arn:                             "" => "<computed>"
  assign_ipv6_address_on_creation: "" => "false"
  availability_zone:               "" => "us-east-1c"
  availability_zone_id:            "" => "<computed>"
  cidr_block:                      "" => "10.0.103.0/24"
  ipv6_cidr_block:                 "" => "<computed>"
  ipv6_cidr_block_association_id:  "" => "<computed>"
  map_public_ip_on_launch:         "" => "true"
  owner_id:                        "" => "<computed>"
  tags.%:                          "" => "3"
  tags.Environment:                "" => "dev"
  tags.Name:                       "" => "mao-683-pr-220-worker-foo"
  tags.Owner:                      "" => "user"
  vpc_id:                          "" => "vpc-006255ee085f7e6a4"
module.vpc.aws_route_table.private[1]: Creating...
  owner_id:           "" => "<computed>"
  propagating_vgws.#: "" => "<computed>"
  route.#:            "" => "<computed>"
  tags.%:             "" => "3"
  tags.Environment:   "" => "dev"
  tags.Name:          "" => "mao-683-pr-220-private-us-east-1b"
  tags.Owner:         "" => "user"
  vpc_id:             "" => "vpc-006255ee085f7e6a4"
aws_security_group.test: Creating...
  arn:                    "" => "<computed>"
  description:            "" => "Managed by Terraform"
  egress.#:               "" => "<computed>"
  ingress.#:              "" => "<computed>"
  name:                   "" => "mao-683-pr-220-sg"
  owner_id:               "" => "<computed>"
  revoke_rules_on_delete: "" => "false"
  tags.%:                 "" => "1"
  tags.Name:              "" => "mao-683-pr-220_worker_sg"
  vpc_id:                 "" => "vpc-006255ee085f7e6a4"
module.vpc.aws_subnet.public[0]: Creating...
  arn:                             "" => "<computed>"
  assign_ipv6_address_on_creation: "" => "false"
  availability_zone:               "" => "us-east-1a"
  availability_zone_id:            "" => "<computed>"
  cidr_block:                      "" => "10.0.101.0/24"
  ipv6_cidr_block:                 "" => "<computed>"
  ipv6_cidr_block_association_id:  "" => "<computed>"
  map_public_ip_on_launch:         "" => "true"
  owner_id:                        "" => "<computed>"
  tags.%:                          "" => "3"
  tags.Environment:                "" => "dev"
  tags.Name:                       "" => "mao-683-pr-220-worker-foo"
  tags.Owner:                      "" => "user"
  vpc_id:                          "" => "vpc-006255ee085f7e6a4"
module.vpc.aws_route_table.public: Creating...
  owner_id:           "" => "<computed>"
  propagating_vgws.#: "" => "<computed>"
  route.#:            "" => "<computed>"
  tags.%:             "" => "3"
  tags.Environment:   "" => "dev"
  tags.Name:          "" => "mao-683-pr-220-public"
  tags.Owner:         "" => "user"
  vpc_id:             "" => "vpc-006255ee085f7e6a4"
module.vpc.aws_route_table.private[2]: Creating...
  owner_id:           "" => "<computed>"
  propagating_vgws.#: "" => "<computed>"
  route.#:            "" => "<computed>"
  tags.%:             "" => "3"
  tags.Environment:   "" => "dev"
  tags.Name:          "" => "mao-683-pr-220-private-us-east-1c"
  tags.Owner:         "" => "user"
  vpc_id:             "" => "vpc-006255ee085f7e6a4"
module.vpc.aws_subnet.private[1]: Creating...
  arn:                             "" => "<computed>"
  assign_ipv6_address_on_creation: "" => "false"
  availability_zone:               "" => "us-east-1b"
  availability_zone_id:            "" => "<computed>"
  cidr_block:                      "" => "10.0.2.0/24"
  ipv6_cidr_block:                 "" => "<computed>"
  ipv6_cidr_block_association_id:  "" => "<computed>"
  map_public_ip_on_launch:         "" => "false"
  owner_id:                        "" => "<computed>"
  tags.%:                          "" => "3"
  tags.Environment:                "" => "dev"
  tags.Name:                       "" => "mao-683-pr-220-private-us-east-1b"
  tags.Owner:                      "" => "user"
  vpc_id:                          "" => "vpc-006255ee085f7e6a4"
module.vpc.aws_route_table.private[0]: Creating...
  owner_id:           "" => "<computed>"
  propagating_vgws.#: "" => "<computed>"
  route.#:            "" => "<computed>"
  tags.%:             "" => "3"
  tags.Environment:   "" => "dev"
  tags.Name:          "" => "mao-683-pr-220-private-us-east-1a"
  tags.Owner:         "" => "user"
  vpc_id:             "" => "vpc-006255ee085f7e6a4"
module.vpc.aws_subnet.private[0]: Creating...
  arn:                             "" => "<computed>"
  assign_ipv6_address_on_creation: "" => "false"
  availability_zone:               "" => "us-east-1a"
  availability_zone_id:            "" => "<computed>"
  cidr_block:                      "" => "10.0.1.0/24"
  ipv6_cidr_block:                 "" => "<computed>"
  ipv6_cidr_block_association_id:  "" => "<computed>"
  map_public_ip_on_launch:         "" => "false"
  owner_id:                        "" => "<computed>"
  tags.%:                          "" => "3"
  tags.Environment:                "" => "dev"
  tags.Name:                       "" => "mao-683-pr-220-private-us-east-1a"
  tags.Owner:                      "" => "user"
  vpc_id:                          "" => "vpc-006255ee085f7e6a4"
module.vpc.aws_subnet.private[2]: Creating...
  arn:                             "" => "<computed>"
  assign_ipv6_address_on_creation: "" => "false"
  availability_zone:               "" => "us-east-1c"
  availability_zone_id:            "" => "<computed>"
  cidr_block:                      "" => "10.0.3.0/24"
  ipv6_cidr_block:                 "" => "<computed>"
  ipv6_cidr_block_association_id:  "" => "<computed>"
  map_public_ip_on_launch:         "" => "false"
  owner_id:                        "" => "<computed>"
  tags.%:                          "" => "3"
  tags.Environment:                "" => "dev"
  tags.Name:                       "" => "mao-683-pr-220-private-us-east-1c"
  tags.Owner:                      "" => "user"
  vpc_id:                          "" => "vpc-006255ee085f7e6a4"
module.vpc.aws_route_table.public: Creation complete after 1s (ID: rtb-0b8b9bdd0c0ca8f03)
module.vpc.aws_internet_gateway.this: Creating...
  owner_id:         "" => "<computed>"
  tags.%:           "0" => "3"
  tags.Environment: "" => "dev"
  tags.Name:        "" => "mao-683-pr-220"
  tags.Owner:       "" => "user"
  vpc_id:           "" => "vpc-006255ee085f7e6a4"
module.vpc.aws_route_table.private[1]: Creation complete after 1s (ID: rtb-026e9dc61f9d41ab0)
module.vpc.aws_subnet.public[1]: Creating...
  arn:                             "" => "<computed>"
  assign_ipv6_address_on_creation: "" => "false"
  availability_zone:               "" => "us-east-1b"
  availability_zone_id:            "" => "<computed>"
  cidr_block:                      "" => "10.0.102.0/24"
  ipv6_cidr_block:                 "" => "<computed>"
  ipv6_cidr_block_association_id:  "" => "<computed>"
  map_public_ip_on_launch:         "" => "true"
  owner_id:                        "" => "<computed>"
  tags.%:                          "" => "3"
  tags.Environment:                "" => "dev"
  tags.Name:                       "" => "mao-683-pr-220-worker-foo"
  tags.Owner:                      "" => "user"
  vpc_id:                          "" => "vpc-006255ee085f7e6a4"
module.vpc.aws_route_table.private[2]: Creation complete after 1s (ID: rtb-0dee653484b3d3204)
module.vpc.aws_route_table.private[0]: Creation complete after 1s (ID: rtb-0d1b62d75bced2e14)
aws_security_group.test: Creation complete after 1s (ID: sg-0ddd7b21c36f37fb9)
aws_security_group_rule.test: Creating...
  cidr_blocks.#:            "" => "1"
  cidr_blocks.0:            "" => "0.0.0.0/0"
  from_port:                "" => "0"
  protocol:                 "" => "tcp"
  security_group_id:        "" => "sg-0ddd7b21c36f37fb9"
  self:                     "" => "false"
  source_security_group_id: "" => "<computed>"
  to_port:                  "" => "0"
  type:                     "" => "ingress"
module.vpc.aws_subnet.private[1]: Creation complete after 1s (ID: subnet-0221627cfdc1787b3)
module.vpc.aws_subnet.private[0]: Creation complete after 1s (ID: subnet-04c91f411dae7a31a)
module.vpc.aws_subnet.private[2]: Creation complete after 1s (ID: subnet-0ef2e64382ec232e4)
module.vpc.aws_route_table_association.private[1]: Creating...
  route_table_id: "" => "rtb-026e9dc61f9d41ab0"
  subnet_id:      "" => "subnet-0221627cfdc1787b3"
module.vpc.aws_route_table_association.private[2]: Creating...
  route_table_id: "" => "rtb-0dee653484b3d3204"
  subnet_id:      "" => "subnet-0ef2e64382ec232e4"
module.vpc.aws_route_table_association.private[0]: Creating...
  route_table_id: "" => "rtb-0d1b62d75bced2e14"
  subnet_id:      "" => "subnet-04c91f411dae7a31a"
module.vpc.aws_internet_gateway.this: Creation complete after 0s (ID: igw-0a88c50c888967c00)
module.vpc.aws_route.public_internet_gateway: Creating...
  destination_cidr_block:     "" => "0.0.0.0/0"
  destination_prefix_list_id: "" => "<computed>"
  egress_only_gateway_id:     "" => "<computed>"
  gateway_id:                 "" => "igw-0a88c50c888967c00"
  instance_id:                "" => "<computed>"
  instance_owner_id:          "" => "<computed>"
  nat_gateway_id:             "" => "<computed>"
  network_interface_id:       "" => "<computed>"
  origin:                     "" => "<computed>"
  route_table_id:             "" => "rtb-0b8b9bdd0c0ca8f03"
  state:                      "" => "<computed>"
module.vpc.aws_subnet.public[0]: Creation complete after 1s (ID: subnet-0f3fc58f1c124ecba)
module.vpc.aws_route_table_association.private[2]: Creation complete after 0s (ID: rtbassoc-0ed0186fa9c6a966d)
module.vpc.aws_subnet.public[2]: Creation complete after 1s (ID: subnet-0de8ea66ce59a577d)
module.vpc.aws_route_table_association.private[0]: Creation complete after 0s (ID: rtbassoc-0f12f06c4ebebc816)
module.vpc.aws_route_table_association.private[1]: Creation complete after 0s (ID: rtbassoc-0028faddca2e39a83)
aws_security_group_rule.test: Creation complete after 0s (ID: sgrule-3270019908)
module.vpc.aws_subnet.public[1]: Creation complete after 0s (ID: subnet-00fd6755daa16640f)
module.vpc.aws_route_table_association.public[2]: Creating...
  route_table_id: "" => "rtb-0b8b9bdd0c0ca8f03"
  subnet_id:      "" => "subnet-0de8ea66ce59a577d"
module.vpc.aws_route_table_association.public[1]: Creating...
  route_table_id: "" => "rtb-0b8b9bdd0c0ca8f03"
  subnet_id:      "" => "subnet-00fd6755daa16640f"
module.vpc.aws_route_table_association.public[0]: Creating...
  route_table_id: "" => "rtb-0b8b9bdd0c0ca8f03"
  subnet_id:      "" => "subnet-0f3fc58f1c124ecba"
module.vpc.aws_route.public_internet_gateway: Creation complete after 0s (ID: r-rtb-0b8b9bdd0c0ca8f031080289494)
module.vpc.aws_route_table_association.public[0]: Creation complete after 0s (ID: rtbassoc-01ffbc193e9d3c8e2)
module.vpc.aws_route_table_association.public[1]: Creation complete after 0s (ID: rtbassoc-0bbd7c0c5defa83f5)
module.vpc.aws_route_table_association.public[2]: Creation complete after 1s (ID: rtbassoc-0e8a41c0a424f9bd4)

Apply complete! Resources: 23 added, 0 changed, 0 destroyed.
time="2019-03-01T11:33:15Z" level=info msg="Creating \"openshift-machine-api\" namespace..."
time="2019-03-01T11:33:15Z" level=info msg="Creating \"default-account-openshift-machine-api\" ClusterRoleBinding..."
time="2019-03-01T11:33:15Z" level=info msg="Creating \"clusteroperators.config.openshift.io\" CRD..."
time="2019-03-01T11:33:20Z" level=info msg="Creating \"machines.machine.openshift.io\" CRD..."
time="2019-03-01T11:33:21Z" level=info msg="Creating \"machinesets.machine.openshift.io\" CRD..."
time="2019-03-01T11:33:22Z" level=info msg="Creating \"machinedeployments.machine.openshift.io\" CRD..."
time="2019-03-01T11:33:23Z" level=info msg="Creating \"clusters.machine.openshift.io\" CRD..."
time="2019-03-01T11:33:24Z" level=info msg="Creating \"machinehealthchecks.healthchecking.openshift.io\" CRD..."
time="2019-03-01T11:33:25Z" level=info msg="Creating \"kube-system/cluster-config-v1\" ConfigMap..."
time="2019-03-01T11:33:25Z" level=info msg="Creating \"openshift-machine-api/machine-api-operator-images\" ConfigMap..."
time="2019-03-01T11:33:25Z" level=info msg="Creating \"openshift-machine-api/aws-credentials-secret\" secret..."
time="2019-03-01T11:33:25Z" level=info msg="Creating \"openshift-machine-api/ignition-worker\" secret..."
time="2019-03-01T11:33:25Z" level=info msg="Creating machine-api-operator..."
time="2019-03-01T11:33:26Z" level=info msg="Waiting for clusterapi-manager-controllers deployment to be created: deployments.apps \"clusterapi-manager-controllers\" not found"
time="2019-03-01T11:33:27Z" level=info msg="Waiting for clusterapi-manager-controllers deployment to be created: deployments.apps \"clusterapi-manager-controllers\" not found"
time="2019-03-01T11:33:28Z" level=info msg="Waiting for clusterapi-manager-controllers deployment to be created: deployments.apps \"clusterapi-manager-controllers\" not found"
time="2019-03-01T11:33:29Z" level=info msg="Waiting for clusterapi-manager-controllers deployment to be created: deployments.apps \"clusterapi-manager-controllers\" not found"
time="2019-03-01T11:33:30Z" level=info msg="Waiting for clusterapi-manager-controllers deployment to be created: deployments.apps \"clusterapi-manager-controllers\" not found"
time="2019-03-01T11:33:31Z" level=info msg="Waiting for all clusterapi-manager-controllers deployment pods to be ready, have 0, expecting 1"
time="2019-03-01T11:33:32Z" level=info msg="Waiting for all clusterapi-manager-controllers deployment pods to be ready, have 0, expecting 1"
time="2019-03-01T11:33:33Z" level=info msg="Waiting for all clusterapi-manager-controllers deployment pods to be ready, have 0, expecting 1"
time="2019-03-01T11:33:34Z" level=info msg="Waiting for all clusterapi-manager-controllers deployment pods to be ready, have 0, expecting 1"
time="2019-03-01T11:33:35Z" level=info msg="Waiting for all clusterapi-manager-controllers deployment pods to be ready, have 1, expecting 1"
time="2019-03-01T11:33:36Z" level=info msg="Cluster object has been created"
time="2019-03-01T11:33:36Z" level=info msg="MachineSet object has been created"
time="2019-03-01T11:33:36Z" level=info msg="Machine objects has been created"
time="2019-03-01T11:33:36Z" level=info msg="The cluster-api stack is ready"
time="2019-03-01T11:33:36Z" level=info msg="The cluster, the machineSet and the machines have been deployed"
time="2019-03-01T11:33:37Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:39Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:40Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:41Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:42Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:43Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:44Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:45Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:46Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:47Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:48Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:49Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:50Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:51Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:52Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:53Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:54Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:55Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:56Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:57Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:58Z" level=info msg="Waiting for aws instances to come up"
time="2019-03-01T11:33:58Z" level=info msg="Two instances are running on aws"
time="2019-03-01T11:33:58Z" level=info msg="All verified successfully. Tearing down..."
time="2019-03-01T11:33:58Z" level=info msg="Running terraform destroy"
aws_vpc.this: Refreshing state... (ID: vpc-006255ee085f7e6a4)
aws_iam_role.role: Refreshing state... (ID: mao-683-pr-220-role)
aws_iam_instance_profile.test_profile: Refreshing state... (ID: mao-683-pr-220-worker-profile)
aws_security_group.test: Refreshing state... (ID: sg-0ddd7b21c36f37fb9)
aws_subnet.public[0]: Refreshing state... (ID: subnet-0f3fc58f1c124ecba)
aws_subnet.public[1]: Refreshing state... (ID: subnet-00fd6755daa16640f)
aws_internet_gateway.this: Refreshing state... (ID: igw-0a88c50c888967c00)
aws_subnet.private[0]: Refreshing state... (ID: subnet-04c91f411dae7a31a)
aws_subnet.private[1]: Refreshing state... (ID: subnet-0221627cfdc1787b3)
aws_subnet.private[2]: Refreshing state... (ID: subnet-0ef2e64382ec232e4)
aws_subnet.public[2]: Refreshing state... (ID: subnet-0de8ea66ce59a577d)
aws_route_table.private[2]: Refreshing state... (ID: rtb-0dee653484b3d3204)
aws_route_table.private[0]: Refreshing state... (ID: rtb-0d1b62d75bced2e14)
aws_route_table.public: Refreshing state... (ID: rtb-0b8b9bdd0c0ca8f03)
aws_route_table.private[1]: Refreshing state... (ID: rtb-026e9dc61f9d41ab0)
aws_security_group_rule.test: Refreshing state... (ID: sgrule-3270019908)
aws_route.public_internet_gateway: Refreshing state... (ID: r-rtb-0b8b9bdd0c0ca8f031080289494)
aws_route_table_association.public[0]: Refreshing state... (ID: rtbassoc-01ffbc193e9d3c8e2)
aws_route_table_association.public[1]: Refreshing state... (ID: rtbassoc-0bbd7c0c5defa83f5)
aws_route_table_association.public[2]: Refreshing state... (ID: rtbassoc-0e8a41c0a424f9bd4)
aws_route_table_association.private[2]: Refreshing state... (ID: rtbassoc-0ed0186fa9c6a966d)
aws_route_table_association.private[1]: Refreshing state... (ID: rtbassoc-0028faddca2e39a83)
aws_route_table_association.private[0]: Refreshing state... (ID: rtbassoc-0f12f06c4ebebc816)
aws_security_group_rule.test: Destroying... (ID: sgrule-3270019908)
module.vpc.aws_route_table_association.public[2]: Destroying... (ID: rtbassoc-0e8a41c0a424f9bd4)
module.vpc.aws_route_table_association.public[1]: Destroying... (ID: rtbassoc-0bbd7c0c5defa83f5)
module.vpc.aws_route_table_association.public[0]: Destroying... (ID: rtbassoc-01ffbc193e9d3c8e2)
module.vpc.aws_route_table_association.private[1]: Destroying... (ID: rtbassoc-0028faddca2e39a83)
module.vpc.aws_route_table_association.private[2]: Destroying... (ID: rtbassoc-0ed0186fa9c6a966d)
aws_iam_instance_profile.test_profile: Destroying... (ID: mao-683-pr-220-worker-profile)
module.vpc.aws_route_table_association.private[0]: Destroying... (ID: rtbassoc-0f12f06c4ebebc816)
module.vpc.aws_route.public_internet_gateway: Destroying... (ID: r-rtb-0b8b9bdd0c0ca8f031080289494)
aws_iam_instance_profile.test_profile: Destruction complete after 0s
aws_iam_role.role: Destroying... (ID: mao-683-pr-220-role)
module.vpc.aws_route_table_association.public[1]: Destruction complete after 0s
module.vpc.aws_route_table_association.private[1]: Destruction complete after 0s
module.vpc.aws_route_table_association.public[0]: Destruction complete after 0s
module.vpc.aws_route_table_association.public[2]: Destruction complete after 0s
module.vpc.aws_route_table_association.private[0]: Destruction complete after 0s
module.vpc.aws_subnet.public[1]: Destroying... (ID: subnet-00fd6755daa16640f)
module.vpc.aws_subnet.public[2]: Destroying... (ID: subnet-0de8ea66ce59a577d)
module.vpc.aws_subnet.public[0]: Destroying... (ID: subnet-0f3fc58f1c124ecba)
module.vpc.aws_route.public_internet_gateway: Destruction complete after 0s
module.vpc.aws_internet_gateway.this: Destroying... (ID: igw-0a88c50c888967c00)
module.vpc.aws_route_table.public: Destroying... (ID: rtb-0b8b9bdd0c0ca8f03)
module.vpc.aws_route_table_association.private[2]: Destruction complete after 0s
module.vpc.aws_route_table.private[1]: Destroying... (ID: rtb-026e9dc61f9d41ab0)
module.vpc.aws_subnet.private[2]: Destroying... (ID: subnet-0ef2e64382ec232e4)
module.vpc.aws_subnet.private[1]: Destroying... (ID: subnet-0221627cfdc1787b3)
aws_security_group_rule.test: Destruction complete after 0s
module.vpc.aws_route_table.private[2]: Destroying... (ID: rtb-0dee653484b3d3204)
aws_iam_role.role: Destruction complete after 0s
module.vpc.aws_subnet.private[0]: Destroying... (ID: subnet-04c91f411dae7a31a)
module.vpc.aws_route_table.private[2]: Destruction complete after 0s
module.vpc.aws_route_table.public: Destruction complete after 0s
module.vpc.aws_route_table.private[0]: Destroying... (ID: rtb-0d1b62d75bced2e14)
module.vpc.aws_route_table.private[1]: Destruction complete after 0s
aws_security_group.test: Destroying... (ID: sg-0ddd7b21c36f37fb9)
module.vpc.aws_subnet.public[0]: Destruction complete after 1s
module.vpc.aws_subnet.public[2]: Destruction complete after 1s
module.vpc.aws_subnet.private[1]: Destruction complete after 1s
module.vpc.aws_subnet.private[2]: Destruction complete after 1s
module.vpc.aws_subnet.private[0]: Destruction complete after 1s
module.vpc.aws_route_table.private[0]: Destruction complete after 1s
module.vpc.aws_subnet.public.1: Still destroying... (ID: subnet-00fd6755daa16640f, 10s elapsed)
module.vpc.aws_internet_gateway.this: Still destroying... (ID: igw-0a88c50c888967c00, 10s elapsed)
aws_security_group.test: Still destroying... (ID: sg-0ddd7b21c36f37fb9, 10s elapsed)
module.vpc.aws_subnet.public.1: Still destroying... (ID: subnet-00fd6755daa16640f, 20s elapsed)
module.vpc.aws_internet_gateway.this: Still destroying... (ID: igw-0a88c50c888967c00, 20s elapsed)
aws_security_group.test: Still destroying... (ID: sg-0ddd7b21c36f37fb9, 20s elapsed)
module.vpc.aws_subnet.public.1: Still destroying... (ID: subnet-00fd6755daa16640f, 30s elapsed)
module.vpc.aws_internet_gateway.this: Still destroying... (ID: igw-0a88c50c888967c00, 30s elapsed)
aws_security_group.test: Still destroying... (ID: sg-0ddd7b21c36f37fb9, 30s elapsed)
module.vpc.aws_subnet.public.1: Still destroying... (ID: subnet-00fd6755daa16640f, 40s elapsed)
module.vpc.aws_internet_gateway.this: Still destroying... (ID: igw-0a88c50c888967c00, 40s elapsed)
aws_security_group.test: Still destroying... (ID: sg-0ddd7b21c36f37fb9, 40s elapsed)
module.vpc.aws_internet_gateway.this: Destruction complete after 45s
module.vpc.aws_subnet.public[1]: Destruction complete after 47s
aws_security_group.test: Destruction complete after 47s
module.vpc.aws_vpc.this: Destroying... (ID: vpc-006255ee085f7e6a4)
module.vpc.aws_vpc.this: Destruction complete after 1s

Destroy complete! Resources: 23 destroyed.
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPLOY AND TEST THE MACHINE API OPERATOR [00h 02m 35s] ##########
[PostBuildScript] - Executing post build scripts.
[workspace] $ /bin/bash /tmp/jenkins5554856295858522626.sh
########## STARTING STAGE: DOWNLOAD ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/gathered
+ rm -rf /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/gathered
+ mkdir -p /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/gathered
+ tree /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/gathered
/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/gathered

0 directories, 0 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins256560835245120245.sh
########## STARTING STAGE: GENERATE ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/generated
+ rm -rf /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/generated
+ mkdir /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/generated
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a 2>&1'
  WARNING: You're not using the default seccomp profile
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo cat /etc/sysconfig/docker /etc/sysconfig/docker-network /etc/sysconfig/docker-storage /etc/sysconfig/docker-storage-setup /etc/systemd/system/docker.service 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1 2>&1'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo df -T -h && sudo pvs && sudo vgs && sudo lvs && sudo findmnt --all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo yum list installed 2>&1'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl --dmesg --no-pager --all --lines=all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl _PID=1 --no-pager --all --lines=all 2>&1'
+ tree /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/generated
/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/generated
├── avc_denials.log
├── containers.log
├── dmesg.log
├── docker.config
├── docker.info
├── filesystem.info
├── installed_packages.log
└── pid1.journal

0 directories, 8 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins5850260002542291669.sh
########## STARTING STAGE: FETCH SYSTEMD JOURNALS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/journals
+ rm -rf /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/journals
+ mkdir /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/journals
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit docker.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
+ tree /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/journals
/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/artifacts/journals
├── dnsmasq.service
├── docker.service
└── systemd-journald.service

0 directories, 3 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins2822351564612967208.sh
########## STARTING STAGE: ASSEMBLE GCS OUTPUT ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
+ trap 'exit 0' EXIT
+ mkdir -p gcs/artifacts gcs/artifacts/generated gcs/artifacts/journals gcs/artifacts/gathered
++ python -c 'import json; import urllib; print json.load(urllib.urlopen('\''https://ci.openshift.redhat.com/jenkins/job/ci-kubernetes-machine-api-operator/683/api/json'\''))['\''result'\'']'
+ result=SUCCESS
+ cat
++ date +%s
+ cat /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/builds/683/log
+ cp artifacts/generated/avc_denials.log artifacts/generated/containers.log artifacts/generated/dmesg.log artifacts/generated/docker.config artifacts/generated/docker.info artifacts/generated/filesystem.info artifacts/generated/installed_packages.log artifacts/generated/pid1.journal gcs/artifacts/generated/
+ cp artifacts/journals/dnsmasq.service artifacts/journals/docker.service artifacts/journals/systemd-journald.service gcs/artifacts/journals/
++ pwd
+ scp -F ./.config/origin-ci-tool/inventory/.ssh_config -r /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/gcs openshiftdevel:/data
+ scp -F ./.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.config/gcloud/gcs-publisher-credentials.json openshiftdevel:/data/credentials.json
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins4690231284742231779.sh
########## STARTING STAGE: PUSH THE ARTIFACTS AND METADATA ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ mktemp
+ script=/tmp/tmp.w75mrsrHrP
+ cat
+ chmod +x /tmp/tmp.w75mrsrHrP
+ scp -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.w75mrsrHrP openshiftdevel:/tmp/tmp.w75mrsrHrP
+ ssh -F /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 300 /tmp/tmp.w75mrsrHrP"'
+ cd /home/origin
+ trap 'exit 0' EXIT
+ [[ -n {"type":"presubmit","job":"ci-kubernetes-machine-api-operator","buildid":"1101442818523533312","prowjobid":"64e3c8c9-3c14-11e9-b6ae-0a58ac10f509","refs":{"org":"openshift","repo":"machine-api-operator","base_ref":"master","base_sha":"9b9b9f9446267e97478002dd222851919e53c58d","pulls":[{"number":220,"author":"ingvagabund","sha":"86992f9cb25c5b6df83c6c79bfcad3d60253350f"}]}} ]]
++ jq --compact-output '.buildid |= "683"'
+ JOB_SPEC='{"type":"presubmit","job":"ci-kubernetes-machine-api-operator","buildid":"683","prowjobid":"64e3c8c9-3c14-11e9-b6ae-0a58ac10f509","refs":{"org":"openshift","repo":"machine-api-operator","base_ref":"master","base_sha":"9b9b9f9446267e97478002dd222851919e53c58d","pulls":[{"number":220,"author":"ingvagabund","sha":"86992f9cb25c5b6df83c6c79bfcad3d60253350f"}]}}'
+ docker run -e 'JOB_SPEC={"type":"presubmit","job":"ci-kubernetes-machine-api-operator","buildid":"683","prowjobid":"64e3c8c9-3c14-11e9-b6ae-0a58ac10f509","refs":{"org":"openshift","repo":"machine-api-operator","base_ref":"master","base_sha":"9b9b9f9446267e97478002dd222851919e53c58d","pulls":[{"number":220,"author":"ingvagabund","sha":"86992f9cb25c5b6df83c6c79bfcad3d60253350f"}]}}' -v /data:/data:z registry.svc.ci.openshift.org/ci/gcsupload:latest --dry-run=false --gcs-path=gs://origin-ci-test --gcs-credentials-file=/data/credentials.json --path-strategy=single --default-org=openshift --default-repo=origin /data/gcs/artifacts /data/gcs/build-log.txt /data/gcs/finished.json
Unable to find image 'registry.svc.ci.openshift.org/ci/gcsupload:latest' locally
Trying to pull repository registry.svc.ci.openshift.org/ci/gcsupload ... 
latest: Pulling from registry.svc.ci.openshift.org/ci/gcsupload
a073c86ecf9e: Already exists
cc3fc741b1a9: Already exists
404f55af0f52: Pulling fs layer
85cea451eec0: Pulling fs layer
85cea451eec0: Verifying Checksum
85cea451eec0: Download complete
404f55af0f52: Verifying Checksum
404f55af0f52: Download complete
404f55af0f52: Pull complete
85cea451eec0: Pull complete
Digest: sha256:6c1150286a76f9f8faf1f2702e62aecec3081c8dbd93e4c4ba62d7b28bd43caf
Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/gcsupload:latest
{"component":"gcsupload","level":"info","msg":"Gathering artifacts from artifact directory: /data/gcs/artifacts","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/avc_denials.log in artifact directory. Uploading as pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/avc_denials.log\n","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/containers.log in artifact directory. Uploading as pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/containers.log\n","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/dmesg.log in artifact directory. Uploading as pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/dmesg.log\n","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.config in artifact directory. Uploading as pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/docker.config\n","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.info in artifact directory. Uploading as pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/docker.info\n","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/filesystem.info in artifact directory. Uploading as pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/filesystem.info\n","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/installed_packages.log in artifact directory. Uploading as pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/installed_packages.log\n","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/generated/pid1.journal in artifact directory. Uploading as pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/pid1.journal\n","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/dnsmasq.service in artifact directory. Uploading as pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/journals/dnsmasq.service\n","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/docker.service in artifact directory. Uploading as pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/journals/docker.service\n","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","level":"info","msg":"Found /data/gcs/artifacts/journals/systemd-journald.service in artifact directory. Uploading as pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/journals/systemd-journald.service\n","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/docker.config","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/dmesg.log","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/containers.log","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/journals/docker.service","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/latest-build.txt","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/filesystem.info","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/installed_packages.log","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/journals/systemd-journald.service","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/directory/ci-kubernetes-machine-api-operator/latest-build.txt","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/avc_denials.log","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/docker.info","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/pid1.journal","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/journals/dnsmasq.service","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/build-log.txt","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/finished.json","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/directory/ci-kubernetes-machine-api-operator/683.txt","level":"info","msg":"Queued for upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/finished.json","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/docker.config","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/filesystem.info","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/directory/ci-kubernetes-machine-api-operator/latest-build.txt","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/avc_denials.log","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/journals/dnsmasq.service","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/directory/ci-kubernetes-machine-api-operator/683.txt","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/journals/systemd-journald.service","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/docker.info","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/latest-build.txt","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/installed_packages.log","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/dmesg.log","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/pid1.journal","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/build-log.txt","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/generated/containers.log","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/220/ci-kubernetes-machine-api-operator/683/artifacts/journals/docker.service","level":"info","msg":"Finished upload","time":"2019-03-01T11:35:20Z"}
{"component":"gcsupload","level":"info","msg":"Finished upload to GCS","time":"2019-03-01T11:35:20Z"}
+ exit 0
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: PUSH THE ARTIFACTS AND METADATA [00h 00m 05s] ##########
[workspace] $ /bin/bash /tmp/jenkins2375923364863903628.sh
########## STARTING STAGE: DEPROVISION CLOUD RESOURCES ##########
+ [[ -s /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate ]]
+ source /var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config
+ oct deprovision

PLAYBOOK: main.yml *************************************************************
4 plays in /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml

PLAY [ensure we have the parameters necessary to deprovision virtual hosts] ****

TASK [ensure all required variables are set] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:9
skipping: [localhost] => (item=origin_ci_inventory_dir)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:35:22.073480", 
    "item": "origin_ci_inventory_dir", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_region)  => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:35:22.076589", 
    "item": "origin_ci_aws_region", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

PLAY [deprovision virtual hosts in EC2] ****************************************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

TASK [deprovision a virtual EC2 host] ******************************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:28
included: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml for localhost

TASK [update the SSH configuration to remove AWS EC2 specifics] ****************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:2
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-03-01 06:35:22.829861", 
    "msg": ""
}

TASK [rename EC2 instance for termination reaper] ******************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:8
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-03-01 06:35:23.531452", 
    "msg": "Tags {'Name': 'oct-terminate'} created for resource i-015e1b4f6f2a05e43."
}

TASK [tear down the EC2 instance] **********************************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:15
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-03-01 06:35:24.545351", 
    "instance_ids": [
        "i-015e1b4f6f2a05e43"
    ], 
    "instances": [
        {
            "ami_launch_index": "0", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-02bd0a734a29daa54"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-08bf4550ecb1eb195"
                }
            }, 
            "dns_name": "ec2-35-172-227-13.compute-1.amazonaws.com", 
            "ebs_optimized": false, 
            "groups": {
                "sg-7e73221a": "default"
            }, 
            "hypervisor": "xen", 
            "id": "i-015e1b4f6f2a05e43", 
            "image_id": "ami-0b77b87a37c3e662c", 
            "instance_type": "m4.xlarge", 
            "kernel": null, 
            "key_name": "libra", 
            "launch_time": "2019-03-01T11:23:39.000Z", 
            "placement": "us-east-1c", 
            "private_dns_name": "ip-172-18-25-241.ec2.internal", 
            "private_ip": "172.18.25.241", 
            "public_dns_name": "ec2-35-172-227-13.compute-1.amazonaws.com", 
            "public_ip": "35.172.227.13", 
            "ramdisk": null, 
            "region": "us-east-1", 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "running", 
            "state_code": 16, 
            "tags": {
                "Name": "oct-terminate", 
                "openshift_etcd": "", 
                "openshift_master": "", 
                "openshift_node": ""
            }, 
            "tenancy": "default", 
            "virtualization_type": "hvm"
        }
    ], 
    "tagged_instances": []
}

TASK [remove the serialized host variables] ************************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:22
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-03-01 06:35:24.785548", 
    "path": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory/host_vars/172.18.25.241.yml", 
    "state": "absent"
}

PLAY [deprovision virtual hosts locally manged by Vagrant] *********************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

PLAY [clean up local configuration for deprovisioned instances] ****************

TASK [remove inventory configuration directory] ********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:61
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-03-01 06:35:25.220713", 
    "path": "/var/lib/jenkins/jobs/ci-kubernetes-machine-api-operator/workspace/.config/origin-ci-tool/inventory", 
    "state": "absent"
}

PLAY RECAP *********************************************************************
localhost                  : ok=8    changed=4    unreachable=0    failed=0   

+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPROVISION CLOUD RESOURCES [00h 00m 05s] ##########
Archiving artifacts
Recording test results
[WS-CLEANUP] Deleting project workspace...[WS-CLEANUP] done
Finished: SUCCESS