Console Output

Started by user OpenShift CI Robot
[EnvInject] - Loading node environment variables.
Building in workspace /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace
[WS-CLEANUP] Deleting project workspace...
[WS-CLEANUP] Deferred wipeout is used...
[workspace] $ /bin/bash /tmp/jenkins2609453306702544274.sh
########## STARTING STAGE: INSTALL THE ORIGIN-CI-TOOL ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
++ readlink /var/lib/jenkins/origin-ci-tool/latest
+ latest=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
+ touch /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
+ cp /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin/activate /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
+ cat
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
+ mkdir -p /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
+ rm -rf /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool
+ oct configure ansible-client verbosity 2
Option verbosity updated to be 2.
+ oct configure aws-client keypair_name libra
Option keypair_name updated to be libra.
+ oct configure aws-client private_key_path /var/lib/jenkins/.ssh/devenv.pem
Option private_key_path updated to be /var/lib/jenkins/.ssh/devenv.pem.
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: INSTALL THE ORIGIN-CI-TOOL [00h 00m 02s] ##########
[workspace] $ /bin/bash /tmp/jenkins690907187694611320.sh
########## STARTING STAGE: PROVISION CLOUD RESOURCES ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
+ oct provision remote all-in-one --os rhel --stage base --provider aws --discrete-ssh-config --name pull-ci-openshift-machine-api-operator-master-e2e_716

PLAYBOOK: aws-up.yml ***********************************************************
2 plays in /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/aws-up.yml

PLAY [ensure we have the parameters necessary to bring up the AWS EC2 instance] ***

TASK [ensure all required variables are set] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/aws-up.yml:9
skipping: [localhost] => (item=origin_ci_inventory_dir)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.058798", 
    "item": "origin_ci_inventory_dir", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_keypair_name)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.062427", 
    "item": "origin_ci_aws_keypair_name", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_private_key_path)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.065522", 
    "item": "origin_ci_aws_private_key_path", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_region)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.068791", 
    "item": "origin_ci_aws_region", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_ami_tags)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.071861", 
    "item": "origin_ci_aws_ami_tags", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_instance_name)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.074924", 
    "item": "origin_ci_aws_instance_name", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_master_instance_type)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.077990", 
    "item": "origin_ci_aws_master_instance_type", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_identifying_tag_key)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.082220", 
    "item": "origin_ci_aws_identifying_tag_key", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_hostname)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.085280", 
    "item": "origin_ci_aws_hostname", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_ssh_config_strategy)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.088318", 
    "item": "origin_ci_ssh_config_strategy", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=openshift_schedulable)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.092523", 
    "item": "openshift_schedulable", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=openshift_node_labels)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.096117", 
    "item": "openshift_node_labels", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

TASK [ensure all required variables are set] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/aws-up.yml:27
skipping: [localhost] => (item=origin_ci_aws_master_subnet)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.129332", 
    "item": "origin_ci_aws_master_subnet", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_etcd_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.134012", 
    "item": "origin_ci_aws_etcd_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_node_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.138165", 
    "item": "origin_ci_aws_node_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_master_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.143530", 
    "item": "origin_ci_aws_master_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_master_external_elb_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.147746", 
    "item": "origin_ci_aws_master_external_elb_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_master_internal_elb_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.153108", 
    "item": "origin_ci_aws_master_internal_elb_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_router_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.157366", 
    "item": "origin_ci_aws_router_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_router_elb_security_group)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.163313", 
    "item": "origin_ci_aws_router_elb_security_group", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

PLAY [provision an AWS EC2 instance] *******************************************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

TASK [inventory : initialize the inventory directory] **************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:2
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:23.956164", 
    "gid": 995, 
    "group": "jenkins", 
    "mode": "0755", 
    "owner": "jenkins", 
    "path": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 6, 
    "state": "directory", 
    "uid": 997
}

TASK [inventory : add the nested group mapping] ********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:7
changed: [localhost] => {
    "changed": true, 
    "checksum": "18aaee00994df38cc3a63b635893175235331a9c", 
    "dest": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/nested_group_mappings", 
    "generated_timestamp": "2019-09-06 10:54:24.419671", 
    "gid": 995, 
    "group": "jenkins", 
    "md5sum": "b30c3226ea63efa3ff9c5e346c14a16e", 
    "mode": "0644", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 93, 
    "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1567781664.2-110173740647490/source", 
    "state": "file", 
    "uid": 997
}

TASK [inventory : initialize the OSEv3 group variables directory] **************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:12
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-09-06 10:54:24.583878", 
    "gid": 995, 
    "group": "jenkins", 
    "mode": "0755", 
    "owner": "jenkins", 
    "path": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/group_vars/OSEv3", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 6, 
    "state": "directory", 
    "uid": 997
}

TASK [inventory : initialize the host variables directory] *********************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:17
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-09-06 10:54:24.752192", 
    "gid": 995, 
    "group": "jenkins", 
    "mode": "0755", 
    "owner": "jenkins", 
    "path": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/host_vars", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 6, 
    "state": "directory", 
    "uid": 997
}

TASK [inventory : add the default Origin installation configuration] ***********
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:22
changed: [localhost] => {
    "changed": true, 
    "checksum": "4c06ba508f055c20f13426e8587342e8765a7b66", 
    "dest": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/group_vars/OSEv3/general.yml", 
    "generated_timestamp": "2019-09-06 10:54:25.046171", 
    "gid": 995, 
    "group": "jenkins", 
    "md5sum": "8aec71c75f7d512b278ae7c6f2959b12", 
    "mode": "0644", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 331, 
    "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1567781664.91-108422028835334/source", 
    "state": "file", 
    "uid": 997
}

TASK [aws-up : determine if we are inside AWS EC2] *****************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:2
changed: [localhost] => {
    "changed": true, 
    "cmd": [
        "curl", 
        "-s", 
        "http://instance-data.ec2.internal"
    ], 
    "delta": "0:00:00.015330", 
    "end": "2019-09-06 10:54:25.272763", 
    "failed": false, 
    "failed_when_result": false, 
    "generated_timestamp": "2019-09-06 10:54:25.289640", 
    "rc": 0, 
    "start": "2019-09-06 10:54:25.257433", 
    "stderr": [], 
    "stdout": [
        "1.0", 
        "2007-01-19", 
        "2007-03-01", 
        "2007-08-29", 
        "2007-10-10", 
        "2007-12-15", 
        "2008-02-01", 
        "2008-09-01", 
        "2009-04-04", 
        "2011-01-01", 
        "2011-05-01", 
        "2012-01-12", 
        "2014-02-25", 
        "2014-11-05", 
        "2015-10-20", 
        "2016-04-19", 
        "2016-06-30", 
        "2016-09-02", 
        "2018-03-28", 
        "2018-08-17", 
        "2018-09-24", 
        "latest"
    ], 
    "warnings": [
        "Consider using get_url or uri module rather than running curl"
    ]
}
 [WARNING]: Consider using get_url or uri module rather than running curl

TASK [aws-up : configure EC2 parameters for inventory when controlling from inside EC2] ***
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:7
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_destination_variable": "private_dns_name", 
        "origin_ci_aws_host_address_variable": "private_ip", 
        "origin_ci_aws_vpc_destination_variable": "private_ip_address"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:25.328095"
}

TASK [aws-up : determine where to put the AWS API cache] ***********************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:14
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_cache_dir": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ec2_cache"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:25.363811"
}

TASK [aws-up : ensure we have a place to put the AWS API cache] ****************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:18
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-09-06 10:54:25.527700", 
    "gid": 995, 
    "group": "jenkins", 
    "mode": "0755", 
    "owner": "jenkins", 
    "path": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ec2_cache", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 6, 
    "state": "directory", 
    "uid": 997
}

TASK [aws-up : place the EC2 dynamic inventory script] *************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:23
changed: [localhost] => {
    "changed": true, 
    "checksum": "625b8af723189db3b96ba0026d0f997a0025bc47", 
    "dest": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/ec2.py", 
    "generated_timestamp": "2019-09-06 10:54:25.818870", 
    "gid": 995, 
    "group": "jenkins", 
    "md5sum": "cac06c14065dac74904232b89d4ba24c", 
    "mode": "0755", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 63725, 
    "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1567781665.68-235483521267226/source", 
    "state": "file", 
    "uid": 997
}

TASK [aws-up : place the EC2 dynamic inventory configuration] ******************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:29
changed: [localhost] => {
    "changed": true, 
    "checksum": "f94cf21fdbe6046b2b885ad1f93991e252011e31", 
    "dest": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/ec2.ini", 
    "generated_timestamp": "2019-09-06 10:54:26.101301", 
    "gid": 995, 
    "group": "jenkins", 
    "md5sum": "b483bdefdeb766739eb3202824fed18d", 
    "mode": "0644", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 421, 
    "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1567781665.85-58358705775797/source", 
    "state": "file", 
    "uid": 997
}

TASK [aws-up : place the EC2 tag to group mappings] ****************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:34
changed: [localhost] => {
    "changed": true, 
    "checksum": "b4205a33dc73f62bd4f77f35d045cf8e09ae62b0", 
    "dest": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/tag_to_group_mappings", 
    "generated_timestamp": "2019-09-06 10:54:26.387827", 
    "gid": 995, 
    "group": "jenkins", 
    "md5sum": "bc3a567a1b6f342e1005182efc1b66be", 
    "mode": "0644", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 287, 
    "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1567781666.25-109057332304890/source", 
    "state": "file", 
    "uid": 997
}

TASK [aws-up : list available AMIs] ********************************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:40
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:28.881907", 
    "results": [
        {
            "ami_id": "ami-04f9b88b6b0571f20", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "encrypted": false, 
                    "size": 75, 
                    "snapshot_id": "snap-0655d2d962c590c8c", 
                    "volume_type": "gp2"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "encrypted": false, 
                    "size": 50, 
                    "snapshot_id": "snap-0d86ae865b17f4def", 
                    "volume_type": "gp2"
                }
            }, 
            "creationDate": "2018-06-26T12:22:31.000Z", 
            "description": "OpenShift Origin development AMI on rhel at the base stage.", 
            "hypervisor": "xen", 
            "is_public": false, 
            "location": "531415883065/ami_build_origin_int_rhel_base_758", 
            "name": "ami_build_origin_int_rhel_base_758", 
            "owner_id": "531415883065", 
            "platform": null, 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "available", 
            "tags": {
                "Name": "ami_build_origin_int_rhel_base_758", 
                "image_stage": "base", 
                "operating_system": "rhel", 
                "ready": "yes"
            }, 
            "virtualization_type": "hvm"
        }, 
        {
            "ami_id": "ami-0b77b87a37c3e662c", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "encrypted": false, 
                    "size": 75, 
                    "snapshot_id": "snap-02ec23d4818f2747e", 
                    "volume_type": "gp2"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "encrypted": false, 
                    "size": 50, 
                    "snapshot_id": "snap-0d8726e441d4ca329", 
                    "volume_type": "gp2"
                }
            }, 
            "creationDate": "2018-06-26T22:18:53.000Z", 
            "description": "OpenShift Origin development AMI on rhel at the base stage.", 
            "hypervisor": "xen", 
            "is_public": false, 
            "location": "531415883065/ami_build_origin_int_rhel_base_760", 
            "name": "ami_build_origin_int_rhel_base_760", 
            "owner_id": "531415883065", 
            "platform": null, 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "available", 
            "tags": {
                "Name": "ami_build_origin_int_rhel_base_760", 
                "image_stage": "base", 
                "operating_system": "rhel", 
                "ready": "yes"
            }, 
            "virtualization_type": "hvm"
        }
    ]
}

TASK [aws-up : choose appropriate AMIs for use] ********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:49
ok: [localhost] => (item={u'ami_id': u'ami-04f9b88b6b0571f20', u'root_device_type': u'ebs', u'description': u'OpenShift Origin development AMI on rhel at the base stage.', u'tags': {u'ready': u'yes', u'image_stage': u'base', u'Name': u'ami_build_origin_int_rhel_base_758', u'operating_system': u'rhel'}, u'hypervisor': u'xen', u'block_device_mapping': {u'/dev/sdb': {u'encrypted': False, u'snapshot_id': u'snap-0d86ae865b17f4def', u'delete_on_termination': True, u'volume_type': u'gp2', u'size': 50}, u'/dev/sda1': {u'encrypted': False, u'snapshot_id': u'snap-0655d2d962c590c8c', u'delete_on_termination': True, u'volume_type': u'gp2', u'size': 75}}, u'architecture': u'x86_64', u'owner_id': u'531415883065', u'platform': None, u'state': u'available', u'location': u'531415883065/ami_build_origin_int_rhel_base_758', u'is_public': False, u'creationDate': u'2018-06-26T12:22:31.000Z', u'root_device_name': u'/dev/sda1', u'virtualization_type': u'hvm', u'name': u'ami_build_origin_int_rhel_base_758'}) => {
    "ansible_facts": {
        "origin_ci_aws_ami_id_candidate": "ami-04f9b88b6b0571f20"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:28.925900", 
    "item": {
        "ami_id": "ami-04f9b88b6b0571f20", 
        "architecture": "x86_64", 
        "block_device_mapping": {
            "/dev/sda1": {
                "delete_on_termination": true, 
                "encrypted": false, 
                "size": 75, 
                "snapshot_id": "snap-0655d2d962c590c8c", 
                "volume_type": "gp2"
            }, 
            "/dev/sdb": {
                "delete_on_termination": true, 
                "encrypted": false, 
                "size": 50, 
                "snapshot_id": "snap-0d86ae865b17f4def", 
                "volume_type": "gp2"
            }
        }, 
        "creationDate": "2018-06-26T12:22:31.000Z", 
        "description": "OpenShift Origin development AMI on rhel at the base stage.", 
        "hypervisor": "xen", 
        "is_public": false, 
        "location": "531415883065/ami_build_origin_int_rhel_base_758", 
        "name": "ami_build_origin_int_rhel_base_758", 
        "owner_id": "531415883065", 
        "platform": null, 
        "root_device_name": "/dev/sda1", 
        "root_device_type": "ebs", 
        "state": "available", 
        "tags": {
            "Name": "ami_build_origin_int_rhel_base_758", 
            "image_stage": "base", 
            "operating_system": "rhel", 
            "ready": "yes"
        }, 
        "virtualization_type": "hvm"
    }
}
ok: [localhost] => (item={u'ami_id': u'ami-0b77b87a37c3e662c', u'root_device_type': u'ebs', u'description': u'OpenShift Origin development AMI on rhel at the base stage.', u'tags': {u'ready': u'yes', u'image_stage': u'base', u'Name': u'ami_build_origin_int_rhel_base_760', u'operating_system': u'rhel'}, u'hypervisor': u'xen', u'block_device_mapping': {u'/dev/sdb': {u'encrypted': False, u'snapshot_id': u'snap-0d8726e441d4ca329', u'delete_on_termination': True, u'volume_type': u'gp2', u'size': 50}, u'/dev/sda1': {u'encrypted': False, u'snapshot_id': u'snap-02ec23d4818f2747e', u'delete_on_termination': True, u'volume_type': u'gp2', u'size': 75}}, u'architecture': u'x86_64', u'owner_id': u'531415883065', u'platform': None, u'state': u'available', u'location': u'531415883065/ami_build_origin_int_rhel_base_760', u'is_public': False, u'creationDate': u'2018-06-26T22:18:53.000Z', u'root_device_name': u'/dev/sda1', u'virtualization_type': u'hvm', u'name': u'ami_build_origin_int_rhel_base_760'}) => {
    "ansible_facts": {
        "origin_ci_aws_ami_id_candidate": "ami-0b77b87a37c3e662c"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:28.934105", 
    "item": {
        "ami_id": "ami-0b77b87a37c3e662c", 
        "architecture": "x86_64", 
        "block_device_mapping": {
            "/dev/sda1": {
                "delete_on_termination": true, 
                "encrypted": false, 
                "size": 75, 
                "snapshot_id": "snap-02ec23d4818f2747e", 
                "volume_type": "gp2"
            }, 
            "/dev/sdb": {
                "delete_on_termination": true, 
                "encrypted": false, 
                "size": 50, 
                "snapshot_id": "snap-0d8726e441d4ca329", 
                "volume_type": "gp2"
            }
        }, 
        "creationDate": "2018-06-26T22:18:53.000Z", 
        "description": "OpenShift Origin development AMI on rhel at the base stage.", 
        "hypervisor": "xen", 
        "is_public": false, 
        "location": "531415883065/ami_build_origin_int_rhel_base_760", 
        "name": "ami_build_origin_int_rhel_base_760", 
        "owner_id": "531415883065", 
        "platform": null, 
        "root_device_name": "/dev/sda1", 
        "root_device_type": "ebs", 
        "state": "available", 
        "tags": {
            "Name": "ami_build_origin_int_rhel_base_760", 
            "image_stage": "base", 
            "operating_system": "rhel", 
            "ready": "yes"
        }, 
        "virtualization_type": "hvm"
    }
}

TASK [aws-up : determine which AMI to use] *************************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:55
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_ami_id": "ami-0b77b87a37c3e662c"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:28.969498"
}

TASK [aws-up : determine which subnets are available] **************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:60
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:29.621021", 
    "subnets": [
        {
            "availability_zone": "us-east-1c", 
            "available_ip_address_count": 4053, 
            "cidr_block": "172.18.16.0/20", 
            "default_for_az": "false", 
            "id": "subnet-8bdb5ac2", 
            "map_public_ip_on_launch": "true", 
            "state": "available", 
            "tags": {
                "Name": "devenv-subnet-2", 
                "origin_ci_aws_cluster_component": "master_subnet"
            }, 
            "vpc_id": "vpc-69705d0c"
        }, 
        {
            "availability_zone": "us-east-1d", 
            "available_ip_address_count": 4036, 
            "cidr_block": "172.18.0.0/20", 
            "default_for_az": "false", 
            "id": "subnet-cf57c596", 
            "map_public_ip_on_launch": "true", 
            "state": "available", 
            "tags": {
                "Name": "devenv-subnet-1", 
                "origin_ci_aws_cluster_component": "master_subnet"
            }, 
            "vpc_id": "vpc-69705d0c"
        }
    ]
}

TASK [aws-up : determine which subnets to use for the master] ******************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:67
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_master_subnet_ids": [
            "subnet-8bdb5ac2", 
            "subnet-cf57c596"
        ]
    }, 
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:29.660871"
}

TASK [aws-up : determine which security groups are available] ******************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:72
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:30.452406", 
    "security_groups": [
        {
            "description": "default VPC security group", 
            "group_id": "sg-7e73221a", 
            "group_name": "default", 
            "ip_permissions": [
                {
                    "ip_protocol": "-1", 
                    "ip_ranges": [], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "user_id_group_pairs": [
                        {
                            "group_id": "sg-7e73221a", 
                            "user_id": "531415883065"
                        }
                    ]
                }, 
                {
                    "from_port": 80, 
                    "ip_protocol": "tcp", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "54.241.19.245/32"
                        }, 
                        {
                            "cidr_ip": "97.65.119.184/29"
                        }, 
                        {
                            "cidr_ip": "107.20.219.35/32"
                        }, 
                        {
                            "cidr_ip": "108.166.48.153/32"
                        }, 
                        {
                            "cidr_ip": "212.199.177.64/27"
                        }, 
                        {
                            "cidr_ip": "212.72.208.162/32"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": 443, 
                    "user_id_group_pairs": []
                }, 
                {
                    "from_port": 53, 
                    "ip_protocol": "tcp", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "119.254.120.64/26"
                        }, 
                        {
                            "cidr_ip": "209.132.176.0/20"
                        }, 
                        {
                            "cidr_ip": "209.132.186.34/32"
                        }, 
                        {
                            "cidr_ip": "213.175.37.10/32"
                        }, 
                        {
                            "cidr_ip": "62.40.79.66/32"
                        }, 
                        {
                            "cidr_ip": "66.187.224.0/20"
                        }, 
                        {
                            "cidr_ip": "66.187.239.0/24"
                        }, 
                        {
                            "cidr_ip": "38.140.108.0/24"
                        }, 
                        {
                            "cidr_ip": "213.175.37.9/32"
                        }, 
                        {
                            "cidr_ip": "38.99.12.232/29"
                        }, 
                        {
                            "cidr_ip": "4.14.33.72/30"
                        }, 
                        {
                            "cidr_ip": "4.14.35.88/29"
                        }, 
                        {
                            "cidr_ip": "50.227.40.96/29"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": 8444, 
                    "user_id_group_pairs": []
                }, 
                {
                    "from_port": 22, 
                    "ip_protocol": "tcp", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "0.0.0.0/0"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": 22, 
                    "user_id_group_pairs": []
                }, 
                {
                    "from_port": 53, 
                    "ip_protocol": "udp", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "209.132.176.0/20"
                        }, 
                        {
                            "cidr_ip": "66.187.224.0/20"
                        }, 
                        {
                            "cidr_ip": "66.187.239.0/24"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": 53, 
                    "user_id_group_pairs": []
                }, 
                {
                    "from_port": 0, 
                    "ip_protocol": "udp", 
                    "ip_ranges": [], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": 65535, 
                    "user_id_group_pairs": [
                        {
                            "group_id": "sg-0d1a5377", 
                            "user_id": "531415883065"
                        }, 
                        {
                            "group_id": "sg-5875023f", 
                            "user_id": "531415883065"
                        }, 
                        {
                            "group_id": "sg-7e73221a", 
                            "user_id": "531415883065"
                        }, 
                        {
                            "group_id": "sg-e1760186", 
                            "user_id": "531415883065"
                        }
                    ]
                }, 
                {
                    "from_port": 3389, 
                    "ip_protocol": "tcp", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "0.0.0.0/0"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": 3389, 
                    "user_id_group_pairs": []
                }, 
                {
                    "from_port": -1, 
                    "ip_protocol": "icmp", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "0.0.0.0/0"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "to_port": -1, 
                    "user_id_group_pairs": []
                }
            ], 
            "ip_permissions_egress": [
                {
                    "ip_protocol": "-1", 
                    "ip_ranges": [
                        {
                            "cidr_ip": "0.0.0.0/0"
                        }
                    ], 
                    "ipv6_ranges": [], 
                    "prefix_list_ids": [], 
                    "user_id_group_pairs": []
                }
            ], 
            "owner_id": "531415883065", 
            "tags": {
                "Name": "devenv-vpc", 
                "openshift_infra": "true", 
                "origin_ci_aws_cluster_component": "master_security_group"
            }, 
            "vpc_id": "vpc-69705d0c"
        }
    ]
}

TASK [aws-up : determine which security group to use] **************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:79
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_master_security_group_ids": [
            "sg-7e73221a"
        ]
    }, 
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:30.493234"
}

TASK [aws-up : provision an AWS EC2 instance] **********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:84
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-09-06 10:54:53.241195", 
    "instance_ids": [
        "i-06550787d42cc325e"
    ], 
    "instances": [
        {
            "ami_launch_index": "0", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-03d9d644224906960"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-0bbfd421d51201f8f"
                }
            }, 
            "dns_name": "ec2-52-200-5-193.compute-1.amazonaws.com", 
            "ebs_optimized": false, 
            "groups": {
                "sg-7e73221a": "default"
            }, 
            "hypervisor": "xen", 
            "id": "i-06550787d42cc325e", 
            "image_id": "ami-0b77b87a37c3e662c", 
            "instance_type": "m4.xlarge", 
            "kernel": null, 
            "key_name": "libra", 
            "launch_time": "2019-09-06T14:54:32.000Z", 
            "placement": "us-east-1c", 
            "private_dns_name": "ip-172-18-28-208.ec2.internal", 
            "private_ip": "172.18.28.208", 
            "public_dns_name": "ec2-52-200-5-193.compute-1.amazonaws.com", 
            "public_ip": "52.200.5.193", 
            "ramdisk": null, 
            "region": "us-east-1", 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "running", 
            "state_code": 16, 
            "tags": {
                "Name": "pull-ci-openshift-machine-api-operator-master-e2e_716", 
                "openshift_etcd": "", 
                "openshift_master": "", 
                "openshift_node": ""
            }, 
            "tenancy": "default", 
            "virtualization_type": "hvm"
        }
    ], 
    "tagged_instances": []
}

TASK [aws-up : determine the host address] *************************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:110
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_host": "172.18.28.208"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:53.278534"
}

TASK [aws-up : determine the default user to use for SSH] **********************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:114
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:53.310224", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

TASK [aws-up : determine the default user to use for SSH] **********************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:119
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_aws_ssh_user": "origin"
    }, 
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:53.345125"
}

TASK [aws-up : update variables for the host] **********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:124
changed: [localhost] => {
    "changed": true, 
    "checksum": "98d42c4cfc35d9a77439cd2692aafff6a7d6568f", 
    "dest": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/host_vars/172.18.28.208.yml", 
    "generated_timestamp": "2019-09-06 10:54:53.636528", 
    "gid": 995, 
    "group": "jenkins", 
    "md5sum": "87b1cba59abb1228d1809ea4e7434b85", 
    "mode": "0644", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 769, 
    "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1567781693.5-249965212557336/source", 
    "state": "file", 
    "uid": 997
}

TASK [aws-up : determine where updated SSH configuration should go] ************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:141
ok: [localhost] => {
    "ansible_facts": {
        "origin_ci_ssh_config_files": [
            "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config"
        ]
    }, 
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:53.673621"
}

TASK [aws-up : determine where updated SSH configuration should go] ************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:146
skipping: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 10:54:53.703676", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

TASK [aws-up : ensure the targeted SSH configuration file exists] **************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:151
changed: [localhost] => (item=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config) => {
    "changed": true, 
    "dest": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config", 
    "generated_timestamp": "2019-09-06 10:54:53.866543", 
    "gid": 995, 
    "group": "jenkins", 
    "item": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config", 
    "mode": "0644", 
    "owner": "jenkins", 
    "secontext": "system_u:object_r:var_lib_t:s0", 
    "size": 0, 
    "state": "file", 
    "uid": 997
}

TASK [aws-up : update the SSH configuration] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:157
changed: [localhost] => (item=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config) => {
    "changed": true, 
    "generated_timestamp": "2019-09-06 10:54:54.145978", 
    "item": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config", 
    "msg": "Block inserted"
}

TASK [aws-up : wait for SSH to be available] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:175
ok: [localhost] => {
    "changed": false, 
    "elapsed": 63, 
    "generated_timestamp": "2019-09-06 10:55:57.518832", 
    "path": null, 
    "port": 22, 
    "search_regex": null, 
    "state": "started"
}

PLAY RECAP *********************************************************************
localhost                  : ok=28   changed=13   unreachable=0    failed=0   

+ set +o xtrace
########## FINISHED STAGE: SUCCESS: PROVISION CLOUD RESOURCES [00h 01m 35s] ##########
[workspace] $ /bin/bash /tmp/jenkins232126286039077755.sh
########## STARTING STAGE: FORWARD GCS CREDENTIALS TO REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
+ (( i = 0 ))
+ (( i < 10 ))
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.config/gcloud/gcs-publisher-credentials.json openshiftdevel:/data/credentials.json
+ break
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: FORWARD GCS CREDENTIALS TO REMOTE HOST [00h 00m 02s] ##########
[workspace] $ /bin/bash /tmp/jenkins6929665636970042056.sh
########## STARTING STAGE: FORWARD PARAMETERS TO THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod o+rw /etc/environment
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''JOB_SPEC={"type":"presubmit","job":"pull-ci-openshift-machine-api-operator-master-e2e","buildid":"1169987167749935104","prowjobid":"2695ec19-d0b6-11e9-a06a-0a58ac108d5e","refs":{"org":"openshift","repo":"machine-api-operator","repo_link":"https://github.com/openshift/machine-api-operator","base_ref":"master","base_sha":"474e14e4965a8c5e6788417c851ccc7fad1acb3a","base_link":"https://github.com/openshift/machine-api-operator/commit/474e14e4965a8c5e6788417c851ccc7fad1acb3a","pulls":[{"number":389,"author":"sadasu","sha":"229c7ea627e98ef3b7c1927a25352d366fea7023","link":"https://github.com/openshift/machine-api-operator/pull/389","commit_link":"https://github.com/openshift/machine-api-operator/pull/389/commits/229c7ea627e98ef3b7c1927a25352d366fea7023","author_link":"https://github.com/sadasu"}]}}'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''buildId='\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''BUILD_ID=1169987167749935104'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''REPO_OWNER=openshift'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''REPO_NAME=machine-api-operator'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_BASE_REF=master'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_BASE_SHA=474e14e4965a8c5e6788417c851ccc7fad1acb3a'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_REFS=master:474e14e4965a8c5e6788417c851ccc7fad1acb3a,389:229c7ea627e98ef3b7c1927a25352d366fea7023'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_NUMBER=389'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_PULL_SHA=229c7ea627e98ef3b7c1927a25352d366fea7023'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''JOB_SPEC={"type":"presubmit","job":"pull-ci-openshift-machine-api-operator-master-e2e","buildid":"1169987167749935104","prowjobid":"2695ec19-d0b6-11e9-a06a-0a58ac108d5e","refs":{"org":"openshift","repo":"machine-api-operator","repo_link":"https://github.com/openshift/machine-api-operator","base_ref":"master","base_sha":"474e14e4965a8c5e6788417c851ccc7fad1acb3a","base_link":"https://github.com/openshift/machine-api-operator/commit/474e14e4965a8c5e6788417c851ccc7fad1acb3a","pulls":[{"number":389,"author":"sadasu","sha":"229c7ea627e98ef3b7c1927a25352d366fea7023","link":"https://github.com/openshift/machine-api-operator/pull/389","commit_link":"https://github.com/openshift/machine-api-operator/pull/389/commits/229c7ea627e98ef3b7c1927a25352d366fea7023","author_link":"https://github.com/sadasu"}]}}'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''BUILD_NUMBER=716'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''CLONEREFS_ARGS='\'' >> /etc/environment'
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: FORWARD PARAMETERS TO THE REMOTE HOST [00h 00m 05s] ##########
[workspace] $ /bin/bash /tmp/jenkins5952074927674220828.sh
########## STARTING STAGE: SYNC REPOSITORIES ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.SGyLBz8p28
+ cat
+ chmod +x /tmp/tmp.SGyLBz8p28
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.SGyLBz8p28 openshiftdevel:/tmp/tmp.SGyLBz8p28
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.SGyLBz8p28"'
+ cd /home/origin
++ jq --compact-output '.buildid |= "716"'
+ JOB_SPEC='{"type":"presubmit","job":"pull-ci-openshift-machine-api-operator-master-e2e","buildid":"716","prowjobid":"2695ec19-d0b6-11e9-a06a-0a58ac108d5e","refs":{"org":"openshift","repo":"machine-api-operator","repo_link":"https://github.com/openshift/machine-api-operator","base_ref":"master","base_sha":"474e14e4965a8c5e6788417c851ccc7fad1acb3a","base_link":"https://github.com/openshift/machine-api-operator/commit/474e14e4965a8c5e6788417c851ccc7fad1acb3a","pulls":[{"number":389,"author":"sadasu","sha":"229c7ea627e98ef3b7c1927a25352d366fea7023","link":"https://github.com/openshift/machine-api-operator/pull/389","commit_link":"https://github.com/openshift/machine-api-operator/pull/389/commits/229c7ea627e98ef3b7c1927a25352d366fea7023","author_link":"https://github.com/sadasu"}]}}'
+ for image in ''\''registry.svc.ci.openshift.org/ci/clonerefs:latest'\''' ''\''registry.svc.ci.openshift.org/ci/initupload:latest'\'''
+ (( i = 0 ))
+ (( i < 5 ))
+ docker pull registry.svc.ci.openshift.org/ci/clonerefs:latest
Trying to pull repository registry.svc.ci.openshift.org/ci/clonerefs ... 
latest: Pulling from registry.svc.ci.openshift.org/ci/clonerefs
1160f4abea84: Pulling fs layer
be60dbe7622d: Pulling fs layer
d26b76701841: Pulling fs layer
1b90cab916ea: Pulling fs layer
3a00cbb24bdb: Pulling fs layer
1b90cab916ea: Waiting
3a00cbb24bdb: Waiting
be60dbe7622d: Verifying Checksum
be60dbe7622d: Download complete
1160f4abea84: Verifying Checksum
1160f4abea84: Download complete
3a00cbb24bdb: Verifying Checksum
3a00cbb24bdb: Download complete
d26b76701841: Verifying Checksum
d26b76701841: Download complete
1b90cab916ea: Verifying Checksum
1b90cab916ea: Download complete
1160f4abea84: Pull complete
be60dbe7622d: Pull complete
d26b76701841: Pull complete
1b90cab916ea: Pull complete
3a00cbb24bdb: Pull complete
Digest: sha256:d68e1c6c2de5c1167a79b24d5ba4f909349ca7a44fb634e214bdadc2c8b010cd
Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/clonerefs:latest
+ break
+ for image in ''\''registry.svc.ci.openshift.org/ci/clonerefs:latest'\''' ''\''registry.svc.ci.openshift.org/ci/initupload:latest'\'''
+ (( i = 0 ))
+ (( i < 5 ))
+ docker pull registry.svc.ci.openshift.org/ci/initupload:latest
Trying to pull repository registry.svc.ci.openshift.org/ci/initupload ... 
latest: Pulling from registry.svc.ci.openshift.org/ci/initupload
a073c86ecf9e: Pulling fs layer
cc3fc741b1a9: Pulling fs layer
8f72556ef119: Pulling fs layer
8e5b170ec95b: Pulling fs layer
8e5b170ec95b: Waiting
cc3fc741b1a9: Verifying Checksum
cc3fc741b1a9: Download complete
a073c86ecf9e: Verifying Checksum
a073c86ecf9e: Download complete
8e5b170ec95b: Verifying Checksum
8e5b170ec95b: Download complete
8f72556ef119: Verifying Checksum
8f72556ef119: Download complete
a073c86ecf9e: Pull complete
cc3fc741b1a9: Pull complete
8f72556ef119: Pull complete
8e5b170ec95b: Pull complete
Digest: sha256:e651a6455ada7c070c439eddcd753e2e2ac1fb934c4f2a526c37a4674c8eaee4
Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/initupload:latest
+ break
+ clonerefs_args='--repo=openshift,cluster-api-provider-kubemark=master --repo=openshift,cluster-autoscaler-operator=master --repo=openshift,kubernetes-autoscaler=master '
+ docker run -v /data:/data:z registry.svc.ci.openshift.org/ci/clonerefs:latest --src-root=/data --log=/data/clone.json --repo=openshift,machine-api-operator=master:474e14e4965a8c5e6788417c851ccc7fad1acb3a,389:229c7ea627e98ef3b7c1927a25352d366fea7023 --repo=openshift,cluster-api-provider-kubemark=master --repo=openshift,cluster-autoscaler-operator=master --repo=openshift,kubernetes-autoscaler=master
{"component":"clonerefs","file":"prow/pod-utils/clone/clone.go:34","func":"k8s.io/test-infra/prow/pod-utils/clone.Run","level":"info","msg":"Cloning refs","refs":{"org":"openshift","repo":"machine-api-operator","base_ref":"master","base_sha":"474e14e4965a8c5e6788417c851ccc7fad1acb3a","pulls":[{"number":389,"author":"","sha":"229c7ea627e98ef3b7c1927a25352d366fea7023"}]},"time":"2019-09-06T14:56:59Z"}
{"component":"clonerefs","file":"prow/pod-utils/clone/clone.go:34","func":"k8s.io/test-infra/prow/pod-utils/clone.Run","level":"info","msg":"Cloning refs","refs":{"org":"openshift","repo":"kubernetes-autoscaler","base_ref":"master"},"time":"2019-09-06T14:56:59Z"}
{"component":"clonerefs","file":"prow/pod-utils/clone/clone.go:34","func":"k8s.io/test-infra/prow/pod-utils/clone.Run","level":"info","msg":"Cloning refs","refs":{"org":"openshift","repo":"cluster-api-provider-kubemark","base_ref":"master"},"time":"2019-09-06T14:56:59Z"}
{"component":"clonerefs","file":"prow/pod-utils/clone/clone.go:34","func":"k8s.io/test-infra/prow/pod-utils/clone.Run","level":"info","msg":"Cloning refs","refs":{"org":"openshift","repo":"cluster-autoscaler-operator","base_ref":"master"},"time":"2019-09-06T14:56:59Z"}
{"command":"mkdir -p /data/src/github.com/openshift/cluster-api-provider-kubemark","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:56:59Z"}
{"command":"mkdir -p /data/src/github.com/openshift/kubernetes-autoscaler","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:56:59Z"}
{"command":"mkdir -p /data/src/github.com/openshift/cluster-autoscaler-operator","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:56:59Z"}
{"command":"mkdir -p /data/src/github.com/openshift/machine-api-operator","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:56:59Z"}
{"command":"git init","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Initialized empty Git repository in /data/src/github.com/openshift/kubernetes-autoscaler/.git/\n","time":"2019-09-06T14:56:59Z"}
{"command":"git init","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Initialized empty Git repository in /data/src/github.com/openshift/cluster-api-provider-kubemark/.git/\n","time":"2019-09-06T14:56:59Z"}
{"command":"git init","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Initialized empty Git repository in /data/src/github.com/openshift/cluster-autoscaler-operator/.git/\n","time":"2019-09-06T14:57:00Z"}
{"command":"git init","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Initialized empty Git repository in /data/src/github.com/openshift/machine-api-operator/.git/\n","time":"2019-09-06T14:57:00Z"}
{"command":"git config user.name ci-robot","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:00Z"}
{"command":"git config user.name ci-robot","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:00Z"}
{"command":"git config user.name ci-robot","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:00Z"}
{"command":"git config user.name ci-robot","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:00Z"}
{"command":"git config user.email ci-robot@k8s.io","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:00Z"}
{"command":"git config user.email ci-robot@k8s.io","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:00Z"}
{"command":"git config user.email ci-robot@k8s.io","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:00Z"}
{"command":"git config user.email ci-robot@k8s.io","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:00Z"}
{"command":"git fetch https://github.com/openshift/machine-api-operator.git --tags --prune","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"From https://github.com/openshift/machine-api-operator\n * branch            HEAD       -\u003e FETCH_HEAD\n * [new tag]         v0.1.0     -\u003e v0.1.0\n * [new tag]         v0.2.0     -\u003e v0.2.0\n","time":"2019-09-06T14:57:02Z"}
{"command":"git fetch https://github.com/openshift/machine-api-operator.git master","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"From https://github.com/openshift/machine-api-operator\n * branch            master     -\u003e FETCH_HEAD\n","time":"2019-09-06T14:57:02Z"}
{"command":"git checkout 474e14e4965a8c5e6788417c851ccc7fad1acb3a","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Note: checking out '474e14e4965a8c5e6788417c851ccc7fad1acb3a'.\n\nYou are in 'detached HEAD' state. You can look around, make experimental\nchanges and commit them, and you can discard any commits you make in this\nstate without impacting any branches by performing another checkout.\n\nIf you want to create a new branch to retain commits you create, you may\ndo so (now or later) by using -b with the checkout command again. Example:\n\n  git checkout -b \u003cnew-branch-name\u003e\n\nHEAD is now at 474e14e... Merge pull request #391 from mgugino-upstream-stage/related-resources\n","time":"2019-09-06T14:57:03Z"}
{"command":"git branch --force master 474e14e4965a8c5e6788417c851ccc7fad1acb3a","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:03Z"}
{"command":"git checkout master","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Switched to branch 'master'\n","time":"2019-09-06T14:57:03Z"}
{"command":"git fetch https://github.com/openshift/machine-api-operator.git pull/389/head","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"From https://github.com/openshift/machine-api-operator\n * branch            refs/pull/389/head -\u003e FETCH_HEAD\n","time":"2019-09-06T14:57:03Z"}
{"command":"git merge --no-ff 229c7ea627e98ef3b7c1927a25352d366fea7023","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Merge made by the 'recursive' strategy.\n pkg/operator/baremetal_pod.go | 64 +++++++++++++++++++++++++++++++++++++++----\n pkg/operator/sync.go          |  7 +++++\n 2 files changed, 65 insertions(+), 6 deletions(-)\n","time":"2019-09-06T14:57:03Z"}
{"command":"git fetch https://github.com/openshift/cluster-autoscaler-operator.git --tags --prune","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"From https://github.com/openshift/cluster-autoscaler-operator\n * branch              HEAD       -\u003e FETCH_HEAD\n * [new tag]           v0.0.0     -\u003e v0.0.0\n","time":"2019-09-06T14:57:03Z"}
{"command":"git submodule update --init --recursive","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:03Z"}
{"command":"git fetch https://github.com/openshift/cluster-autoscaler-operator.git master","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"From https://github.com/openshift/cluster-autoscaler-operator\n * branch              master     -\u003e FETCH_HEAD\n","time":"2019-09-06T14:57:03Z"}
{"command":"git checkout FETCH_HEAD","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Note: checking out 'FETCH_HEAD'.\n\nYou are in 'detached HEAD' state. You can look around, make experimental\nchanges and commit them, and you can discard any commits you make in this\nstate without impacting any branches by performing another checkout.\n\nIf you want to create a new branch to retain commits you create, you may\ndo so (now or later) by using -b with the checkout command again. Example:\n\n  git checkout -b \u003cnew-branch-name\u003e\n\nHEAD is now at 045aea45... Merge pull request #117 from enxebre/more-related-objects\n","time":"2019-09-06T14:57:04Z"}
{"command":"git branch --force master FETCH_HEAD","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:04Z"}
{"command":"git checkout master","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Switched to branch 'master'\n","time":"2019-09-06T14:57:04Z"}
{"command":"git submodule update --init --recursive","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:04Z"}
{"command":"git fetch https://github.com/openshift/cluster-api-provider-kubemark.git --tags --prune","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"From https://github.com/openshift/cluster-api-provider-kubemark\n * branch              HEAD       -\u003e FETCH_HEAD\n * [new tag]           v1.0       -\u003e v1.0\n","time":"2019-09-06T14:57:04Z"}
{"command":"git fetch https://github.com/openshift/cluster-api-provider-kubemark.git master","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"From https://github.com/openshift/cluster-api-provider-kubemark\n * branch              master     -\u003e FETCH_HEAD\n","time":"2019-09-06T14:57:04Z"}
{"command":"git checkout FETCH_HEAD","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Note: checking out 'FETCH_HEAD'.\n\nYou are in 'detached HEAD' state. You can look around, make experimental\nchanges and commit them, and you can discard any commits you make in this\nstate without impacting any branches by performing another checkout.\n\nIf you want to create a new branch to retain commits you create, you may\ndo so (now or later) by using -b with the checkout command again. Example:\n\n  git checkout -b \u003cnew-branch-name\u003e\n\nHEAD is now at 45659b31... Merge pull request #27 from frobware/bump-openshift-cluster-api-deps\n","time":"2019-09-06T14:57:06Z"}
{"command":"git branch --force master FETCH_HEAD","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:06Z"}
{"command":"git checkout master","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Switched to branch 'master'\n","time":"2019-09-06T14:57:06Z"}
{"command":"git submodule update --init --recursive","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:06Z"}
{"command":"git fetch https://github.com/openshift/kubernetes-autoscaler.git --tags --prune","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"From https://github.com/openshift/kubernetes-autoscaler\n * branch                HEAD                        -\u003e FETCH_HEAD\n * [new tag]             addon-resizer-1.8.0         -\u003e addon-resizer-1.8.0\n * [new tag]             addon-resizer-1.8.1         -\u003e addon-resizer-1.8.1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.37.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.37.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.38.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.38.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.39.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.39.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.40.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.40.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.41.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.41.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.42.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.42.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.43.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.43.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.44.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.44.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.46.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.46.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.47.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.47.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.50.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.50.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.51.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.51.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.52.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.52.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.53.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.53.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.54.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.54.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.56.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.56.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.57.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.57.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.58.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.58.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.60.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.60.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.61.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.61.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.63.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.63.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.64.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.64.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.65.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.65.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.66.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.66.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.67.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.67.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.68.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.68.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-0.69.0 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-0.69.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.0-1666 -\u003e atomic-openshift-cluster-autoscaler-3.10.0-1666\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.1-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.1-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.10-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.10-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.11-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.11-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.12-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.12-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.13-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.13-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.14-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.14-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.15-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.15-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.16-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.16-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.17-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.17-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.18-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.18-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.2-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.2-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.21-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.21-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.22-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.22-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.23-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.23-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.3-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.3-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.5-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.5-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.6-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.6-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.7-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.7-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.8-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.8-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.10.9-1 -\u003e atomic-openshift-cluster-autoscaler-3.10.9-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.10.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.10.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.11.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.11.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.13.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.13.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.14.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.14.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.15.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.15.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.16.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.16.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.17.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.17.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.18.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.18.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.19.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.19.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.20.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.20.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.21.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.21.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.22.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.22.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.23.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.23.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.24.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.24.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.25.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.25.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.26.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.26.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.27.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.27.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.28.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.28.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.30.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.30.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.32.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.32.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.5.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.5.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.7.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.7.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.8.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.8.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.0-0.9.0 -\u003e atomic-openshift-cluster-autoscaler-3.11.0-0.9.0\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.100-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.100-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.104-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.104-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.105-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.105-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.106-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.106-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.107-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.107-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.108-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.108-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.109-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.109-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.11-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.11-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.110-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.110-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.111-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.111-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.112-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.112-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.113-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.113-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.114-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.114-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.115-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.115-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.116-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.116-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.117-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.117-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.119-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.119-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.12-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.12-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.121-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.121-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.122-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.122-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.123-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.123-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.124-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.124-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.125-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.125-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.126-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.126-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.127-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.127-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.128-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.128-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.129-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.129-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.13-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.13-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.130-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.130-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.131-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.131-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.132-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.132-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.133-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.133-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.134-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.134-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.135-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.135-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.136-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.136-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.137-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.137-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.138-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.138-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.139-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.139-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.14-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.14-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.140-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.140-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.141-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.141-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.142-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.142-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.143-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.143-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.15-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.15-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.16-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.16-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.17-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.17-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.18-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.18-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.19-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.19-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.20-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.20-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.21-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.21-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.22-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.22-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.23-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.23-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.24-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.24-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.25-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.25-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.26-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.26-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.27-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.27-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.28-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.28-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.29-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.29-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.3-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.3-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.30-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.30-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.31-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.31-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.32-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.32-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.33-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.33-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.34-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.34-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.35-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.35-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.36-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.36-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.37-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.37-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.38-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.38-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.39-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.39-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.4-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.4-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.40-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.40-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.41-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.41-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.42-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.42-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.43-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.43-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.44-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.44-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.45-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.45-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.46-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.46-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.47-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.47-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.49-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.49-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.5-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.5-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.50-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.50-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.51-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.51-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.53-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.53-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.54-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.54-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.55-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.55-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.56-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.56-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.57-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.57-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.58-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.58-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.59-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.59-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.6-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.6-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.60-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.60-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.61-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.61-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.62-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.62-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.63-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.63-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.64-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.64-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.65-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.65-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.66-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.66-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.67-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.67-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.69-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.69-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.7-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.7-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.71-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.71-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.72-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.72-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.73-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.73-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.74-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.74-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.75-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.75-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.76-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.76-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.77-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.77-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.78-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.78-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.79-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.79-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.8-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.8-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.81-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.81-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.82-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.82-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.83-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.83-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.85-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.85-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.86-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.86-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.87-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.87-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.88-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.88-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.9-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.9-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.90-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.90-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.91-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.91-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.92-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.92-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.93-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.93-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.94-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.94-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.95-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.95-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.96-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.96-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.97-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.97-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.98-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.98-1\n * [new tag]             atomic-openshift-cluster-autoscaler-3.11.99-1 -\u003e atomic-openshift-cluster-autoscaler-3.11.99-1\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.10.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.10.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.100.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.100.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.101.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.101.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.102.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.102.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.103.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.103.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.104.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.104.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.105.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.105.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.106.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.106.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.107.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.107.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.109.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.109.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.110.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.110.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.112.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.112.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.114.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.114.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.115.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.115.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.116.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.116.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.117.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.117.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.118.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.118.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.119.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.119.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.12.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.12.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.122.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.122.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.123.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.123.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.124.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.124.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.125.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.125.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.128.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.128.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.13.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.13.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.130.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.130.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.131.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.131.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.132.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.132.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.136.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.136.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.137.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.137.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.138.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.138.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.139.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.139.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.14.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.14.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.140.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.140.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.141.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.141.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.142.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.142.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.143.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.143.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.144.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.144.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.145.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.145.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.146.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.146.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.147.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.147.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.148.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.148.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.149.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.149.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.15.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.15.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.17.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.17.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.18.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.18.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.19.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.19.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.20.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.20.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.21.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.21.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.22.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.22.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.23.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.23.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.24.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.24.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.25.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.25.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.26.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.26.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.27.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.27.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.28.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.28.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.29.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.29.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.30.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.30.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.31.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.31.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.32.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.32.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.33.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.33.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.36.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.36.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.37.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.37.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.38.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.38.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.39.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.39.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.4.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.4.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.40.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.40.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.41.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.41.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.42.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.42.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.43.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.43.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.44.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.44.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.45.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.45.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.46.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.46.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.47.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.47.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.48.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.48.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.49.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.49.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.5.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.5.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.50.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.50.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.51.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.51.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.52.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.52.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.55.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.55.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.56.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.56.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.57.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.57.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.58.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.58.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.59.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.59.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.6.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.6.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.60.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.60.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.62.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.62.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.63.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.63.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.66.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.66.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.69.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.69.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.7.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.7.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.70.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.70.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.72.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.72.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.74.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.74.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.75.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.75.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.76.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.76.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.77.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.77.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.79.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.79.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.8.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.8.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.80.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.80.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.81.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.81.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.82.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.82.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.83.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.83.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.84.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.84.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.85.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.85.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.87.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.87.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.88.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.88.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.89.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.89.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.9.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.9.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.91.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.91.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.92.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.92.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.93.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.93.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.94.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.94.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.95.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.95.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.96.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.96.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.97.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.97.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.98.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.98.0\n * [new tag]             atomic-openshift-cluster-autoscaler-4.0.0-0.99.0 -\u003e atomic-openshift-cluster-autoscaler-4.0.0-0.99.0\n * [new tag]             cluster-autoscaler-0.5.2    -\u003e cluster-autoscaler-0.5.2\n * [new tag]             cluster-autoscaler-0.5.3    -\u003e cluster-autoscaler-0.5.3\n * [new tag]             cluster-autoscaler-0.5.4    -\u003e cluster-autoscaler-0.5.4\n * [new tag]             cluster-autoscaler-0.6.0    -\u003e cluster-autoscaler-0.6.0\n * [new tag]             cluster-autoscaler-0.6.1    -\u003e cluster-autoscaler-0.6.1\n * [new tag]             cluster-autoscaler-0.6.2    -\u003e cluster-autoscaler-0.6.2\n * [new tag]             cluster-autoscaler-0.6.3    -\u003e cluster-autoscaler-0.6.3\n * [new tag]             cluster-autoscaler-0.6.4    -\u003e cluster-autoscaler-0.6.4\n * [new tag]             cluster-autoscaler-1.0.0    -\u003e cluster-autoscaler-1.0.0\n * [new tag]             cluster-autoscaler-1.0.1    -\u003e cluster-autoscaler-1.0.1\n * [new tag]             cluster-autoscaler-1.0.2    -\u003e cluster-autoscaler-1.0.2\n * [new tag]             cluster-autoscaler-1.0.3    -\u003e cluster-autoscaler-1.0.3\n * [new tag]             cluster-autoscaler-1.0.4    -\u003e cluster-autoscaler-1.0.4\n * [new tag]             cluster-autoscaler-1.0.5    -\u003e cluster-autoscaler-1.0.5\n * [new tag]             cluster-autoscaler-1.1.0    -\u003e cluster-autoscaler-1.1.0\n * [new tag]             cluster-autoscaler-1.1.1    -\u003e cluster-autoscaler-1.1.1\n * [new tag]             cluster-autoscaler-1.1.2    -\u003e cluster-autoscaler-1.1.2\n * [new tag]             cluster-autoscaler-1.2.0    -\u003e cluster-autoscaler-1.2.0\n * [new tag]             cluster-autoscaler-1.2.1    -\u003e cluster-autoscaler-1.2.1\n * [new tag]             cluster-autoscaler-1.2.2    -\u003e cluster-autoscaler-1.2.2\n * [new tag]             v3.10.0                     -\u003e v3.10.0\n * [new tag]             v3.10.0-alpha.0             -\u003e v3.10.0-alpha.0\n * [new tag]             v3.10.0-rc.0                -\u003e v3.10.0-rc.0\n * [new tag]             v3.11                       -\u003e v3.11\n * [new tag]             v3.11.0                     -\u003e v3.11.0\n * [new tag]             v3.11.0-alpha.0             -\u003e v3.11.0-alpha.0\n * [new tag]             vertical-pod-autoscaler-0.1 -\u003e vertical-pod-autoscaler-0.1\n","time":"2019-09-06T14:57:08Z"}
{"command":"git fetch https://github.com/openshift/kubernetes-autoscaler.git master","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"From https://github.com/openshift/kubernetes-autoscaler\n * branch                master     -\u003e FETCH_HEAD\n","time":"2019-09-06T14:57:08Z"}
{"command":"git checkout FETCH_HEAD","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Note: checking out 'FETCH_HEAD'.\n\nYou are in 'detached HEAD' state. You can look around, make experimental\nchanges and commit them, and you can discard any commits you make in this\nstate without impacting any branches by performing another checkout.\n\nIf you want to create a new branch to retain commits you create, you may\ndo so (now or later) by using -b with the checkout command again. Example:\n\n  git checkout -b \u003cnew-branch-name\u003e\n\nHEAD is now at 18a08df11... Merge pull request #114 from ingvagabund/goimports-makefile\n","time":"2019-09-06T14:57:11Z"}
{"command":"git branch --force master FETCH_HEAD","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:11Z"}
{"command":"git checkout master","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Switched to branch 'master'\n","time":"2019-09-06T14:57:11Z"}
{"command":"git submodule update --init --recursive","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-06T14:57:11Z"}
{"component":"clonerefs","file":"prow/cmd/clonerefs/main.go:43","func":"main.main","level":"info","msg":"Finished cloning refs","time":"2019-09-06T14:57:11Z"}
+ docker run -e 'JOB_SPEC={"type":"presubmit","job":"pull-ci-openshift-machine-api-operator-master-e2e","buildid":"716","prowjobid":"2695ec19-d0b6-11e9-a06a-0a58ac108d5e","refs":{"org":"openshift","repo":"machine-api-operator","repo_link":"https://github.com/openshift/machine-api-operator","base_ref":"master","base_sha":"474e14e4965a8c5e6788417c851ccc7fad1acb3a","base_link":"https://github.com/openshift/machine-api-operator/commit/474e14e4965a8c5e6788417c851ccc7fad1acb3a","pulls":[{"number":389,"author":"sadasu","sha":"229c7ea627e98ef3b7c1927a25352d366fea7023","link":"https://github.com/openshift/machine-api-operator/pull/389","commit_link":"https://github.com/openshift/machine-api-operator/pull/389/commits/229c7ea627e98ef3b7c1927a25352d366fea7023","author_link":"https://github.com/sadasu"}]}}' -v /data:/data:z registry.svc.ci.openshift.org/ci/initupload:latest --clone-log=/data/clone.json --dry-run=false --gcs-path=gs://origin-ci-test --gcs-credentials-file=/data/credentials.json --path-strategy=single --default-org=openshift --default-repo=origin
{"component":"initupload","dest":"pr-logs/directory/pull-ci-openshift-machine-api-operator-master-e2e/716.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-06T14:57:14Z"}
{"component":"initupload","dest":"pr-logs/directory/pull-ci-openshift-machine-api-operator-master-e2e/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-06T14:57:14Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/389/pull-ci-openshift-machine-api-operator-master-e2e/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-06T14:57:14Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/389/pull-ci-openshift-machine-api-operator-master-e2e/716/clone-records.json","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-06T14:57:14Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/389/pull-ci-openshift-machine-api-operator-master-e2e/716/started.json","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-06T14:57:14Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/389/pull-ci-openshift-machine-api-operator-master-e2e/716/clone-log.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-06T14:57:14Z"}
{"component":"initupload","dest":"pr-logs/directory/pull-ci-openshift-machine-api-operator-master-e2e/716.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-06T14:57:15Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/389/pull-ci-openshift-machine-api-operator-master-e2e/716/clone-records.json","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-06T14:57:15Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/389/pull-ci-openshift-machine-api-operator-master-e2e/716/started.json","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-06T14:57:15Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/389/pull-ci-openshift-machine-api-operator-master-e2e/716/clone-log.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-06T14:57:15Z"}
{"component":"initupload","dest":"pr-logs/directory/pull-ci-openshift-machine-api-operator-master-e2e/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-06T14:57:15Z"}
{"component":"initupload","dest":"pr-logs/pull/openshift_machine-api-operator/389/pull-ci-openshift-machine-api-operator-master-e2e/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-06T14:57:15Z"}
{"component":"initupload","file":"prow/gcsupload/run.go:65","func":"k8s.io/test-infra/prow/gcsupload.Options.Run","level":"info","msg":"Finished upload to GCS","time":"2019-09-06T14:57:15Z"}
+ sudo chmod -R a+rwX /data
+ sudo chown -R origin:origin-git /data
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: SYNC REPOSITORIES [00h 01m 14s] ##########
[workspace] $ /bin/bash /tmp/jenkins3575262279710769501.sh
########## STARTING STAGE: FORWARD PARAMETERS TO THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod o+rw /etc/environment
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''JOB_NAME=pull-ci-openshift-machine-api-operator-master-e2e'\'' >> /etc/environment'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''BUILD_NUMBER=716'\'' >> /etc/environment'
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: FORWARD PARAMETERS TO THE REMOTE HOST [00h 00m 01s] ##########
[workspace] $ /bin/bash /tmp/jenkins1464767535790543971.sh
########## STARTING STAGE: INSTALL MINIKUBE ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.LPItgphIRQ
+ cat
+ chmod +x /tmp/tmp.LPItgphIRQ
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.LPItgphIRQ openshiftdevel:/tmp/tmp.LPItgphIRQ
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.LPItgphIRQ"'
+ cd /home/origin
+ curl -Lo minikube https://storage.googleapis.com/minikube/releases/v0.30.0/minikube-linux-amd64
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
 18 40.3M   18 7792k    0     0  8489k      0  0:00:04 --:--:--  0:00:04 8488k
100 40.3M  100 40.3M    0     0  31.0M      0  0:00:01  0:00:01 --:--:-- 31.0M
+ chmod +x minikube
+ sudo mv minikube /usr/bin/
+ curl -Lo kubectl https://storage.googleapis.com/kubernetes-release/release/v1.10.0/bin/linux/amd64/kubectl
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
 47 51.7M   47 24.4M    0     0  41.1M      0  0:00:01 --:--:--  0:00:01 41.1M
100 51.7M  100 51.7M    0     0  59.9M      0 --:--:-- --:--:-- --:--:-- 59.9M
+ chmod +x kubectl
+ sudo mv kubectl /usr/bin/
+ sudo yum install -y ebtables
Loaded plugins: amazon-id, rhui-lb, search-disabled-repos
Resolving Dependencies
--> Running transaction check
---> Package ebtables.x86_64 0:2.0.10-16.el7 will be installed
--> Finished Dependency Resolution

Dependencies Resolved

================================================================================
 Package     Arch      Version           Repository                        Size
================================================================================
Installing:
 ebtables    x86_64    2.0.10-16.el7     oso-rhui-rhel-server-releases    123 k

Transaction Summary
================================================================================
Install  1 Package

Total download size: 123 k
Installed size: 343 k
Downloading packages:
Running transaction check
Running transaction test
Transaction test succeeded
Running transaction
  Installing : ebtables-2.0.10-16.el7.x86_64                                1/1 
  Verifying  : ebtables-2.0.10-16.el7.x86_64                                1/1 

Installed:
  ebtables.x86_64 0:2.0.10-16.el7                                               

Complete!
+ VERSION=v1.13.0
+ wget https://github.com/kubernetes-sigs/cri-tools/releases/download/v1.13.0/crictl-v1.13.0-linux-amd64.tar.gz
--2019-09-06 14:58:34--  https://github.com/kubernetes-sigs/cri-tools/releases/download/v1.13.0/crictl-v1.13.0-linux-amd64.tar.gz
Resolving github.com (github.com)... 192.30.253.113
Connecting to github.com (github.com)|192.30.253.113|:443... connected.
HTTP request sent, awaiting response... 302 Found
Location: https://github-production-release-asset-2e65be.s3.amazonaws.com/80172100/61627180-fed9-11e8-9958-15e7eb90aa9e?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAIWNJYAX4CSVEH53A%2F20190906%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20190906T145834Z&X-Amz-Expires=300&X-Amz-Signature=24d06e03131cdff72203a5fe92a5ecfab493cd498332b2020c675ae64e6ad3bd&X-Amz-SignedHeaders=host&actor_id=0&response-content-disposition=attachment%3B%20filename%3Dcrictl-v1.13.0-linux-amd64.tar.gz&response-content-type=application%2Foctet-stream [following]
--2019-09-06 14:58:34--  https://github-production-release-asset-2e65be.s3.amazonaws.com/80172100/61627180-fed9-11e8-9958-15e7eb90aa9e?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAIWNJYAX4CSVEH53A%2F20190906%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20190906T145834Z&X-Amz-Expires=300&X-Amz-Signature=24d06e03131cdff72203a5fe92a5ecfab493cd498332b2020c675ae64e6ad3bd&X-Amz-SignedHeaders=host&actor_id=0&response-content-disposition=attachment%3B%20filename%3Dcrictl-v1.13.0-linux-amd64.tar.gz&response-content-type=application%2Foctet-stream
Resolving github-production-release-asset-2e65be.s3.amazonaws.com (github-production-release-asset-2e65be.s3.amazonaws.com)... 52.217.37.132
Connecting to github-production-release-asset-2e65be.s3.amazonaws.com (github-production-release-asset-2e65be.s3.amazonaws.com)|52.217.37.132|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: 10631149 (10M) [application/octet-stream]
Saving to: ‘crictl-v1.13.0-linux-amd64.tar.gz’

     0K .......... .......... .......... .......... ..........  0%  120M 0s
    50K .......... .......... .......... .......... ..........  0% 75.1M 0s
   100K .......... .......... .......... .......... ..........  1% 65.3M 0s
   150K .......... .......... .......... .......... ..........  1% 70.6M 0s
   200K .......... .......... .......... .......... ..........  2% 82.7M 0s
   250K .......... .......... .......... .......... ..........  2% 98.7M 0s
   300K .......... .......... .......... .......... ..........  3% 91.8M 0s
   350K .......... .......... .......... .......... ..........  3% 99.1M 0s
   400K .......... .......... .......... .......... ..........  4%  106M 0s
   450K .......... .......... .......... .......... ..........  4%  114M 0s
   500K .......... .......... .......... .......... ..........  5% 94.6M 0s
   550K .......... .......... .......... .......... ..........  5%  121M 0s
   600K .......... .......... .......... .......... ..........  6%  119M 0s
   650K .......... .......... .......... .......... ..........  6%  101M 0s
   700K .......... .......... .......... .......... ..........  7%  133M 0s
   750K .......... .......... .......... .......... ..........  7%  127M 0s
   800K .......... .......... .......... .......... ..........  8%  137M 0s
   850K .......... .......... .......... .......... ..........  8%  176M 0s
   900K .......... .......... .......... .......... ..........  9%  149M 0s
   950K .......... .......... .......... .......... ..........  9%  138M 0s
  1000K .......... .......... .......... .......... .......... 10%  111M 0s
  1050K .......... .......... .......... .......... .......... 10%  137M 0s
  1100K .......... .......... .......... .......... .......... 11%  138M 0s
  1150K .......... .......... .......... .......... .......... 11%  136M 0s
  1200K .......... .......... .......... .......... .......... 12%  131M 0s
  1250K .......... .......... .......... .......... .......... 12%  125M 0s
  1300K .......... .......... .......... .......... .......... 13%  162M 0s
  1350K .......... .......... .......... .......... .......... 13%  136M 0s
  1400K .......... .......... .......... .......... .......... 13%  133M 0s
  1450K .......... .......... .......... .......... .......... 14%  132M 0s
  1500K .......... .......... .......... .......... .......... 14%  150M 0s
  1550K .......... .......... .......... .......... .......... 15%  181M 0s
  1600K .......... .......... .......... .......... .......... 15%  126M 0s
  1650K .......... .......... .......... .......... .......... 16%  123M 0s
  1700K .......... .......... .......... .......... .......... 16%  120M 0s
  1750K .......... .......... .......... .......... .......... 17%  178M 0s
  1800K .......... .......... .......... .......... .......... 17%  185M 0s
  1850K .......... .......... .......... .......... .......... 18%  189M 0s
  1900K .......... .......... .......... .......... .......... 18%  134M 0s
  1950K .......... .......... .......... .......... .......... 19%  140M 0s
  2000K .......... .......... .......... .......... .......... 19%  124M 0s
  2050K .......... .......... .......... .......... .......... 20%  199M 0s
  2100K .......... .......... .......... .......... .......... 20%  198M 0s
  2150K .......... .......... .......... .......... .......... 21%  207M 0s
  2200K .......... .......... .......... .......... .......... 21%  119M 0s
  2250K .......... .......... .......... .......... .......... 22%  120M 0s
  2300K .......... .......... .......... .......... .......... 22%  130M 0s
  2350K .......... .......... .......... .......... .......... 23%  165M 0s
  2400K .......... .......... .......... .......... .......... 23%  184M 0s
  2450K .......... .......... .......... .......... .......... 24% 55.4M 0s
  2500K .......... .......... .......... .......... .......... 24%  118M 0s
  2550K .......... .......... .......... .......... .......... 25%  128M 0s
  2600K .......... .......... .......... .......... .......... 25%  164M 0s
  2650K .......... .......... .......... .......... .......... 26%  179M 0s
  2700K .......... .......... .......... .......... .......... 26%  138M 0s
  2750K .......... .......... .......... .......... .......... 26%  113M 0s
  2800K .......... .......... .......... .......... .......... 27%  138M 0s
  2850K .......... .......... .......... .......... .......... 27%  132M 0s
  2900K .......... .......... .......... .......... .......... 28%  141M 0s
  2950K .......... .......... .......... .......... .......... 28% 76.8M 0s
  3000K .......... .......... .......... .......... .......... 29%  121M 0s
  3050K .......... .......... .......... .......... .......... 29% 85.0M 0s
  3100K .......... .......... .......... .......... .......... 30% 86.6M 0s
  3150K .......... .......... .......... .......... .......... 30% 98.3M 0s
  3200K .......... .......... .......... .......... .......... 31% 85.1M 0s
  3250K .......... .......... .......... .......... .......... 31% 86.0M 0s
  3300K .......... .......... .......... .......... .......... 32% 98.9M 0s
  3350K .......... .......... .......... .......... .......... 32% 85.3M 0s
  3400K .......... .......... .......... .......... .......... 33%  117M 0s
  3450K .......... .......... .......... .......... .......... 33% 98.0M 0s
  3500K .......... .......... .......... .......... .......... 34% 83.5M 0s
  3550K .......... .......... .......... .......... .......... 34%  103M 0s
  3600K .......... .......... .......... .......... .......... 35% 88.6M 0s
  3650K .......... .......... .......... .......... .......... 35% 87.8M 0s
  3700K .......... .......... .......... .......... .......... 36% 85.5M 0s
  3750K .......... .......... .......... .......... .......... 36%  108M 0s
  3800K .......... .......... .......... .......... .......... 37% 84.3M 0s
  3850K .......... .......... .......... .......... .......... 37%  103M 0s
  3900K .......... .......... .......... .......... .......... 38% 84.3M 0s
  3950K .......... .......... .......... .......... .......... 38% 86.4M 0s
  4000K .......... .......... .......... .......... .......... 39%  102M 0s
  4050K .......... .......... .......... .......... .......... 39% 95.9M 0s
  4100K .......... .......... .......... .......... .......... 39% 85.8M 0s
  4150K .......... .......... .......... .......... .......... 40% 85.3M 0s
  4200K .......... .......... .......... .......... .......... 40% 98.1M 0s
  4250K .......... .......... .......... .......... .......... 41%  119M 0s
  4300K .......... .......... .......... .......... .......... 41% 90.9M 0s
  4350K .......... .......... .......... .......... .......... 42% 89.2M 0s
  4400K .......... .......... .......... .......... .......... 42%  102M 0s
  4450K .......... .......... .......... .......... .......... 43% 78.1M 0s
  4500K .......... .......... .......... .......... .......... 43%  102M 0s
  4550K .......... .......... .......... .......... .......... 44% 94.8M 0s
  4600K .......... .......... .......... .......... .......... 44% 88.4M 0s
  4650K .......... .......... .......... .......... .......... 45% 96.9M 0s
  4700K .......... .......... .......... .......... .......... 45% 98.6M 0s
  4750K .......... .......... .......... .......... .......... 46% 83.3M 0s
  4800K .......... .......... .......... .......... .......... 46%  102M 0s
  4850K .......... .......... .......... .......... .......... 47%  100M 0s
  4900K .......... .......... .......... .......... .......... 47% 81.9M 0s
  4950K .......... .......... .......... .......... .......... 48%  104M 0s
  5000K .......... .......... .......... .......... .......... 48% 82.7M 0s
  5050K .......... .......... .......... .......... .......... 49%  101M 0s
  5100K .......... .......... .......... .......... .......... 49%  111M 0s
  5150K .......... .......... .......... .......... .......... 50% 93.3M 0s
  5200K .......... .......... .......... .......... .......... 50% 97.7M 0s
  5250K .......... .......... .......... .......... .......... 51% 86.2M 0s
  5300K .......... .......... .......... .......... .......... 51% 99.8M 0s
  5350K .......... .......... .......... .......... .......... 52% 88.9M 0s
  5400K .......... .......... .......... .......... .......... 52%  102M 0s
  5450K .......... .......... .......... .......... .......... 52% 93.0M 0s
  5500K .......... .......... .......... .......... .......... 53% 99.1M 0s
  5550K .......... .......... .......... .......... .......... 53% 93.6M 0s
  5600K .......... .......... .......... .......... .......... 54% 97.4M 0s
  5650K .......... .......... .......... .......... .......... 54% 86.2M 0s
  5700K .......... .......... .......... .......... .......... 55% 99.6M 0s
  5750K .......... .......... .......... .......... .......... 55% 98.8M 0s
  5800K .......... .......... .......... .......... .......... 56% 85.2M 0s
  5850K .......... .......... .......... .......... .......... 56%  103M 0s
  5900K .......... .......... .......... .......... .......... 57% 80.8M 0s
  5950K .......... .......... .......... .......... .......... 57%  119M 0s
  6000K .......... .......... .......... .......... .......... 58% 15.6M 0s
  6050K .......... .......... .......... .......... .......... 58% 85.8M 0s
  6100K .......... .......... .......... .......... .......... 59%  124M 0s
  6150K .......... .......... .......... .......... .......... 59%  123M 0s
  6200K .......... .......... .......... .......... .......... 60%  124M 0s
  6250K .......... .......... .......... .......... .......... 60%  141M 0s
  6300K .......... .......... .......... .......... .......... 61%  138M 0s
  6350K .......... .......... .......... .......... .......... 61%  147M 0s
  6400K .......... .......... .......... .......... .......... 62%  153M 0s
  6450K .......... .......... .......... .......... .......... 62%  133M 0s
  6500K .......... .......... .......... .......... .......... 63%  143M 0s
  6550K .......... .......... .......... .......... .......... 63%  144M 0s
  6600K .......... .......... .......... .......... .......... 64%  141M 0s
  6650K .......... .......... .......... .......... .......... 64%  129M 0s
  6700K .......... .......... .......... .......... .......... 65%  157M 0s
  6750K .......... .......... .......... .......... .......... 65%  141M 0s
  6800K .......... .......... .......... .......... .......... 65%  157M 0s
  6850K .......... .......... .......... .......... .......... 66%  148M 0s
  6900K .......... .......... .......... .......... .......... 66%  102M 0s
  6950K .......... .......... .......... .......... .......... 67% 82.9M 0s
  7000K .......... .......... .......... .......... .......... 67% 86.1M 0s
  7050K .......... .......... .......... .......... .......... 68% 99.4M 0s
  7100K .......... .......... .......... .......... .......... 68% 84.9M 0s
  7150K .......... .......... .......... .......... .......... 69% 86.1M 0s
  7200K .......... .......... .......... .......... .......... 69% 99.8M 0s
  7250K .......... .......... .......... .......... .......... 70% 92.2M 0s
  7300K .......... .......... .......... .......... .......... 70%  101M 0s
  7350K .......... .......... .......... .......... .......... 71% 86.0M 0s
  7400K .......... .......... .......... .......... .......... 71% 84.6M 0s
  7450K .......... .......... .......... .......... .......... 72% 85.2M 0s
  7500K .......... .......... .......... .......... .......... 72%  103M 0s
  7550K .......... .......... .......... .......... .......... 73% 82.0M 0s
  7600K .......... .......... .......... .......... .......... 73% 85.8M 0s
  7650K .......... .......... .......... .......... .......... 74%  118M 0s
  7700K .......... .......... .......... .......... .......... 74% 95.0M 0s
  7750K .......... .......... .......... .......... .......... 75% 92.2M 0s
  7800K .......... .......... .......... .......... .......... 75% 94.6M 0s
  7850K .......... .......... .......... .......... .......... 76% 82.8M 0s
  7900K .......... .......... .......... .......... .......... 76% 99.4M 0s
  7950K .......... .......... .......... .......... .......... 77% 79.9M 0s
  8000K .......... .......... .......... .......... .......... 77% 86.7M 0s
  8050K .......... .......... .......... .......... .......... 78% 85.4M 0s
  8100K .......... .......... .......... .......... .......... 78%  113M 0s
  8150K .......... .......... .......... .......... .......... 78% 84.9M 0s
  8200K .......... .......... .......... .......... .......... 79% 84.6M 0s
  8250K .......... .......... .......... .......... .......... 79% 97.9M 0s
  8300K .......... .......... .......... .......... .......... 80% 86.6M 0s
  8350K .......... .......... .......... .......... .......... 80% 97.6M 0s
  8400K .......... .......... .......... .......... .......... 81% 85.0M 0s
  8450K .......... .......... .......... .......... .......... 81% 85.2M 0s
  8500K .......... .......... .......... .......... .......... 82%  118M 0s
  8550K .......... .......... .......... .......... .......... 82% 90.2M 0s
  8600K .......... .......... .......... .......... .......... 83% 89.8M 0s
  8650K .......... .......... .......... .......... .......... 83% 92.7M 0s
  8700K .......... .......... .......... .......... .......... 84% 90.8M 0s
  8750K .......... .......... .......... .......... .......... 84% 92.1M 0s
  8800K .......... .......... .......... .......... .......... 85% 87.5M 0s
  8850K .......... .......... .......... .......... .......... 85% 91.6M 0s
  8900K .......... .......... .......... .......... .......... 86% 82.5M 0s
  8950K .......... .......... .......... .......... .......... 86%  105M 0s
  9000K .......... .......... .......... .......... .......... 87% 85.4M 0s
  9050K .......... .......... .......... .......... .......... 87% 86.9M 0s
  9100K .......... .......... .......... .......... .......... 88% 97.6M 0s
  9150K .......... .......... .......... .......... .......... 88% 84.9M 0s
  9200K .......... .......... .......... .......... .......... 89% 99.2M 0s
  9250K .......... .......... .......... .......... .......... 89% 85.8M 0s
  9300K .......... .......... .......... .......... .......... 90% 85.7M 0s
  9350K .......... .......... .......... .......... .......... 90%  123M 0s
  9400K .......... .......... .......... .......... .......... 91% 84.4M 0s
  9450K .......... .......... .......... .......... .......... 91% 87.3M 0s
  9500K .......... .......... .......... .......... .......... 91% 97.7M 0s
  9550K .......... .......... .......... .......... .......... 92% 86.9M 0s
  9600K .......... .......... .......... .......... .......... 92% 88.9M 0s
  9650K .......... .......... .......... .......... .......... 93% 94.7M 0s
  9700K .......... .......... .......... .......... .......... 93% 85.3M 0s
  9750K .......... .......... .......... .......... .......... 94% 88.5M 0s
  9800K .......... .......... .......... .......... .......... 94%  105M 0s
  9850K .......... .......... .......... .......... .......... 95% 85.3M 0s
  9900K .......... .......... .......... .......... .......... 95% 86.5M 0s
  9950K .......... .......... .......... .......... .......... 96% 97.4M 0s
 10000K .......... .......... .......... .......... .......... 96% 85.6M 0s
 10050K .......... .......... .......... .......... .......... 97% 85.7M 0s
 10100K .......... .......... .......... .......... .......... 97% 97.6M 0s
 10150K .......... .......... .......... .......... .......... 98% 82.6M 0s
 10200K .......... .......... .......... .......... .......... 98% 90.8M 0s
 10250K .......... .......... .......... .......... .......... 99% 97.0M 0s
 10300K .......... .......... .......... .......... .......... 99% 98.6M 0s
 10350K .......... .......... .......... .                    100% 88.9M=0.1s

2019-09-06 14:58:35 (99.6 MB/s) - ‘crictl-v1.13.0-linux-amd64.tar.gz’ saved [10631149/10631149]

+ sudo tar zxvf crictl-v1.13.0-linux-amd64.tar.gz -C /usr/bin
crictl
+ rm -f crictl-v1.13.0-linux-amd64.tar.gz
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: INSTALL MINIKUBE [00h 01m 16s] ##########
[workspace] $ /bin/bash /tmp/jenkins181950169828746718.sh
########## STARTING STAGE: DEPLOY KUBERNETES ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.X3XuRj0mqp
+ cat
+ chmod +x /tmp/tmp.X3XuRj0mqp
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.X3XuRj0mqp openshiftdevel:/tmp/tmp.X3XuRj0mqp
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.X3XuRj0mqp"'
+ cd /home/origin
+ sudo setenforce 0
+ sudo minikube start --vm-driver=none --extra-config=kubelet.cgroup-driver=systemd --kubernetes-version v1.12.0 --v 5
Starting local Kubernetes v1.12.0 cluster...
Starting VM...
There is a newer version of minikube available (v1.3.1).  Download it here:
https://github.com/kubernetes/minikube/releases/tag/v1.3.1

To disable this notification, run the following:
minikube config set WantUpdateNotification false
Creating CA: /root/.minikube/certs/ca.pem
Creating client certificate: /root/.minikube/certs/cert.pem
Getting VM IP address...
Moving files into cluster...
Downloading kubeadm v1.12.0
Downloading kubelet v1.12.0
Finished Downloading kubeadm v1.12.0
Finished Downloading kubelet v1.12.0
Setting up certs...
Connecting to cluster...
Setting up kubeconfig...
Starting cluster components...
Kubectl is now configured to use the cluster.
===================
WARNING: IT IS RECOMMENDED NOT TO RUN THE NONE DRIVER ON PERSONAL WORKSTATIONS
	The 'none' driver will run an insecure kubernetes apiserver as root that may leave the host vulnerable to CSRF attacks

When using the none driver, the kubectl config and credentials generated will be root owned and will appear in the root home directory.
You will need to move the files to the appropriate location and then set the correct permissions.  An example of this is below:

	sudo mv /root/.kube $HOME/.kube # this will write over any previous configuration
	sudo chown -R $USER $HOME/.kube
	sudo chgrp -R $USER $HOME/.kube

	sudo mv /root/.minikube $HOME/.minikube # this will write over any previous configuration
	sudo chown -R $USER $HOME/.minikube
	sudo chgrp -R $USER $HOME/.minikube

This can also be done automatically by setting the env var CHANGE_MINIKUBE_NONE_USER=true
Loading cached images from config file.
+ sudo cp /etc/ssl/certs/ca-bundle.crt /etc/ssl/certs/ca-certificates.crt
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPLOY KUBERNETES [00h 00m 58s] ##########
[workspace] $ /bin/bash /tmp/jenkins4635626498762949358.sh
########## STARTING STAGE: INSTALL KUSTOMIZE ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.F9eJHPMX4m
+ cat
+ chmod +x /tmp/tmp.F9eJHPMX4m
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.F9eJHPMX4m openshiftdevel:/tmp/tmp.F9eJHPMX4m
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.F9eJHPMX4m"'
+ cd /home/origin
+ curl -Lo kustomize https://github.com/kubernetes-sigs/kustomize/releases/download/v2.1.0/kustomize_2.1.0_linux_amd64
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100   618    0   618    0     0   2095      0 --:--:-- --:--:-- --:--:--  2102

100 22.9M  100 22.9M    0     0  37.9M      0 --:--:-- --:--:-- --:--:-- 37.9M
+ chmod u+x kustomize
+ sudo mv kustomize /usr/bin/kustomize
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: INSTALL KUSTOMIZE [00h 00m 02s] ##########
[workspace] $ /bin/bash /tmp/jenkins7872975445483960859.sh
########## STARTING STAGE: INSTALL IMAGEBUILDER ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.mPZOqNDcWU
+ cat
+ chmod +x /tmp/tmp.mPZOqNDcWU
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.mPZOqNDcWU openshiftdevel:/tmp/tmp.mPZOqNDcWU
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.mPZOqNDcWU"'
+ cd /home/origin
+ export GOPATH=/data
+ GOPATH=/data
+ go get -u github.com/openshift/imagebuilder/cmd/imagebuilder
+ sudo mv /data/bin/imagebuilder /usr/bin
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: INSTALL IMAGEBUILDER [00h 00m 22s] ##########
[workspace] $ /bin/bash /tmp/jenkins2698717931702222361.sh
########## STARTING STAGE: BUILD KUBEMARK MACHINE CONTROLLERS ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.m4qUzCZVZC
+ cat
+ chmod +x /tmp/tmp.m4qUzCZVZC
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.m4qUzCZVZC openshiftdevel:/tmp/tmp.m4qUzCZVZC
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.m4qUzCZVZC"'
+ cd /home/origin
+ export GOPATH=/data
+ GOPATH=/data
+ cd /data/src/github.com/openshift/cluster-api-provider-kubemark
+ sudo make images IMAGE=docker.io/gofed/kubemark-machine-controllers VERSION=v1.0 NO_DOCKER=1
imagebuilder -t "docker.io/gofed/kubemark-machine-controllers:v1.0" -t "docker.io/gofed/kubemark-machine-controllers:latest" ./
--> Image registry.svc.ci.openshift.org/openshift/release:golang-1.10 was not found, pulling ...
--> Pulled 0/2 layers, 20% complete
--> Pulled 1/2 layers, 51% complete
--> Pulled 2/2 layers, 100% complete
--> Extracting
--> FROM registry.svc.ci.openshift.org/openshift/release:golang-1.10 as builder
--> WORKDIR /go/src/github.com/openshift/cluster-api-provider-kubemark
--> COPY . .
--> RUN go build -o ./machine-controller-manager ./cmd/manager
--> RUN go build -o ./manager ./vendor/github.com/openshift/cluster-api/cmd/manager
--> Image docker.io/gofed/base:baseci was not found, pulling ...
--> Pulled 1/2 layers, 73% complete
--> Pulled 2/2 layers, 100% complete
--> Extracting
--> FROM docker.io/gofed/base:baseci as 1
--> RUN INSTALL_PKGS="       openssh       " &&     yum install -y $INSTALL_PKGS &&     rpm -V $INSTALL_PKGS &&     yum clean all &&     curl -LO https://storage.googleapis.com/kubernetes-release/release/$(curl -s https://storage.googleapis.com/kubernetes-release/release/stable.txt)/bin/linux/amd64/kubectl &&     chmod +x ./kubectl &&     mv ./kubectl /bin/kubectl &&     curl -LO https://github.com/stedolan/jq/releases/download/jq-1.5/jq-linux64 &&     chmod +x ./jq-linux64 &&     mv ./jq-linux64 /bin/jq
Loaded plugins: fastestmirror, ovl
Determining fastest mirrors
 * base: mirrors.advancedhosters.com
 * extras: mirrors.advancedhosters.com
 * updates: mirrors.advancedhosters.com
Resolving Dependencies
--> Running transaction check
---> Package openssh.x86_64 0:7.4p1-16.el7 will be installed
--> Processing Dependency: libfipscheck.so.1()(64bit) for package: openssh-7.4p1-16.el7.x86_64
--> Running transaction check
---> Package fipscheck-lib.x86_64 0:1.4.1-6.el7 will be installed
--> Processing Dependency: /usr/bin/fipscheck for package: fipscheck-lib-1.4.1-6.el7.x86_64
--> Running transaction check
---> Package fipscheck.x86_64 0:1.4.1-6.el7 will be installed
--> Finished Dependency Resolution

Dependencies Resolved

================================================================================
 Package               Arch           Version                Repository    Size
================================================================================
Installing:
 openssh               x86_64         7.4p1-16.el7           base         510 k
Installing for dependencies:
 fipscheck             x86_64         1.4.1-6.el7            base          21 k
 fipscheck-lib         x86_64         1.4.1-6.el7            base          11 k

Transaction Summary
================================================================================
Install  1 Package (+2 Dependent packages)

Total download size: 542 k
Installed size: 2.0 M
Downloading packages:
--------------------------------------------------------------------------------
Total                                              759 kB/s | 542 kB  00:00     
Running transaction check
Running transaction test
Transaction test succeeded
Running transaction
  Installing : fipscheck-1.4.1-6.el7.x86_64                                 1/3 
  Installing : fipscheck-lib-1.4.1-6.el7.x86_64                             2/3 
  Installing : openssh-7.4p1-16.el7.x86_64                                  3/3 
  Verifying  : fipscheck-lib-1.4.1-6.el7.x86_64                             1/3 
  Verifying  : fipscheck-1.4.1-6.el7.x86_64                                 2/3 
  Verifying  : openssh-7.4p1-16.el7.x86_64                                  3/3 

Installed:
  openssh.x86_64 0:7.4p1-16.el7                                                 

Dependency Installed:
  fipscheck.x86_64 0:1.4.1-6.el7       fipscheck-lib.x86_64 0:1.4.1-6.el7      

Complete!
Loaded plugins: fastestmirror, ovl
Cleaning repos: base cbs-paas7-openshift-multiarch-el7-build extras updates
Cleaning up list of fastest mirrors
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100 40.9M  100 40.9M    0     0  66.7M      0 --:--:-- --:--:-- --:--:-- 66.8M
  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                 Dload  Upload   Total   Spent    Left  Speed

  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0
100   599    0   599    0     0   2593      0 --:--:-- --:--:-- --:--:--  2593

100 2956k  100 2956k    0     0  10.0M      0 --:--:-- --:--:-- --:--:-- 10.0M
--> COPY --from=builder /go/src/github.com/openshift/cluster-api-provider-kubemark/manager /
--> COPY --from=builder /go/src/github.com/openshift/cluster-api-provider-kubemark/machine-controller-manager /
--> Committing changes to docker.io/gofed/kubemark-machine-controllers:v1.0 ...
--> Tagged as docker.io/gofed/kubemark-machine-controllers:latest
--> Done
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: BUILD KUBEMARK MACHINE CONTROLLERS [00h 01m 37s] ##########
[workspace] $ /bin/bash /tmp/jenkins5384870064827320769.sh
########## STARTING STAGE: BUILD CLUSTER AUTOSCALER ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.P2s4wyJuCi
+ cat
+ chmod +x /tmp/tmp.P2s4wyJuCi
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.P2s4wyJuCi openshiftdevel:/tmp/tmp.P2s4wyJuCi
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.P2s4wyJuCi"'
+ cd /home/origin
+ export GOPATH=/data
+ GOPATH=/data
+ cd /data/src/github.com/openshift/kubernetes-autoscaler
+ sudo imagebuilder -f images/cluster-autoscaler/Dockerfile -t docker.io/openshift/origin-cluster-autoscaler:v4.0 .
--> Image registry.svc.ci.openshift.org/openshift/release:golang-1.12 was not found, pulling ...
--> Pulled 1/2 layers, 65% complete
--> Pulled 2/2 layers, 100% complete
--> Extracting
--> FROM registry.svc.ci.openshift.org/openshift/release:golang-1.12 as builder
--> WORKDIR /go/src/k8s.io/autoscaler
--> COPY . .
--> RUN go build -o cluster-autoscaler/cluster-autoscaler ./cluster-autoscaler
--> Image registry.svc.ci.openshift.org/openshift/origin-v4.0:base was not found, pulling ...
--> Pulled 2/4 layers, 54% complete
--> Pulled 3/4 layers, 82% complete
--> Pulled 4/4 layers, 100% complete
--> Extracting
--> FROM registry.svc.ci.openshift.org/openshift/origin-v4.0:base as 1
--> COPY --from=builder /go/src/k8s.io/autoscaler/cluster-autoscaler/cluster-autoscaler /usr/bin/
--> CMD /usr/bin/cluster-autoscaler
--> LABEL summary="Cluster Autoscaler for OpenShift and Kubernetes"
--> Committing changes to docker.io/openshift/origin-cluster-autoscaler:v4.0 ...
--> Done
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: BUILD CLUSTER AUTOSCALER [00h 02m 22s] ##########
[workspace] $ /bin/bash /tmp/jenkins3961425016706187314.sh
########## STARTING STAGE: DEPLOY MACHINE API OPERATOR ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.VcEYgLp21m
+ cat
+ chmod +x /tmp/tmp.VcEYgLp21m
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.VcEYgLp21m openshiftdevel:/tmp/tmp.VcEYgLp21m
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.VcEYgLp21m"'
+ cd /home/origin
+ export GOPATH=/data
+ GOPATH=/data
+ cd /data/src/github.com/openshift/machine-api-operator
+ make build
docker run --rm -v "/data/src/github.com/openshift/machine-api-operator":/go/src/github.com/openshift/machine-api-operator:Z -w /go/src/github.com/openshift/machine-api-operator golang:1.12 ./hack/go-build.sh machine-api-operator
Unable to find image 'golang:1.12' locally
Trying to pull repository registry.access.redhat.com/golang ... 
Pulling repository registry.access.redhat.com/golang
Trying to pull repository docker.io/library/golang ... 
1.12: Pulling from docker.io/library/golang
4ae16bd47783: Pulling fs layer
bbab4ec87ac4: Pulling fs layer
2ea1f7804402: Pulling fs layer
96465440c208: Pulling fs layer
16a3d8aca6cd: Pulling fs layer
e0ec5610455a: Pulling fs layer
96d705baf026: Pulling fs layer
16a3d8aca6cd: Waiting
e0ec5610455a: Waiting
96d705baf026: Waiting
96465440c208: Waiting
2ea1f7804402: Verifying Checksum
2ea1f7804402: Download complete
bbab4ec87ac4: Verifying Checksum
bbab4ec87ac4: Download complete
16a3d8aca6cd: Verifying Checksum
16a3d8aca6cd: Download complete
4ae16bd47783: Verifying Checksum
4ae16bd47783: Download complete
96d705baf026: Verifying Checksum
96d705baf026: Download complete
96465440c208: Verifying Checksum
96465440c208: Download complete
e0ec5610455a: Verifying Checksum
e0ec5610455a: Download complete
4ae16bd47783: Pull complete
bbab4ec87ac4: Pull complete
2ea1f7804402: Pull complete
96465440c208: Pull complete
16a3d8aca6cd: Pull complete
e0ec5610455a: Pull complete
96d705baf026: Pull complete
Digest: sha256:584daefca21d7b8f0702e70bfe6a31c3a6def12ed4145a8d6fe78c205f124f1e
Status: Downloaded newer image for docker.io/golang:1.12
Using version from git...
Building github.com/openshift/machine-api-operator/cmd/machine-api-operator (v0.1.0-526-g6d1a51d5)
docker run --rm -v "/data/src/github.com/openshift/machine-api-operator":/go/src/github.com/openshift/machine-api-operator:Z -w /go/src/github.com/openshift/machine-api-operator golang:1.12 ./hack/go-build.sh nodelink-controller
Using version from git...
Building github.com/openshift/machine-api-operator/cmd/nodelink-controller (v0.1.0-526-g6d1a51d5)
docker run --rm -v "/data/src/github.com/openshift/machine-api-operator":/go/src/github.com/openshift/machine-api-operator:Z -w /go/src/github.com/openshift/machine-api-operator golang:1.12 ./hack/go-build.sh machine-healthcheck
Using version from git...
Building github.com/openshift/machine-api-operator/cmd/machine-healthcheck (v0.1.0-526-g6d1a51d5)
+ sudo imagebuilder -t docker.io/openshift/origin-machine-api-operator:v4.0.0 .
--> FROM registry.svc.ci.openshift.org/openshift/release:golang-1.12 as builder
--> WORKDIR /go/src/github.com/openshift/machine-api-operator
--> COPY . .
--> RUN NO_DOCKER=1 make build
./hack/go-build.sh machine-api-operator
Using version from git...
Building github.com/openshift/machine-api-operator/cmd/machine-api-operator (v0.1.0-526-g6d1a51d5)
./hack/go-build.sh nodelink-controller
Using version from git...
Building github.com/openshift/machine-api-operator/cmd/nodelink-controller (v0.1.0-526-g6d1a51d5)
./hack/go-build.sh machine-healthcheck
Using version from git...
Building github.com/openshift/machine-api-operator/cmd/machine-healthcheck (v0.1.0-526-g6d1a51d5)
--> FROM registry.svc.ci.openshift.org/openshift/origin-v4.0:base as 1
--> COPY --from=builder /go/src/github.com/openshift/machine-api-operator/install manifests
--> COPY --from=builder /go/src/github.com/openshift/machine-api-operator/bin/machine-api-operator .
--> COPY --from=builder /go/src/github.com/openshift/machine-api-operator/bin/nodelink-controller .
--> COPY --from=builder /go/src/github.com/openshift/machine-api-operator/bin/machine-healthcheck .
--> LABEL io.openshift.release.operator true
--> Committing changes to docker.io/openshift/origin-machine-api-operator:v4.0.0 ...
--> Done
+ sudo make deploy-kubemark
kustomize build config | kubectl apply -f -
namespace "kubemark-actuator" created
serviceaccount "kubemark" created
clusterrole.rbac.authorization.k8s.io "kubemark-actuator-role" created
clusterrolebinding.rbac.authorization.k8s.io "kubemark-actuator-rolebinding" created
configmap "deleteunreadynodes" created
deployment.apps "machineapi-kubemark-controllers" created
kustomize build | kubectl apply -f -
namespace "openshift-machine-api" created
customresourcedefinition.apiextensions.k8s.io "clusteroperators.config.openshift.io" created
customresourcedefinition.apiextensions.k8s.io "featuregates.config.openshift.io" created
customresourcedefinition.apiextensions.k8s.io "machinedisruptionbudgets.healthchecking.openshift.io" created
customresourcedefinition.apiextensions.k8s.io "machinehealthchecks.healthchecking.openshift.io" created
customresourcedefinition.apiextensions.k8s.io "machines.machine.openshift.io" created
customresourcedefinition.apiextensions.k8s.io "machinesets.machine.openshift.io" created
customresourcedefinition.apiextensions.k8s.io "prometheusrules.monitoring.coreos.com" created
customresourcedefinition.apiextensions.k8s.io "servicemonitors.monitoring.coreos.com" created
serviceaccount "machine-api-controllers" created
serviceaccount "machine-api-operator" created
role.rbac.authorization.k8s.io "machine-api-controllers" created
role.rbac.authorization.k8s.io "machine-api-operator" created
role.rbac.authorization.k8s.io "prometheus-k8s-machine-api-operator" created
clusterrole.rbac.authorization.k8s.io "machine-api-controllers" created
clusterrole.rbac.authorization.k8s.io "machine-api-operator" created
rolebinding.rbac.authorization.k8s.io "machine-api-controllers" created
rolebinding.rbac.authorization.k8s.io "machine-api-operator" created
rolebinding.rbac.authorization.k8s.io "prometheus-k8s-machine-api-operator" created
clusterrolebinding.rbac.authorization.k8s.io "machine-api-controllers" created
clusterrolebinding.rbac.authorization.k8s.io "machine-api-operator" created
configmap "machine-api-operator-images" created
service "machine-api-operator" created
deployment.apps "machine-api-operator" created
clusteroperator.config.openshift.io "machine-api" created
kubectl apply -f config/kubemark-config-infra.yaml
customresourcedefinition.apiextensions.k8s.io "infrastructures.config.openshift.io" created
infrastructure.config.openshift.io "cluster" created
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPLOY MACHINE API OPERATOR [00h 03m 36s] ##########
[workspace] $ /bin/bash /tmp/jenkins6161104429701592219.sh
########## STARTING STAGE: DEPLOY CLUSTER AUTOSCALER OPERATOR ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.MX5xcQ1Enl
+ cat
+ chmod +x /tmp/tmp.MX5xcQ1Enl
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.MX5xcQ1Enl openshiftdevel:/tmp/tmp.MX5xcQ1Enl
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.MX5xcQ1Enl"'
+ cd /home/origin
+ export GOPATH=/data
+ GOPATH=/data
+ cd /data/src/github.com/openshift/cluster-autoscaler-operator/
+ sudo imagebuilder -t quay.io/openshift/origin-cluster-autoscaler-operator:v4.0 .
--> FROM registry.svc.ci.openshift.org/openshift/release:golang-1.12 as builder
--> WORKDIR /go/src/github.com/openshift/cluster-autoscaler-operator
--> COPY . .
--> ENV NO_DOCKER=1
--> ENV BUILD_DEST=/go/bin/cluster-autoscaler-operator
--> RUN unset VERSION && make build
go build  -ldflags "-X github.com/openshift/cluster-autoscaler-operator/pkg/version.Raw=v0.0.0-213-g045aea4" -o "/go/bin/cluster-autoscaler-operator" "github.com/openshift/cluster-autoscaler-operator/cmd/manager"
--> FROM registry.svc.ci.openshift.org/openshift/origin-v4.0:base as 1
--> COPY --from=builder /go/bin/cluster-autoscaler-operator /usr/bin/
--> COPY --from=builder /go/src/github.com/openshift/cluster-autoscaler-operator/install /manifests
--> CMD ["/usr/bin/cluster-autoscaler-operator"]
--> LABEL io.openshift.release.operator true
--> Committing changes to quay.io/openshift/origin-cluster-autoscaler-operator:v4.0 ...
--> Done
+ kustomize build
+ sudo kubectl apply -f -
customresourcedefinition.apiextensions.k8s.io "clusterautoscalers.autoscaling.openshift.io" created
customresourcedefinition.apiextensions.k8s.io "machineautoscalers.autoscaling.openshift.io" created
serviceaccount "cluster-autoscaler" created
serviceaccount "cluster-autoscaler-operator" created
role.rbac.authorization.k8s.io "cluster-autoscaler" created
role.rbac.authorization.k8s.io "prometheus-k8s-cluster-autoscaler-operator" created
role.rbac.authorization.k8s.io "cluster-autoscaler-operator" created
clusterrole.rbac.authorization.k8s.io "cluster-autoscaler" created
clusterrole.rbac.authorization.k8s.io "cluster-autoscaler-operator" created
rolebinding.rbac.authorization.k8s.io "cluster-autoscaler" created
rolebinding.rbac.authorization.k8s.io "prometheus-k8s-cluster-autoscaler-operator" created
rolebinding.rbac.authorization.k8s.io "cluster-autoscaler-operator" created
clusterrolebinding.rbac.authorization.k8s.io "cluster-autoscaler" created
clusterrolebinding.rbac.authorization.k8s.io "cluster-autoscaler-operator" created
configmap "cluster-autoscaler-operator-ca" created
secret "cluster-autoscaler-operator-cert" created
service "cluster-autoscaler-operator" created
deployment.apps "cluster-autoscaler-operator" created
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPLOY CLUSTER AUTOSCALER OPERATOR [00h 00m 46s] ##########
[workspace] $ /bin/bash /tmp/jenkins6141386697928010368.sh
########## STARTING STAGE: DEPLOY CLUSTER RESOURCES ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.noKQS6YIDK
+ cat
+ chmod +x /tmp/tmp.noKQS6YIDK
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.noKQS6YIDK openshiftdevel:/tmp/tmp.noKQS6YIDK
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.noKQS6YIDK"'
+ cd /home/origin
+ export GOPATH=/data
+ GOPATH=/data
+ cd /data/src/github.com/openshift/cluster-api-provider-kubemark
+ sudo kubectl apply -f examples/machine-set.yaml
machineset.machine.openshift.io "kubemark-actuator-testing-machineset" created
+ sudo kubectl apply -f examples/static-machine.yaml
machine.machine.openshift.io "minikube-static-machine" created
+ sudo kubectl apply -f examples/worker-machinesets.yaml
machineset.machine.openshift.io "kubemark-actuator-testing-machineset-red" created
machineset.machine.openshift.io "kubemark-actuator-testing-machineset-green" created
machineset.machine.openshift.io "kubemark-actuator-testing-machineset-blue" created
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPLOY CLUSTER RESOURCES [00h 00m 02s] ##########
[workspace] $ /bin/bash /tmp/jenkins3850359014073959748.sh
########## STARTING STAGE: INSTALL GO 1.10.1 ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.xutgEMoO0Z
+ cat
+ chmod +x /tmp/tmp.xutgEMoO0Z
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.xutgEMoO0Z openshiftdevel:/tmp/tmp.xutgEMoO0Z
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.xutgEMoO0Z"'
+ cd /home/origin
+ mkdir -p /home/origin/bin
+ curl -sL -o /home/origin/bin/gimme https://raw.githubusercontent.com/travis-ci/gimme/master/gimme
+ chmod +x /home/origin/bin/gimme
+ gimme 1.10.1

unset GOOS;
unset GOARCH;
export GOROOT='/home/origin/.gimme/versions/go1.10.1.linux.amd64';
export PATH="/home/origin/.gimme/versions/go1.10.1.linux.amd64/bin:${PATH}";
go version >&2;

export GIMME_ENV="/home/origin/.gimme/envs/go1.10.1.env"
+ source /home/origin/.gimme/envs/go1.10.1.env
++ unset GOOS
++ unset GOARCH
++ export GOROOT=/home/origin/.gimme/versions/go1.10.1.linux.amd64
++ GOROOT=/home/origin/.gimme/versions/go1.10.1.linux.amd64
++ export PATH=/home/origin/.gimme/versions/go1.10.1.linux.amd64/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/home/origin/.local/bin:/home/origin/bin
++ PATH=/home/origin/.gimme/versions/go1.10.1.linux.amd64/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/home/origin/.local/bin:/home/origin/bin
++ go version
go version go1.10.1 linux/amd64
+ sudo cp /home/origin/.gimme/versions/go1.10.1.linux.amd64/bin/go /bin/go
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: INSTALL GO 1.10.1 [00h 00m 07s] ##########
[workspace] $ /bin/bash /tmp/jenkins1990061321328044952.sh
########## STARTING STAGE: RUN E2E TESTS ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.FHlRwRMQYA
+ cat
+ chmod +x /tmp/tmp.FHlRwRMQYA
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.FHlRwRMQYA openshiftdevel:/tmp/tmp.FHlRwRMQYA
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.FHlRwRMQYA"'
+ cd /home/origin
+ set +x
go version go1.10.1 linux/amd64
./hack/e2e.sh test-e2e
make[1]: Entering directory `/tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg'
# Run operator tests first to preserve logs for troubleshooting test
# failures and flakes.
# Feature:Operator tests remove deployments. Thus loosing all the logs
# previously acquired.
hack/ci-integration.sh  -ginkgo.v -ginkgo.noColor=true -ginkgo.focus "Feature:Operators" -ginkgo.failFast
=== RUN   TestE2E
Running Suite: Machine Suite
============================
Random Seed: 1567782566
Will run 7 of 16 specs

[Feature:Operators] Cluster autoscaler operator deployment should 
  be available
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/operators/cluster-autoscaler-operator.go:79
I0906 15:09:26.652799    5009 framework.go:406] >>> kubeConfig: /root/.kube/config
I0906 15:09:26.689216    5009 deloyment.go:58] Deployment "cluster-autoscaler-operator" is available. Status: (replicas: 1, updated: 1, ready: 1, available: 1, unavailable: 0)
•SSSSSSS
------------------------------
[Feature:Operators] Cluster autoscaler cluster operator status should 
  be available
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/operators/cluster-autoscaler-operator.go:90
I0906 15:09:26.689334    5009 framework.go:406] >>> kubeConfig: /root/.kube/config
•
------------------------------
[Feature:Operators] Cluster autoscaler operator should 
  reject invalid ClusterAutoscaler resources early via webhook
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/operators/cluster-autoscaler-operator.go:33
I0906 15:09:26.706063    5009 framework.go:406] >>> kubeConfig: /root/.kube/config
•
------------------------------
[Feature:Operators] Cluster autoscaler operator should 
  reject invalid MachineAutoscaler resources early via webhook
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/operators/cluster-autoscaler-operator.go:49
I0906 15:09:26.733424    5009 framework.go:406] >>> kubeConfig: /root/.kube/config
•
------------------------------
[Feature:Operators] Machine API cluster operator status should 
  be available
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/operators/machine-api-operator.go:53
I0906 15:09:26.758277    5009 framework.go:406] >>> kubeConfig: /root/.kube/config
•
------------------------------
[Feature:Operators] Machine API operator deployment should 
  be available
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/operators/machine-api-operator.go:18
I0906 15:09:26.770987    5009 framework.go:406] >>> kubeConfig: /root/.kube/config
I0906 15:09:26.786446    5009 deloyment.go:58] Deployment "machine-api-operator" is available. Status: (replicas: 1, updated: 1, ready: 1, available: 1, unavailable: 0)
•
------------------------------
[Feature:Operators] Machine API operator deployment should 
  reconcile controllers deployment
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/operators/machine-api-operator.go:25
I0906 15:09:26.786512    5009 framework.go:406] >>> kubeConfig: /root/.kube/config
STEP: checking deployment "machine-api-controllers" is available
I0906 15:09:26.799549    5009 deloyment.go:58] Deployment "machine-api-controllers" is available. Status: (replicas: 1, updated: 1, ready: 1, available: 1, unavailable: 0)
STEP: deleting deployment "machine-api-controllers"
STEP: checking deployment "machine-api-controllers" is available again
E0906 15:09:26.807416    5009 deloyment.go:25] Error querying api for Deployment object "machine-api-controllers": deployments.apps "machine-api-controllers" not found, retrying...
E0906 15:09:27.810060    5009 deloyment.go:55] Deployment "machine-api-controllers" is not available. Status: (replicas: 1, updated: 1, ready: 0, available: 0, unavailable: 1)
E0906 15:09:28.812152    5009 deloyment.go:55] Deployment "machine-api-controllers" is not available. Status: (replicas: 1, updated: 1, ready: 0, available: 0, unavailable: 1)
I0906 15:09:29.812670    5009 deloyment.go:58] Deployment "machine-api-controllers" is available. Status: (replicas: 1, updated: 1, ready: 1, available: 1, unavailable: 0)
•SS
Ran 7 of 16 Specs in 3.160 seconds
SUCCESS! -- 7 Passed | 0 Failed | 0 Pending | 9 Skipped
--- PASS: TestE2E (3.16s)
PASS
ok  	github.com/openshift/cluster-api-actuator-pkg/pkg/e2e	3.213s
hack/ci-integration.sh  -ginkgo.v -ginkgo.noColor=true -ginkgo.skip "Feature:Operators|TechPreview" -ginkgo.failFast -ginkgo.seed=1
=== RUN   TestE2E
Running Suite: Machine Suite
============================
Random Seed: 1
Will run 7 of 16 specs

SSSSSSSS
------------------------------
[Feature:Machines] Autoscaler should 
  scale up and down
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/autoscaler/autoscaler.go:234
I0906 15:09:33.015012    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
I0906 15:09:33.021567    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
I0906 15:09:33.052537    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
STEP: Getting existing machinesets
STEP: Getting existing machines
STEP: Getting existing nodes
I0906 15:09:33.071035    5527 autoscaler.go:286] Have 4 existing machinesets
I0906 15:09:33.071058    5527 autoscaler.go:287] Have 5 existing machines
I0906 15:09:33.071069    5527 autoscaler.go:288] Have 5 existing nodes
STEP: Creating 3 transient machinesets
STEP: [15m0s remaining] Waiting for nodes to be Ready in 3 transient machinesets
E0906 15:09:33.099071    5527 utils.go:157] Machine "e2e-5508c-w-0-mlv9z" has no NodeRef
STEP: [14m57s remaining] Waiting for nodes to be Ready in 3 transient machinesets
I0906 15:09:36.121878    5527 utils.go:165] Machine "e2e-5508c-w-0-mlv9z" is backing node "00db6f98-a3ac-4b9d-9b06-baeefad63df4"
I0906 15:09:36.121907    5527 utils.go:149] MachineSet "e2e-5508c-w-0" have 1 nodes
E0906 15:09:36.131240    5527 utils.go:157] Machine "e2e-5508c-w-1-xxftr" has no NodeRef
STEP: [14m54s remaining] Waiting for nodes to be Ready in 3 transient machinesets
I0906 15:09:39.137754    5527 utils.go:165] Machine "e2e-5508c-w-0-mlv9z" is backing node "00db6f98-a3ac-4b9d-9b06-baeefad63df4"
I0906 15:09:39.137783    5527 utils.go:149] MachineSet "e2e-5508c-w-0" have 1 nodes
I0906 15:09:39.143301    5527 utils.go:165] Machine "e2e-5508c-w-1-xxftr" is backing node "7ab053ab-5975-4dd7-a60f-6db3990be26f"
I0906 15:09:39.143323    5527 utils.go:149] MachineSet "e2e-5508c-w-1" have 1 nodes
I0906 15:09:39.148525    5527 utils.go:165] Machine "e2e-5508c-w-2-wj4jh" is backing node "66cf1356-1533-4fee-8ea0-24d40b6aef5f"
I0906 15:09:39.148547    5527 utils.go:149] MachineSet "e2e-5508c-w-2" have 1 nodes
I0906 15:09:39.148555    5527 utils.go:177] Node "00db6f98-a3ac-4b9d-9b06-baeefad63df4" is ready. Conditions are: [{OutOfDisk False 2019-09-06 15:09:37 +0000 UTC 2019-09-06 15:09:35 +0000 UTC KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2019-09-06 15:09:37 +0000 UTC 2019-09-06 15:09:35 +0000 UTC KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2019-09-06 15:09:37 +0000 UTC 2019-09-06 15:09:35 +0000 UTC KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2019-09-06 15:09:37 +0000 UTC 2019-09-06 15:09:35 +0000 UTC KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2019-09-06 15:09:37 +0000 UTC 2019-09-06 15:09:35 +0000 UTC KubeletReady kubelet is posting ready status}]
I0906 15:09:39.148635    5527 utils.go:177] Node "7ab053ab-5975-4dd7-a60f-6db3990be26f" is ready. Conditions are: [{OutOfDisk False 2019-09-06 15:09:37 +0000 UTC 2019-09-06 15:09:35 +0000 UTC KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2019-09-06 15:09:37 +0000 UTC 2019-09-06 15:09:35 +0000 UTC KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2019-09-06 15:09:37 +0000 UTC 2019-09-06 15:09:35 +0000 UTC KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2019-09-06 15:09:37 +0000 UTC 2019-09-06 15:09:35 +0000 UTC KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2019-09-06 15:09:37 +0000 UTC 2019-09-06 15:09:35 +0000 UTC KubeletReady kubelet is posting ready status}]
I0906 15:09:39.148659    5527 utils.go:177] Node "66cf1356-1533-4fee-8ea0-24d40b6aef5f" is ready. Conditions are: [{OutOfDisk False 2019-09-06 15:09:38 +0000 UTC 2019-09-06 15:09:36 +0000 UTC KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2019-09-06 15:09:38 +0000 UTC 2019-09-06 15:09:36 +0000 UTC KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2019-09-06 15:09:38 +0000 UTC 2019-09-06 15:09:36 +0000 UTC KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2019-09-06 15:09:38 +0000 UTC 2019-09-06 15:09:36 +0000 UTC KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2019-09-06 15:09:38 +0000 UTC 2019-09-06 15:09:36 +0000 UTC KubeletReady kubelet is posting ready status}]
STEP: Getting nodes
STEP: Creating 3 machineautoscalers
I0906 15:09:39.151562    5527 autoscaler.go:340] Create MachineAutoscaler backed by MachineSet kube-system/e2e-5508c-w-0 - min:1, max:2
I0906 15:09:39.158479    5527 autoscaler.go:340] Create MachineAutoscaler backed by MachineSet kube-system/e2e-5508c-w-1 - min:1, max:2
I0906 15:09:39.162577    5527 autoscaler.go:340] Create MachineAutoscaler backed by MachineSet kube-system/e2e-5508c-w-2 - min:1, max:2
STEP: Creating ClusterAutoscaler configured with maxNodesTotal:10
STEP: Deriving Memory capacity from machine "kubemark-actuator-testing-machineset"
I0906 15:09:39.276486    5527 autoscaler.go:377] Memory capacity of worker node "359b0676-397f-402c-b209-ed17aa0a216c" is 3840Mi
STEP: Creating scale-out workload: jobs: 11, memory: 2818572300
I0906 15:09:39.304637    5527 autoscaler.go:399] [15m0s remaining] Expecting 2 "ScaledUpGroup" events; observed 0
I0906 15:09:40.379733    5527 autoscaler.go:361] cluster-autoscaler: cluster-autoscaler-default-598c649f66-tgmls became leader
I0906 15:09:42.304866    5527 autoscaler.go:399] [14m57s remaining] Expecting 2 "ScaledUpGroup" events; observed 0
I0906 15:09:45.305082    5527 autoscaler.go:399] [14m54s remaining] Expecting 2 "ScaledUpGroup" events; observed 0
I0906 15:09:48.305196    5527 autoscaler.go:399] [14m51s remaining] Expecting 2 "ScaledUpGroup" events; observed 0
I0906 15:09:50.515652    5527 autoscaler.go:361] cluster-autoscaler-status: Max total nodes in cluster reached: 10
I0906 15:09:50.518206    5527 autoscaler.go:361] cluster-autoscaler-status: Scale-up: setting group kube-system/e2e-5508c-w-1 size to 2
I0906 15:09:50.523617    5527 autoscaler.go:361] cluster-autoscaler-status: Scale-up: group kube-system/e2e-5508c-w-1 size set to 2
I0906 15:09:50.526317    5527 autoscaler.go:361] e2e-autoscaler-workload-k7d25: pod triggered scale-up: [{kube-system/e2e-5508c-w-1 1->2 (max: 2)}]
I0906 15:09:50.532696    5527 autoscaler.go:361] e2e-autoscaler-workload-7745h: pod triggered scale-up: [{kube-system/e2e-5508c-w-1 1->2 (max: 2)}]
I0906 15:09:50.538502    5527 autoscaler.go:361] e2e-autoscaler-workload-x9srw: pod triggered scale-up: [{kube-system/e2e-5508c-w-1 1->2 (max: 2)}]
I0906 15:09:50.544645    5527 autoscaler.go:361] e2e-autoscaler-workload-n5xxj: pod triggered scale-up: [{kube-system/e2e-5508c-w-1 1->2 (max: 2)}]
I0906 15:09:50.552081    5527 autoscaler.go:361] e2e-autoscaler-workload-2h24c: pod triggered scale-up: [{kube-system/e2e-5508c-w-1 1->2 (max: 2)}]
I0906 15:09:50.563329    5527 autoscaler.go:361] e2e-autoscaler-workload-hl5bk: pod triggered scale-up: [{kube-system/e2e-5508c-w-1 1->2 (max: 2)}]
I0906 15:09:50.570678    5527 autoscaler.go:361] e2e-autoscaler-workload-2lbwq: pod triggered scale-up: [{kube-system/e2e-5508c-w-1 1->2 (max: 2)}]
I0906 15:09:50.715740    5527 autoscaler.go:361] e2e-autoscaler-workload-cks94: pod triggered scale-up: [{kube-system/e2e-5508c-w-1 1->2 (max: 2)}]
I0906 15:09:51.305407    5527 autoscaler.go:399] [14m48s remaining] Expecting 2 "ScaledUpGroup" events; observed 1
I0906 15:09:54.305658    5527 autoscaler.go:399] [14m45s remaining] Expecting 2 "ScaledUpGroup" events; observed 1
I0906 15:09:57.306516    5527 autoscaler.go:399] [14m42s remaining] Expecting 2 "ScaledUpGroup" events; observed 1
I0906 15:10:00.306733    5527 autoscaler.go:399] [14m39s remaining] Expecting 2 "ScaledUpGroup" events; observed 1
I0906 15:10:00.548986    5527 autoscaler.go:361] cluster-autoscaler-status: Scale-up: setting group kube-system/e2e-5508c-w-0 size to 2
I0906 15:10:00.553746    5527 autoscaler.go:361] e2e-autoscaler-workload-cks94: pod triggered scale-up: [{kube-system/e2e-5508c-w-0 1->2 (max: 2)}]
I0906 15:10:00.560145    5527 autoscaler.go:361] cluster-autoscaler-status: Scale-up: group kube-system/e2e-5508c-w-0 size set to 2
I0906 15:10:00.562190    5527 autoscaler.go:361] e2e-autoscaler-workload-7745h: pod triggered scale-up: [{kube-system/e2e-5508c-w-0 1->2 (max: 2)}]
I0906 15:10:00.564159    5527 autoscaler.go:361] e2e-autoscaler-workload-n5xxj: pod triggered scale-up: [{kube-system/e2e-5508c-w-0 1->2 (max: 2)}]
I0906 15:10:00.570030    5527 autoscaler.go:361] e2e-autoscaler-workload-k7d25: pod triggered scale-up: [{kube-system/e2e-5508c-w-0 1->2 (max: 2)}]
I0906 15:10:00.578727    5527 autoscaler.go:361] e2e-autoscaler-workload-x9srw: pod triggered scale-up: [{kube-system/e2e-5508c-w-0 1->2 (max: 2)}]
I0906 15:10:00.587666    5527 autoscaler.go:361] e2e-autoscaler-workload-hl5bk: pod triggered scale-up: [{kube-system/e2e-5508c-w-0 1->2 (max: 2)}]
I0906 15:10:00.591015    5527 autoscaler.go:361] e2e-autoscaler-workload-2lbwq: pod triggered scale-up: [{kube-system/e2e-5508c-w-0 1->2 (max: 2)}]
I0906 15:10:03.306991    5527 autoscaler.go:399] [14m36s remaining] Expecting 2 "ScaledUpGroup" events; observed 2
I0906 15:10:03.307900    5527 autoscaler.go:414] [1m0s remaining] Waiting for cluster-autoscaler to generate a "MaxNodesTotalReached" event; observed 1
I0906 15:10:03.307930    5527 autoscaler.go:422] [1m0s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:06.308135    5527 autoscaler.go:422] [57s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:09.308419    5527 autoscaler.go:422] [54s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:12.308836    5527 autoscaler.go:422] [51s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:15.309087    5527 autoscaler.go:422] [48s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:18.309347    5527 autoscaler.go:422] [45s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:21.309578    5527 autoscaler.go:422] [42s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:24.309822    5527 autoscaler.go:422] [39s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:27.310774    5527 autoscaler.go:422] [36s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:30.310996    5527 autoscaler.go:422] [33s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:33.311243    5527 autoscaler.go:422] [30s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:36.311532    5527 autoscaler.go:422] [27s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:39.311794    5527 autoscaler.go:422] [24s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:42.312009    5527 autoscaler.go:422] [21s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:45.312269    5527 autoscaler.go:422] [18s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:48.312537    5527 autoscaler.go:422] [15s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:51.312789    5527 autoscaler.go:422] [12s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:54.313064    5527 autoscaler.go:422] [9s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:10:57.313292    5527 autoscaler.go:422] [6s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
I0906 15:11:00.313459    5527 autoscaler.go:422] [3s remaining] At max cluster size and expecting no more "ScaledUpGroup" events; currently have 2, max=2
STEP: Deleting workload
I0906 15:11:03.308151    5527 autoscaler.go:249] [cleanup] "e2e-autoscaler-workload" (*v1.Job)
I0906 15:11:03.313341    5527 autoscaler.go:434] [15m0s remaining] Expecting 2 "ScaleDownEmpty" events; observed 2
I0906 15:11:03.348034    5527 autoscaler.go:445] still have workload POD: "e2e-autoscaler-workload-2h24c"
I0906 15:11:03.348073    5527 autoscaler.go:249] [cleanup] "default" (*v1.ClusterAutoscaler)
I0906 15:11:03.452490    5527 autoscaler.go:465] Waiting for cluster-autoscaler POD "cluster-autoscaler-default-598c649f66-tgmls" to disappear
STEP: Scaling transient machinesets to zero
I0906 15:11:03.452550    5527 autoscaler.go:474] Scaling transient machineset "e2e-5508c-w-0" to zero
I0906 15:11:03.458112    5527 autoscaler.go:474] Scaling transient machineset "e2e-5508c-w-1" to zero
I0906 15:11:03.466094    5527 autoscaler.go:474] Scaling transient machineset "e2e-5508c-w-2" to zero
STEP: Waiting for scaled up nodes to be deleted
I0906 15:11:03.522000    5527 autoscaler.go:491] [15m0s remaining] Waiting for cluster to reach original node count of 5; currently have 10
I0906 15:11:06.526461    5527 autoscaler.go:491] [14m57s remaining] Waiting for cluster to reach original node count of 5; currently have 8
I0906 15:11:09.530138    5527 autoscaler.go:491] [14m54s remaining] Waiting for cluster to reach original node count of 5; currently have 5
STEP: Waiting for scaled up machines to be deleted
I0906 15:11:09.533584    5527 autoscaler.go:501] [15m0s remaining] Waiting for cluster to reach original machine count of 5; currently have 5
I0906 15:11:09.533616    5527 autoscaler.go:249] [cleanup] "autoscale-e2e-5508c-w-0mtzfn" (*v1beta1.MachineAutoscaler)
I0906 15:11:09.536918    5527 autoscaler.go:249] [cleanup] "autoscale-e2e-5508c-w-1zmp8d" (*v1beta1.MachineAutoscaler)
I0906 15:11:09.540193    5527 autoscaler.go:249] [cleanup] "autoscale-e2e-5508c-w-2z6hhv" (*v1beta1.MachineAutoscaler)
I0906 15:11:09.545457    5527 autoscaler.go:249] [cleanup] "e2e-5508c-w-0" (*v1beta1.MachineSet)
I0906 15:11:09.549133    5527 autoscaler.go:249] [cleanup] "e2e-5508c-w-1" (*v1beta1.MachineSet)
I0906 15:11:09.554079    5527 autoscaler.go:249] [cleanup] "e2e-5508c-w-2" (*v1beta1.MachineSet)

• [SLOW TEST:96.546 seconds]
[Feature:Machines] Autoscaler should
/tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/autoscaler/autoscaler.go:233
  scale up and down
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/autoscaler/autoscaler.go:234
------------------------------
S
------------------------------
[Feature:Machines] Managed cluster should 
  have machines linked with nodes
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/infra/infra.go:136
I0906 15:11:09.561108    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
I0906 15:11:09.579106    5527 utils.go:47] [remaining 3m0s] Expecting the same number of machines and nodes, have 5 nodes and 5 machines
I0906 15:11:09.579139    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-6pt7l" is linked to node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab"
I0906 15:11:09.579152    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-blue-hpgct" is linked to node "8d76d38d-5446-4aef-802c-ad0fcfdb4546"
I0906 15:11:09.579160    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-green-nr9lx" is linked to node "359b0676-397f-402c-b209-ed17aa0a216c"
I0906 15:11:09.579169    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-red-s4l9g" is linked to node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b"
I0906 15:11:09.579185    5527 utils.go:70] [remaining 3m0s] Machine "minikube-static-machine" is linked to node "minikube"
•
------------------------------
[Feature:Machines] Managed cluster should 
  have ability to additively reconcile taints from machine to nodes
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/infra/infra.go:145
I0906 15:11:09.579237    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
STEP: getting machine "kubemark-actuator-testing-machineset-6pt7l"
I0906 15:11:09.598496    5527 utils.go:165] Machine "kubemark-actuator-testing-machineset-6pt7l" is backing node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab"
STEP: getting the backed node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab"
STEP: updating node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab" with taint: {not-from-machine true NoSchedule <nil>}
STEP: updating machine "kubemark-actuator-testing-machineset-6pt7l" with taint: {from-machine-8e92327e-d0b8-11e9-978c-0a445740e986 true NoSchedule <nil>}
I0906 15:11:09.607997    5527 infra.go:184] Getting node from machine again for verification of taints
I0906 15:11:09.611944    5527 utils.go:165] Machine "kubemark-actuator-testing-machineset-6pt7l" is backing node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab"
I0906 15:11:09.611980    5527 infra.go:194] Expected : map[not-from-machine:{} from-machine-8e92327e-d0b8-11e9-978c-0a445740e986:{}], observed map[kubemark:{} not-from-machine:{} from-machine-8e92327e-d0b8-11e9-978c-0a445740e986:{}] , difference map[], 
STEP: Getting the latest version of the original machine
STEP: Setting back the original machine taints
STEP: Getting the latest version of the node
I0906 15:11:09.625610    5527 utils.go:165] Machine "kubemark-actuator-testing-machineset-6pt7l" is backing node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab"
STEP: Setting back the original node taints
•
------------------------------
[Feature:Machines] Managed cluster should 
  recover from deleted worker machines
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/infra/infra.go:220
I0906 15:11:09.629879    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
STEP: checking initial cluster state
I0906 15:11:09.657202    5527 utils.go:87] Cluster size is 5 nodes
I0906 15:11:09.657230    5527 utils.go:239] [remaining 15m0s] Cluster size expected to be 5 nodes
I0906 15:11:09.661265    5527 utils.go:99] MachineSet "e2e-5508c-w-0" replicas 0. Ready: 0, available 0
I0906 15:11:09.661290    5527 utils.go:99] MachineSet "e2e-5508c-w-1" replicas 0. Ready: 0, available 0
I0906 15:11:09.661299    5527 utils.go:99] MachineSet "e2e-5508c-w-2" replicas 0. Ready: 0, available 0
I0906 15:11:09.661307    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset" replicas 1. Ready: 1, available 1
I0906 15:11:09.661316    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-blue" replicas 1. Ready: 1, available 1
I0906 15:11:09.661325    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-green" replicas 1. Ready: 1, available 1
I0906 15:11:09.661334    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-red" replicas 1. Ready: 1, available 1
I0906 15:11:09.664335    5527 utils.go:231] Node "359b0676-397f-402c-b209-ed17aa0a216c". Ready: true. Unschedulable: false
I0906 15:11:09.664360    5527 utils.go:231] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab". Ready: true. Unschedulable: false
I0906 15:11:09.664371    5527 utils.go:231] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546". Ready: true. Unschedulable: false
I0906 15:11:09.664376    5527 utils.go:231] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b". Ready: true. Unschedulable: false
I0906 15:11:09.664382    5527 utils.go:231] Node "minikube". Ready: true. Unschedulable: false
I0906 15:11:09.667091    5527 utils.go:87] Cluster size is 5 nodes
I0906 15:11:09.667108    5527 utils.go:257] waiting for all nodes to be ready
I0906 15:11:09.670260    5527 utils.go:262] waiting for all nodes to be schedulable
I0906 15:11:09.674411    5527 utils.go:290] [remaining 1m0s] Node "359b0676-397f-402c-b209-ed17aa0a216c" is schedulable
I0906 15:11:09.674440    5527 utils.go:290] [remaining 1m0s] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab" is schedulable
I0906 15:11:09.674450    5527 utils.go:290] [remaining 1m0s] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546" is schedulable
I0906 15:11:09.674457    5527 utils.go:290] [remaining 1m0s] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b" is schedulable
I0906 15:11:09.674463    5527 utils.go:290] [remaining 1m0s] Node "minikube" is schedulable
I0906 15:11:09.674471    5527 utils.go:267] waiting for each node to be backed by a machine
I0906 15:11:09.684919    5527 utils.go:47] [remaining 3m0s] Expecting the same number of machines and nodes, have 5 nodes and 5 machines
I0906 15:11:09.684955    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-6pt7l" is linked to node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab"
I0906 15:11:09.684970    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-blue-hpgct" is linked to node "8d76d38d-5446-4aef-802c-ad0fcfdb4546"
I0906 15:11:09.684984    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-green-nr9lx" is linked to node "359b0676-397f-402c-b209-ed17aa0a216c"
I0906 15:11:09.684997    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-red-s4l9g" is linked to node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b"
I0906 15:11:09.685015    5527 utils.go:70] [remaining 3m0s] Machine "minikube-static-machine" is linked to node "minikube"
STEP: getting worker node
STEP: deleting machine object "kubemark-actuator-testing-machineset-green-nr9lx"
STEP: waiting for node object "359b0676-397f-402c-b209-ed17aa0a216c" to go away
I0906 15:11:09.699018    5527 infra.go:255] Node "359b0676-397f-402c-b209-ed17aa0a216c" still exists. Node conditions are: [{OutOfDisk False 2019-09-06 15:11:09 +0000 UTC 2019-09-06 15:08:33 +0000 UTC KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2019-09-06 15:11:09 +0000 UTC 2019-09-06 15:08:33 +0000 UTC KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2019-09-06 15:11:09 +0000 UTC 2019-09-06 15:08:33 +0000 UTC KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2019-09-06 15:11:09 +0000 UTC 2019-09-06 15:08:33 +0000 UTC KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2019-09-06 15:11:09 +0000 UTC 2019-09-06 15:08:33 +0000 UTC KubeletReady kubelet is posting ready status}]
STEP: waiting for new node object to come up
I0906 15:11:14.703992    5527 utils.go:239] [remaining 15m0s] Cluster size expected to be 5 nodes
I0906 15:11:14.707665    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset" replicas 1. Ready: 1, available 1
I0906 15:11:14.707687    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-blue" replicas 1. Ready: 1, available 1
I0906 15:11:14.707694    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-green" replicas 1. Ready: 1, available 1
I0906 15:11:14.707699    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-red" replicas 1. Ready: 1, available 1
I0906 15:11:14.710430    5527 utils.go:231] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab". Ready: true. Unschedulable: false
I0906 15:11:14.710449    5527 utils.go:231] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546". Ready: true. Unschedulable: false
I0906 15:11:14.710454    5527 utils.go:231] Node "b3408843-b44c-4857-ab9d-3b13ab158aea". Ready: true. Unschedulable: false
I0906 15:11:14.710459    5527 utils.go:231] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b". Ready: true. Unschedulable: false
I0906 15:11:14.710468    5527 utils.go:231] Node "minikube". Ready: true. Unschedulable: false
I0906 15:11:14.713015    5527 utils.go:87] Cluster size is 5 nodes
I0906 15:11:14.713041    5527 utils.go:257] waiting for all nodes to be ready
I0906 15:11:14.715847    5527 utils.go:262] waiting for all nodes to be schedulable
I0906 15:11:14.718805    5527 utils.go:290] [remaining 1m0s] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab" is schedulable
I0906 15:11:14.718828    5527 utils.go:290] [remaining 1m0s] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546" is schedulable
I0906 15:11:14.718835    5527 utils.go:290] [remaining 1m0s] Node "b3408843-b44c-4857-ab9d-3b13ab158aea" is schedulable
I0906 15:11:14.718842    5527 utils.go:290] [remaining 1m0s] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b" is schedulable
I0906 15:11:14.718862    5527 utils.go:290] [remaining 1m0s] Node "minikube" is schedulable
I0906 15:11:14.718868    5527 utils.go:267] waiting for each node to be backed by a machine
I0906 15:11:14.724556    5527 utils.go:47] [remaining 3m0s] Expecting the same number of machines and nodes, have 5 nodes and 5 machines
I0906 15:11:14.724583    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-6pt7l" is linked to node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab"
I0906 15:11:14.724594    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-blue-hpgct" is linked to node "8d76d38d-5446-4aef-802c-ad0fcfdb4546"
I0906 15:11:14.724602    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-green-scthk" is linked to node "b3408843-b44c-4857-ab9d-3b13ab158aea"
I0906 15:11:14.724613    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-red-s4l9g" is linked to node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b"
I0906 15:11:14.724634    5527 utils.go:70] [remaining 3m0s] Machine "minikube-static-machine" is linked to node "minikube"

• [SLOW TEST:5.095 seconds]
[Feature:Machines] Managed cluster should
/tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/infra/infra.go:126
  recover from deleted worker machines
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/infra/infra.go:220
------------------------------
[Feature:Machines] Managed cluster should 
  grow and decrease when scaling different machineSets simultaneously
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/infra/infra.go:267
I0906 15:11:14.724720    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
STEP: checking existing cluster size
I0906 15:11:14.740937    5527 utils.go:87] Cluster size is 5 nodes
STEP: getting worker machineSets
I0906 15:11:14.743851    5527 infra.go:297] Creating transient MachineSet "e2e-91a2d-w-0"
I0906 15:11:14.748839    5527 infra.go:297] Creating transient MachineSet "e2e-91a2d-w-1"
STEP: scaling "e2e-91a2d-w-0" from 0 to 2 replicas
I0906 15:11:14.752871    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
STEP: scaling "e2e-91a2d-w-1" from 0 to 2 replicas
I0906 15:11:14.772375    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
E0906 15:11:14.812124    5527 utils.go:157] Machine "e2e-91a2d-w-0-bcdrm" has no NodeRef
I0906 15:11:19.829884    5527 utils.go:165] Machine "e2e-91a2d-w-0-bcdrm" is backing node "41e2bf6d-a04c-4354-afea-7e711d38300e"
I0906 15:11:19.838522    5527 utils.go:165] Machine "e2e-91a2d-w-0-v2vd5" is backing node "a18bf460-d110-4af6-91a0-4af7a5c1fe76"
I0906 15:11:19.838545    5527 utils.go:149] MachineSet "e2e-91a2d-w-0" have 2 nodes
E0906 15:11:19.852028    5527 utils.go:157] Machine "e2e-91a2d-w-1-kxg8f" has no NodeRef
I0906 15:11:24.860019    5527 utils.go:165] Machine "e2e-91a2d-w-0-bcdrm" is backing node "41e2bf6d-a04c-4354-afea-7e711d38300e"
I0906 15:11:24.862527    5527 utils.go:165] Machine "e2e-91a2d-w-0-v2vd5" is backing node "a18bf460-d110-4af6-91a0-4af7a5c1fe76"
I0906 15:11:24.862548    5527 utils.go:149] MachineSet "e2e-91a2d-w-0" have 2 nodes
I0906 15:11:24.868337    5527 utils.go:165] Machine "e2e-91a2d-w-1-kxg8f" is backing node "86eff62d-6aee-4907-b3a3-b0af551e243b"
I0906 15:11:24.870121    5527 utils.go:165] Machine "e2e-91a2d-w-1-z5zn4" is backing node "f94e2b84-7660-436d-b933-ce06e9220145"
I0906 15:11:24.870145    5527 utils.go:149] MachineSet "e2e-91a2d-w-1" have 2 nodes
I0906 15:11:24.870156    5527 utils.go:177] Node "41e2bf6d-a04c-4354-afea-7e711d38300e" is ready. Conditions are: [{OutOfDisk False 2019-09-06 15:11:23 +0000 UTC 2019-09-06 15:11:17 +0000 UTC KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2019-09-06 15:11:23 +0000 UTC 2019-09-06 15:11:17 +0000 UTC KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2019-09-06 15:11:23 +0000 UTC 2019-09-06 15:11:17 +0000 UTC KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2019-09-06 15:11:23 +0000 UTC 2019-09-06 15:11:17 +0000 UTC KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2019-09-06 15:11:23 +0000 UTC 2019-09-06 15:11:17 +0000 UTC KubeletReady kubelet is posting ready status}]
I0906 15:11:24.870250    5527 utils.go:177] Node "a18bf460-d110-4af6-91a0-4af7a5c1fe76" is ready. Conditions are: [{OutOfDisk False 2019-09-06 15:11:24 +0000 UTC 2019-09-06 15:11:17 +0000 UTC KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2019-09-06 15:11:24 +0000 UTC 2019-09-06 15:11:17 +0000 UTC KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2019-09-06 15:11:24 +0000 UTC 2019-09-06 15:11:17 +0000 UTC KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2019-09-06 15:11:24 +0000 UTC 2019-09-06 15:11:17 +0000 UTC KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2019-09-06 15:11:24 +0000 UTC 2019-09-06 15:11:17 +0000 UTC KubeletReady kubelet is posting ready status}]
I0906 15:11:24.870288    5527 utils.go:177] Node "86eff62d-6aee-4907-b3a3-b0af551e243b" is ready. Conditions are: [{OutOfDisk False 2019-09-06 15:11:24 +0000 UTC 2019-09-06 15:11:19 +0000 UTC KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2019-09-06 15:11:24 +0000 UTC 2019-09-06 15:11:19 +0000 UTC KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2019-09-06 15:11:24 +0000 UTC 2019-09-06 15:11:19 +0000 UTC KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2019-09-06 15:11:24 +0000 UTC 2019-09-06 15:11:19 +0000 UTC KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2019-09-06 15:11:24 +0000 UTC 2019-09-06 15:11:19 +0000 UTC KubeletReady kubelet is posting ready status}]
I0906 15:11:24.870315    5527 utils.go:177] Node "f94e2b84-7660-436d-b933-ce06e9220145" is ready. Conditions are: [{OutOfDisk False 2019-09-06 15:11:22 +0000 UTC 2019-09-06 15:11:18 +0000 UTC KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2019-09-06 15:11:22 +0000 UTC 2019-09-06 15:11:18 +0000 UTC KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2019-09-06 15:11:22 +0000 UTC 2019-09-06 15:11:18 +0000 UTC KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2019-09-06 15:11:22 +0000 UTC 2019-09-06 15:11:18 +0000 UTC KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2019-09-06 15:11:22 +0000 UTC 2019-09-06 15:11:18 +0000 UTC KubeletReady kubelet is posting ready status}]
STEP: scaling "e2e-91a2d-w-0" from 2 to 0 replicas
I0906 15:11:24.870364    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
STEP: scaling "e2e-91a2d-w-1" from 2 to 0 replicas
I0906 15:11:24.892046    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
STEP: waiting for cluster to get back to original size. Final size should be 5 nodes
I0906 15:11:24.924116    5527 utils.go:239] [remaining 15m0s] Cluster size expected to be 5 nodes
I0906 15:11:24.989484    5527 utils.go:99] MachineSet "e2e-91a2d-w-0" replicas 0. Ready: 2, available 2
I0906 15:11:24.989519    5527 utils.go:99] MachineSet "e2e-91a2d-w-1" replicas 0. Ready: 2, available 2
I0906 15:11:24.989529    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset" replicas 1. Ready: 1, available 1
I0906 15:11:24.989539    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-blue" replicas 1. Ready: 1, available 1
I0906 15:11:24.989548    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-green" replicas 1. Ready: 1, available 1
I0906 15:11:24.989558    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-red" replicas 1. Ready: 1, available 1
I0906 15:11:25.006743    5527 utils.go:231] Node "41e2bf6d-a04c-4354-afea-7e711d38300e". Ready: true. Unschedulable: false
I0906 15:11:25.006770    5527 utils.go:231] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab". Ready: true. Unschedulable: false
I0906 15:11:25.006779    5527 utils.go:231] Node "86eff62d-6aee-4907-b3a3-b0af551e243b". Ready: true. Unschedulable: false
I0906 15:11:25.006787    5527 utils.go:231] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546". Ready: true. Unschedulable: false
I0906 15:11:25.006795    5527 utils.go:231] Node "a18bf460-d110-4af6-91a0-4af7a5c1fe76". Ready: true. Unschedulable: false
I0906 15:11:25.006803    5527 utils.go:231] Node "b3408843-b44c-4857-ab9d-3b13ab158aea". Ready: true. Unschedulable: false
I0906 15:11:25.006811    5527 utils.go:231] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b". Ready: true. Unschedulable: false
I0906 15:11:25.006823    5527 utils.go:231] Node "f94e2b84-7660-436d-b933-ce06e9220145". Ready: true. Unschedulable: false
I0906 15:11:25.006831    5527 utils.go:231] Node "minikube". Ready: true. Unschedulable: false
I0906 15:11:25.023990    5527 utils.go:87] Cluster size is 9 nodes
I0906 15:11:30.024230    5527 utils.go:239] [remaining 14m55s] Cluster size expected to be 5 nodes
I0906 15:11:30.029565    5527 utils.go:99] MachineSet "e2e-91a2d-w-0" replicas 0. Ready: 0, available 0
I0906 15:11:30.029588    5527 utils.go:99] MachineSet "e2e-91a2d-w-1" replicas 0. Ready: 0, available 0
I0906 15:11:30.029598    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset" replicas 1. Ready: 1, available 1
I0906 15:11:30.029607    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-blue" replicas 1. Ready: 1, available 1
I0906 15:11:30.029613    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-green" replicas 1. Ready: 1, available 1
I0906 15:11:30.029618    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-red" replicas 1. Ready: 1, available 1
I0906 15:11:30.035465    5527 utils.go:231] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab". Ready: true. Unschedulable: false
I0906 15:11:30.035485    5527 utils.go:231] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546". Ready: true. Unschedulable: false
I0906 15:11:30.035495    5527 utils.go:231] Node "b3408843-b44c-4857-ab9d-3b13ab158aea". Ready: true. Unschedulable: false
I0906 15:11:30.035503    5527 utils.go:231] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b". Ready: true. Unschedulable: false
I0906 15:11:30.035512    5527 utils.go:231] Node "minikube". Ready: true. Unschedulable: false
I0906 15:11:30.038851    5527 utils.go:87] Cluster size is 5 nodes
I0906 15:11:30.038883    5527 utils.go:257] waiting for all nodes to be ready
I0906 15:11:30.042502    5527 utils.go:262] waiting for all nodes to be schedulable
I0906 15:11:30.049309    5527 utils.go:290] [remaining 1m0s] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab" is schedulable
I0906 15:11:30.049334    5527 utils.go:290] [remaining 1m0s] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546" is schedulable
I0906 15:11:30.049346    5527 utils.go:290] [remaining 1m0s] Node "b3408843-b44c-4857-ab9d-3b13ab158aea" is schedulable
I0906 15:11:30.049357    5527 utils.go:290] [remaining 1m0s] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b" is schedulable
I0906 15:11:30.049367    5527 utils.go:290] [remaining 1m0s] Node "minikube" is schedulable
I0906 15:11:30.049376    5527 utils.go:267] waiting for each node to be backed by a machine
I0906 15:11:30.058252    5527 utils.go:47] [remaining 3m0s] Expecting the same number of machines and nodes, have 5 nodes and 5 machines
I0906 15:11:30.058283    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-6pt7l" is linked to node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab"
I0906 15:11:30.058303    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-blue-hpgct" is linked to node "8d76d38d-5446-4aef-802c-ad0fcfdb4546"
I0906 15:11:30.058318    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-green-scthk" is linked to node "b3408843-b44c-4857-ab9d-3b13ab158aea"
I0906 15:11:30.058331    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-red-s4l9g" is linked to node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b"
I0906 15:11:30.058344    5527 utils.go:70] [remaining 3m0s] Machine "minikube-static-machine" is linked to node "minikube"

• [SLOW TEST:15.344 seconds]
[Feature:Machines] Managed cluster should
/tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/infra/infra.go:126
  grow and decrease when scaling different machineSets simultaneously
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/infra/infra.go:267
------------------------------
[Feature:Machines] Managed cluster should 
  drain node before removing machine resource
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/infra/infra.go:346
I0906 15:11:30.068510    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
STEP: checking existing cluster size
I0906 15:11:30.085166    5527 utils.go:87] Cluster size is 5 nodes
STEP: Taking the first worker machineset (assuming only worker machines are backed by machinesets)
STEP: Creating two new machines, one for node about to be drained, other for moving workload from drained node
STEP: Waiting until both new nodes are ready
E0906 15:11:30.096585    5527 utils.go:342] [remaining 15m0s] Expecting 2 nodes with map[string]string{"node-role.kubernetes.io/worker":"", "node-draining-test":"54ff90f1-d0b8-11e9-978c-0a445740e986"} labels in Ready state, got 0
I0906 15:11:35.100243    5527 utils.go:346] [14m55s remaining] Expected number (2) of nodes with map[node-draining-test:54ff90f1-d0b8-11e9-978c-0a445740e986 node-role.kubernetes.io/worker:] label in Ready state found
STEP: Creating RC with workload
STEP: Creating PDB for RC
STEP: Wait until all replicas are ready
I0906 15:11:35.141657    5527 utils.go:396] [15m0s remaining] Waiting for at least one RC ready replica, ReadyReplicas: 0, Replicas: 0
I0906 15:11:40.145072    5527 utils.go:396] [14m55s remaining] Waiting for at least one RC ready replica, ReadyReplicas: 0, Replicas: 20
I0906 15:11:45.143917    5527 utils.go:399] [14m50s remaining] Waiting for RC ready replicas, ReadyReplicas: 20, Replicas: 20
I0906 15:11:45.153706    5527 utils.go:416] POD #0/20: {
  "metadata": {
    "name": "pdb-workload-5wbhf",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-5wbhf",
    "uid": "9dce09a4-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3788",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "927f2a33-8b87-455d-9a89-7c030aa4fcf2",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.23",
    "podIP": "10.211.234.220",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:40Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://2e0cea5ea8d141c4"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.153865    5527 utils.go:416] POD #1/20: {
  "metadata": {
    "name": "pdb-workload-747sq",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-747sq",
    "uid": "9dcb94ff-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3767",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "927f2a33-8b87-455d-9a89-7c030aa4fcf2",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.23",
    "podIP": "10.78.42.110",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:39Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://31ab02dda3e57412"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.154031    5527 utils.go:416] POD #2/20: {
  "metadata": {
    "name": "pdb-workload-bmgt5",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-bmgt5",
    "uid": "9dcdb415-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3816",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "310d2184-6584-443c-83cf-1df6982bea38",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.18",
    "podIP": "10.202.74.28",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:40Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://b137ec1b132a04ce"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.154199    5527 utils.go:416] POD #3/20: {
  "metadata": {
    "name": "pdb-workload-bzxqt",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-bzxqt",
    "uid": "9dcd97eb-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3804",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "310d2184-6584-443c-83cf-1df6982bea38",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.18",
    "podIP": "10.206.214.232",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:39Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://7c95463b3528976d"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.154362    5527 utils.go:416] POD #4/20: {
  "metadata": {
    "name": "pdb-workload-csr24",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-csr24",
    "uid": "9dc9102d-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3779",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "927f2a33-8b87-455d-9a89-7c030aa4fcf2",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.23",
    "podIP": "10.183.109.152",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:39Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://6659cb427942f37b"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.154490    5527 utils.go:416] POD #5/20: {
  "metadata": {
    "name": "pdb-workload-cwsx6",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-cwsx6",
    "uid": "9dcbb286-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3782",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "927f2a33-8b87-455d-9a89-7c030aa4fcf2",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.23",
    "podIP": "10.107.50.185",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:39Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://e6662125b7fe70a2"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.154610    5527 utils.go:416] POD #6/20: {
  "metadata": {
    "name": "pdb-workload-d9knp",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-d9knp",
    "uid": "9dc9dc0a-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3770",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "927f2a33-8b87-455d-9a89-7c030aa4fcf2",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.23",
    "podIP": "10.239.96.124",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:38Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://733660889ad6ab4e"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.154717    5527 utils.go:416] POD #7/20: {
  "metadata": {
    "name": "pdb-workload-fzkkf",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-fzkkf",
    "uid": "9dc9c804-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3841",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "310d2184-6584-443c-83cf-1df6982bea38",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:42Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.18",
    "podIP": "10.16.218.162",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:41Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://84965017a9a309e5"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.154838    5527 utils.go:416] POD #8/20: {
  "metadata": {
    "name": "pdb-workload-hkgss",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-hkgss",
    "uid": "9dd0f0e1-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3838",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "310d2184-6584-443c-83cf-1df6982bea38",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.18",
    "podIP": "10.24.253.172",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:39Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://874196afe775dd5e"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.154958    5527 utils.go:416] POD #9/20: {
  "metadata": {
    "name": "pdb-workload-jq5l8",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-jq5l8",
    "uid": "9dcda9db-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3773",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "927f2a33-8b87-455d-9a89-7c030aa4fcf2",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.23",
    "podIP": "10.188.188.167",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:40Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://4f58ac0b6009aa49"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.155088    5527 utils.go:416] POD #10/20: {
  "metadata": {
    "name": "pdb-workload-jvzrv",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-jvzrv",
    "uid": "9dd1759c-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3827",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "310d2184-6584-443c-83cf-1df6982bea38",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.18",
    "podIP": "10.83.195.238",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:41Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://3e1150987debfa17"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.155225    5527 utils.go:416] POD #11/20: {
  "metadata": {
    "name": "pdb-workload-lf6xd",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-lf6xd",
    "uid": "9dcb7029-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3811",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "310d2184-6584-443c-83cf-1df6982bea38",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.18",
    "podIP": "10.197.89.194",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:38Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://1ed8492024500655"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.155374    5527 utils.go:416] POD #12/20: {
  "metadata": {
    "name": "pdb-workload-qpcqf",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-qpcqf",
    "uid": "9dcde4a1-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3820",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "310d2184-6584-443c-83cf-1df6982bea38",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.18",
    "podIP": "10.61.205.104",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:40Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://1ab2d89934fe92bd"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.155505    5527 utils.go:416] POD #13/20: {
  "metadata": {
    "name": "pdb-workload-rv85j",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-rv85j",
    "uid": "9dcbaeef-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3807",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "310d2184-6584-443c-83cf-1df6982bea38",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.18",
    "podIP": "10.206.128.173",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:41Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://68d2b560e2b67fc7"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.155647    5527 utils.go:416] POD #14/20: {
  "metadata": {
    "name": "pdb-workload-t968g",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-t968g",
    "uid": "9dcdd176-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3800",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "927f2a33-8b87-455d-9a89-7c030aa4fcf2",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.23",
    "podIP": "10.103.218.85",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:40Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://b117c42312eec019"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.155781    5527 utils.go:416] POD #15/20: {
  "metadata": {
    "name": "pdb-workload-tn4bp",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-tn4bp",
    "uid": "9dce0b68-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3823",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "310d2184-6584-443c-83cf-1df6982bea38",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.18",
    "podIP": "10.236.24.225",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:39Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://41e63de3e86f5c12"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.155912    5527 utils.go:416] POD #16/20: {
  "metadata": {
    "name": "pdb-workload-w4kh2",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-w4kh2",
    "uid": "9dd1592e-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3795",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "927f2a33-8b87-455d-9a89-7c030aa4fcf2",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.23",
    "podIP": "10.142.223.35",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:39Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://d279a033e940fab0"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.156065    5527 utils.go:416] POD #17/20: {
  "metadata": {
    "name": "pdb-workload-wfsmn",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-wfsmn",
    "uid": "9dd14795-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3831",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "310d2184-6584-443c-83cf-1df6982bea38",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.18",
    "podIP": "10.220.235.32",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:40Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://36b3b7994c4f0845"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.156207    5527 utils.go:416] POD #18/20: {
  "metadata": {
    "name": "pdb-workload-zs4hj",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-zs4hj",
    "uid": "9dcdc610-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3776",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "927f2a33-8b87-455d-9a89-7c030aa4fcf2",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.23",
    "podIP": "10.28.179.7",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:40Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://3d35c91aed42ae00"
      }
    ],
    "qosClass": "Burstable"
  }
}
I0906 15:11:45.156365    5527 utils.go:416] POD #19/20: {
  "metadata": {
    "name": "pdb-workload-zvpdf",
    "generateName": "pdb-workload-",
    "namespace": "default",
    "selfLink": "/api/v1/namespaces/default/pods/pdb-workload-zvpdf",
    "uid": "9dd1a952-d0b8-11e9-b3bc-0a445740e986",
    "resourceVersion": "3785",
    "creationTimestamp": "2019-09-06T15:11:35Z",
    "labels": {
      "app": "nginx"
    },
    "ownerReferences": [
      {
        "apiVersion": "v1",
        "kind": "ReplicationController",
        "name": "pdb-workload",
        "uid": "9dc65a09-d0b8-11e9-b3bc-0a445740e986",
        "controller": true,
        "blockOwnerDeletion": true
      }
    ]
  },
  "spec": {
    "volumes": [
      {
        "name": "default-token-t266s",
        "secret": {
          "secretName": "default-token-t266s",
          "defaultMode": 420
        }
      }
    ],
    "containers": [
      {
        "name": "work",
        "image": "busybox",
        "command": [
          "sleep",
          "10h"
        ],
        "resources": {
          "requests": {
            "cpu": "50m",
            "memory": "50Mi"
          }
        },
        "volumeMounts": [
          {
            "name": "default-token-t266s",
            "readOnly": true,
            "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount"
          }
        ],
        "terminationMessagePath": "/dev/termination-log",
        "terminationMessagePolicy": "File",
        "imagePullPolicy": "Always"
      }
    ],
    "restartPolicy": "Always",
    "terminationGracePeriodSeconds": 30,
    "dnsPolicy": "ClusterFirst",
    "nodeSelector": {
      "node-draining-test": "54ff90f1-d0b8-11e9-978c-0a445740e986",
      "node-role.kubernetes.io/worker": ""
    },
    "serviceAccountName": "default",
    "serviceAccount": "default",
    "nodeName": "927f2a33-8b87-455d-9a89-7c030aa4fcf2",
    "securityContext": {},
    "schedulerName": "default-scheduler",
    "tolerations": [
      {
        "key": "kubemark",
        "operator": "Exists"
      },
      {
        "key": "node.kubernetes.io/not-ready",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      },
      {
        "key": "node.kubernetes.io/unreachable",
        "operator": "Exists",
        "effect": "NoExecute",
        "tolerationSeconds": 300
      }
    ],
    "priority": 0
  },
  "status": {
    "phase": "Running",
    "conditions": [
      {
        "type": "Initialized",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      },
      {
        "type": "Ready",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:41Z"
      },
      {
        "type": "ContainersReady",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": null
      },
      {
        "type": "PodScheduled",
        "status": "True",
        "lastProbeTime": null,
        "lastTransitionTime": "2019-09-06T15:11:35Z"
      }
    ],
    "hostIP": "172.17.0.23",
    "podIP": "10.218.46.158",
    "startTime": "2019-09-06T15:11:35Z",
    "containerStatuses": [
      {
        "name": "work",
        "state": {
          "running": {
            "startedAt": "2019-09-06T15:11:39Z"
          }
        },
        "lastState": {},
        "ready": true,
        "restartCount": 0,
        "image": "busybox:latest",
        "imageID": "docker://busybox:latest",
        "containerID": "docker://8be426a8c14c7d52"
      }
    ],
    "qosClass": "Burstable"
  }
}
STEP: Delete machine to trigger node draining
STEP: Observing and verifying node draining
E0906 15:11:45.165841    5527 utils.go:451] Node "310d2184-6584-443c-83cf-1df6982bea38" is expected to be marked as unschedulable, it is not
I0906 15:11:50.170823    5527 utils.go:455] [remaining 14m55s] Node "310d2184-6584-443c-83cf-1df6982bea38" is mark unschedulable as expected
I0906 15:11:50.177921    5527 utils.go:474] [remaining 14m55s] Have 9 pods scheduled to node "310d2184-6584-443c-83cf-1df6982bea38"
I0906 15:11:50.179578    5527 utils.go:490] [remaining 14m55s] RC ReadyReplicas: 20, Replicas: 20
I0906 15:11:50.179598    5527 utils.go:500] [remaining 14m55s] Expecting at most 2 pods to be scheduled to drained node "310d2184-6584-443c-83cf-1df6982bea38", got 9
I0906 15:11:55.177703    5527 utils.go:455] [remaining 14m50s] Node "310d2184-6584-443c-83cf-1df6982bea38" is mark unschedulable as expected
I0906 15:11:55.191140    5527 utils.go:474] [remaining 14m50s] Have 8 pods scheduled to node "310d2184-6584-443c-83cf-1df6982bea38"
I0906 15:11:55.195667    5527 utils.go:490] [remaining 14m50s] RC ReadyReplicas: 20, Replicas: 20
I0906 15:11:55.195696    5527 utils.go:500] [remaining 14m50s] Expecting at most 2 pods to be scheduled to drained node "310d2184-6584-443c-83cf-1df6982bea38", got 8
I0906 15:12:00.170299    5527 utils.go:455] [remaining 14m45s] Node "310d2184-6584-443c-83cf-1df6982bea38" is mark unschedulable as expected
I0906 15:12:00.177185    5527 utils.go:474] [remaining 14m45s] Have 7 pods scheduled to node "310d2184-6584-443c-83cf-1df6982bea38"
I0906 15:12:00.178921    5527 utils.go:490] [remaining 14m45s] RC ReadyReplicas: 20, Replicas: 20
I0906 15:12:00.178945    5527 utils.go:500] [remaining 14m45s] Expecting at most 2 pods to be scheduled to drained node "310d2184-6584-443c-83cf-1df6982bea38", got 7
I0906 15:12:05.169999    5527 utils.go:455] [remaining 14m40s] Node "310d2184-6584-443c-83cf-1df6982bea38" is mark unschedulable as expected
I0906 15:12:05.177028    5527 utils.go:474] [remaining 14m40s] Have 6 pods scheduled to node "310d2184-6584-443c-83cf-1df6982bea38"
I0906 15:12:05.179833    5527 utils.go:490] [remaining 14m40s] RC ReadyReplicas: 20, Replicas: 20
I0906 15:12:05.179861    5527 utils.go:500] [remaining 14m40s] Expecting at most 2 pods to be scheduled to drained node "310d2184-6584-443c-83cf-1df6982bea38", got 6
I0906 15:12:10.170902    5527 utils.go:455] [remaining 14m35s] Node "310d2184-6584-443c-83cf-1df6982bea38" is mark unschedulable as expected
I0906 15:12:10.177679    5527 utils.go:474] [remaining 14m35s] Have 5 pods scheduled to node "310d2184-6584-443c-83cf-1df6982bea38"
I0906 15:12:10.179435    5527 utils.go:490] [remaining 14m35s] RC ReadyReplicas: 20, Replicas: 20
I0906 15:12:10.179487    5527 utils.go:500] [remaining 14m35s] Expecting at most 2 pods to be scheduled to drained node "310d2184-6584-443c-83cf-1df6982bea38", got 5
I0906 15:12:15.169974    5527 utils.go:455] [remaining 14m30s] Node "310d2184-6584-443c-83cf-1df6982bea38" is mark unschedulable as expected
I0906 15:12:15.177332    5527 utils.go:474] [remaining 14m30s] Have 4 pods scheduled to node "310d2184-6584-443c-83cf-1df6982bea38"
I0906 15:12:15.178891    5527 utils.go:490] [remaining 14m30s] RC ReadyReplicas: 20, Replicas: 20
I0906 15:12:15.178918    5527 utils.go:500] [remaining 14m30s] Expecting at most 2 pods to be scheduled to drained node "310d2184-6584-443c-83cf-1df6982bea38", got 4
I0906 15:12:20.171174    5527 utils.go:455] [remaining 14m25s] Node "310d2184-6584-443c-83cf-1df6982bea38" is mark unschedulable as expected
I0906 15:12:20.177183    5527 utils.go:474] [remaining 14m25s] Have 3 pods scheduled to node "310d2184-6584-443c-83cf-1df6982bea38"
I0906 15:12:20.178915    5527 utils.go:490] [remaining 14m25s] RC ReadyReplicas: 20, Replicas: 20
I0906 15:12:20.178944    5527 utils.go:500] [remaining 14m25s] Expecting at most 2 pods to be scheduled to drained node "310d2184-6584-443c-83cf-1df6982bea38", got 3
I0906 15:12:25.170112    5527 utils.go:455] [remaining 14m20s] Node "310d2184-6584-443c-83cf-1df6982bea38" is mark unschedulable as expected
I0906 15:12:25.176608    5527 utils.go:474] [remaining 14m20s] Have 2 pods scheduled to node "310d2184-6584-443c-83cf-1df6982bea38"
I0906 15:12:25.178235    5527 utils.go:490] [remaining 14m20s] RC ReadyReplicas: 20, Replicas: 20
I0906 15:12:25.178259    5527 utils.go:504] [remaining 14m20s] Expected result: all pods from the RC up to last one or two got scheduled to a different node while respecting PDB
STEP: Validating the machine is deleted
E0906 15:12:25.179998    5527 infra.go:454] Machine "machine1" not yet deleted
E0906 15:12:30.182527    5527 infra.go:454] Machine "machine1" not yet deleted
I0906 15:12:35.182231    5527 infra.go:463] Machine "machine1" successfully deleted
STEP: Validate underlying node corresponding to machine1 is removed as well
I0906 15:12:35.183733    5527 utils.go:530] [15m0s remaining] Node "310d2184-6584-443c-83cf-1df6982bea38" successfully deleted
STEP: Delete PDB
STEP: Delete machine2
STEP: waiting for cluster to get back to original size. Final size should be 5 nodes
I0906 15:12:35.191084    5527 utils.go:239] [remaining 15m0s] Cluster size expected to be 5 nodes
I0906 15:12:35.197497    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset" replicas 1. Ready: 1, available 1
I0906 15:12:35.197522    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-blue" replicas 1. Ready: 1, available 1
I0906 15:12:35.197532    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-green" replicas 1. Ready: 1, available 1
I0906 15:12:35.197541    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-red" replicas 1. Ready: 1, available 1
I0906 15:12:35.201850    5527 utils.go:231] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab". Ready: true. Unschedulable: false
I0906 15:12:35.201871    5527 utils.go:231] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546". Ready: true. Unschedulable: false
I0906 15:12:35.201881    5527 utils.go:231] Node "927f2a33-8b87-455d-9a89-7c030aa4fcf2". Ready: true. Unschedulable: true
I0906 15:12:35.201889    5527 utils.go:231] Node "b3408843-b44c-4857-ab9d-3b13ab158aea". Ready: true. Unschedulable: false
I0906 15:12:35.201897    5527 utils.go:231] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b". Ready: true. Unschedulable: false
I0906 15:12:35.201909    5527 utils.go:231] Node "minikube". Ready: true. Unschedulable: false
I0906 15:12:35.205873    5527 utils.go:87] Cluster size is 6 nodes
I0906 15:12:40.206153    5527 utils.go:239] [remaining 14m55s] Cluster size expected to be 5 nodes
I0906 15:12:40.209599    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset" replicas 1. Ready: 1, available 1
I0906 15:12:40.209627    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-blue" replicas 1. Ready: 1, available 1
I0906 15:12:40.209637    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-green" replicas 1. Ready: 1, available 1
I0906 15:12:40.209646    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-red" replicas 1. Ready: 1, available 1
I0906 15:12:40.212928    5527 utils.go:231] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab". Ready: true. Unschedulable: false
I0906 15:12:40.212949    5527 utils.go:231] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546". Ready: true. Unschedulable: false
I0906 15:12:40.212955    5527 utils.go:231] Node "927f2a33-8b87-455d-9a89-7c030aa4fcf2". Ready: true. Unschedulable: true
I0906 15:12:40.212963    5527 utils.go:231] Node "b3408843-b44c-4857-ab9d-3b13ab158aea". Ready: true. Unschedulable: false
I0906 15:12:40.212973    5527 utils.go:231] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b". Ready: true. Unschedulable: false
I0906 15:12:40.212981    5527 utils.go:231] Node "minikube". Ready: true. Unschedulable: false
I0906 15:12:40.216851    5527 utils.go:87] Cluster size is 6 nodes
I0906 15:12:45.206118    5527 utils.go:239] [remaining 14m50s] Cluster size expected to be 5 nodes
I0906 15:12:45.209024    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset" replicas 1. Ready: 1, available 1
I0906 15:12:45.209051    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-blue" replicas 1. Ready: 1, available 1
I0906 15:12:45.209061    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-green" replicas 1. Ready: 1, available 1
I0906 15:12:45.209070    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-red" replicas 1. Ready: 1, available 1
I0906 15:12:45.212171    5527 utils.go:231] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab". Ready: true. Unschedulable: false
I0906 15:12:45.212193    5527 utils.go:231] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546". Ready: true. Unschedulable: false
I0906 15:12:45.212203    5527 utils.go:231] Node "927f2a33-8b87-455d-9a89-7c030aa4fcf2". Ready: true. Unschedulable: true
I0906 15:12:45.212212    5527 utils.go:231] Node "b3408843-b44c-4857-ab9d-3b13ab158aea". Ready: true. Unschedulable: false
I0906 15:12:45.212220    5527 utils.go:231] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b". Ready: true. Unschedulable: false
I0906 15:12:45.212228    5527 utils.go:231] Node "minikube". Ready: true. Unschedulable: false
I0906 15:12:45.216040    5527 utils.go:87] Cluster size is 6 nodes
I0906 15:12:50.206134    5527 utils.go:239] [remaining 14m45s] Cluster size expected to be 5 nodes
I0906 15:12:50.209012    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset" replicas 1. Ready: 1, available 1
I0906 15:12:50.209034    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-blue" replicas 1. Ready: 1, available 1
I0906 15:12:50.209040    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-green" replicas 1. Ready: 1, available 1
I0906 15:12:50.209046    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-red" replicas 1. Ready: 1, available 1
I0906 15:12:50.211890    5527 utils.go:231] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab". Ready: true. Unschedulable: false
I0906 15:12:50.211910    5527 utils.go:231] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546". Ready: true. Unschedulable: false
I0906 15:12:50.211916    5527 utils.go:231] Node "927f2a33-8b87-455d-9a89-7c030aa4fcf2". Ready: true. Unschedulable: true
I0906 15:12:50.211921    5527 utils.go:231] Node "b3408843-b44c-4857-ab9d-3b13ab158aea". Ready: true. Unschedulable: false
I0906 15:12:50.211929    5527 utils.go:231] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b". Ready: true. Unschedulable: false
I0906 15:12:50.211937    5527 utils.go:231] Node "minikube". Ready: true. Unschedulable: false
I0906 15:12:50.218996    5527 utils.go:87] Cluster size is 6 nodes
I0906 15:12:55.206486    5527 utils.go:239] [remaining 14m40s] Cluster size expected to be 5 nodes
I0906 15:12:55.209887    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset" replicas 1. Ready: 1, available 1
I0906 15:12:55.209921    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-blue" replicas 1. Ready: 1, available 1
I0906 15:12:55.209933    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-green" replicas 1. Ready: 1, available 1
I0906 15:12:55.209944    5527 utils.go:99] MachineSet "kubemark-actuator-testing-machineset-red" replicas 1. Ready: 1, available 1
I0906 15:12:55.212888    5527 utils.go:231] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab". Ready: true. Unschedulable: false
I0906 15:12:55.212917    5527 utils.go:231] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546". Ready: true. Unschedulable: false
I0906 15:12:55.212928    5527 utils.go:231] Node "b3408843-b44c-4857-ab9d-3b13ab158aea". Ready: true. Unschedulable: false
I0906 15:12:55.212937    5527 utils.go:231] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b". Ready: true. Unschedulable: false
I0906 15:12:55.212947    5527 utils.go:231] Node "minikube". Ready: true. Unschedulable: false
I0906 15:12:55.216206    5527 utils.go:87] Cluster size is 5 nodes
I0906 15:12:55.216237    5527 utils.go:257] waiting for all nodes to be ready
I0906 15:12:55.220294    5527 utils.go:262] waiting for all nodes to be schedulable
I0906 15:12:55.223696    5527 utils.go:290] [remaining 1m0s] Node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab" is schedulable
I0906 15:12:55.223727    5527 utils.go:290] [remaining 1m0s] Node "8d76d38d-5446-4aef-802c-ad0fcfdb4546" is schedulable
I0906 15:12:55.223740    5527 utils.go:290] [remaining 1m0s] Node "b3408843-b44c-4857-ab9d-3b13ab158aea" is schedulable
I0906 15:12:55.223750    5527 utils.go:290] [remaining 1m0s] Node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b" is schedulable
I0906 15:12:55.223761    5527 utils.go:290] [remaining 1m0s] Node "minikube" is schedulable
I0906 15:12:55.223770    5527 utils.go:267] waiting for each node to be backed by a machine
I0906 15:12:55.232480    5527 utils.go:47] [remaining 3m0s] Expecting the same number of machines and nodes, have 5 nodes and 5 machines
I0906 15:12:55.232512    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-6pt7l" is linked to node "6ed3bc5e-d85d-4e5c-bce4-61d11ef633ab"
I0906 15:12:55.232527    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-blue-hpgct" is linked to node "8d76d38d-5446-4aef-802c-ad0fcfdb4546"
I0906 15:12:55.232541    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-green-scthk" is linked to node "b3408843-b44c-4857-ab9d-3b13ab158aea"
I0906 15:12:55.232555    5527 utils.go:70] [remaining 3m0s] Machine "kubemark-actuator-testing-machineset-red-s4l9g" is linked to node "c81bafaa-7edf-4fb4-b5c9-b78f1548066b"
I0906 15:12:55.232569    5527 utils.go:70] [remaining 3m0s] Machine "minikube-static-machine" is linked to node "minikube"
I0906 15:12:55.242816    5527 utils.go:378] [15m0s remaining] Found 0 number of nodes with map[node-role.kubernetes.io/worker: node-draining-test:54ff90f1-d0b8-11e9-978c-0a445740e986] label as expected

• [SLOW TEST:85.174 seconds]
[Feature:Machines] Managed cluster should
/tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/infra/infra.go:126
  drain node before removing machine resource
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/infra/infra.go:346
------------------------------
[Feature:Machines] Managed cluster should 
  reject invalid machinesets
  /tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/infra/infra.go:487
I0906 15:12:55.242925    5527 framework.go:406] >>> kubeConfig: /root/.kube/config
STEP: Creating invalid machineset
STEP: Waiting for ReconcileError MachineSet event
I0906 15:12:55.327608    5527 infra.go:506] Fetching ReconcileError MachineSet invalid-machineset event
I0906 15:12:55.327648    5527 infra.go:512] Found ReconcileError event for "invalid-machineset" machine set with the following message: "invalid-machineset" machineset validation failed: spec.template.metadata.labels: Invalid value: map[string]string{"big-kitty":"i-am-bit-kitty"}: `selector` does not match template `labels`
STEP: Verify no machine from "invalid-machineset" machineset were created
I0906 15:12:55.330968    5527 infra.go:528] Have 0 machines generated from "invalid-machineset" machineset
STEP: Deleting invalid machineset
•
Ran 7 of 16 Specs in 202.323 seconds
SUCCESS! -- 7 Passed | 0 Failed | 0 Pending | 9 Skipped
--- PASS: TestE2E (202.32s)
PASS
ok  	github.com/openshift/cluster-api-actuator-pkg/pkg/e2e	202.381s
make[1]: Leaving directory `/tmp/tmp.3XEIfW31vl/src/github.com/openshift/cluster-api-actuator-pkg'
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: RUN E2E TESTS [00h 04m 28s] ##########
[PostBuildScript] - Executing post build scripts.
[workspace] $ /bin/bash /tmp/jenkins6506429167453858818.sh
########## STARTING STAGE: DOWNLOAD ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/gathered
+ rm -rf /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/gathered
+ mkdir -p /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/gathered
+ tree /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/gathered
/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/gathered

0 directories, 0 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins5975698528833924011.sh
########## STARTING STAGE: GENERATE ARTIFACTS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/generated
+ rm -rf /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/generated
+ mkdir /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/generated
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a 2>&1'
  WARNING: You're not using the default seccomp profile
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo cat /etc/sysconfig/docker /etc/sysconfig/docker-network /etc/sysconfig/docker-storage /etc/sysconfig/docker-storage-setup /etc/systemd/system/docker.service 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1 2>&1'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC 2>&1'
+ true
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo df -T -h && sudo pvs && sudo vgs && sudo lvs && sudo findmnt --all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo yum list installed 2>&1'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl --dmesg --no-pager --all --lines=all 2>&1'
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl _PID=1 --no-pager --all --lines=all 2>&1'
+ tree /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/generated
/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/generated
├── avc_denials.log
├── containers.log
├── dmesg.log
├── docker.config
├── docker.info
├── filesystem.info
├── installed_packages.log
└── pid1.journal

0 directories, 8 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins7054677787600306698.sh
########## STARTING STAGE: FETCH SYSTEMD JOURNALS FROM THE REMOTE HOST ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
+ trap 'exit 0' EXIT
++ pwd
+ ARTIFACT_DIR=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/journals
+ rm -rf /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/journals
+ mkdir /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/journals
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit docker.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all
+ tree /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/journals
/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/artifacts/journals
├── dnsmasq.service
├── docker.service
└── systemd-journald.service

0 directories, 3 files
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins2021010689578215985.sh
########## STARTING STAGE: ASSEMBLE GCS OUTPUT ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
+ trap 'exit 0' EXIT
+ mkdir -p gcs/artifacts gcs/artifacts/generated gcs/artifacts/journals gcs/artifacts/gathered
++ python -c 'import json; import urllib; print json.load(urllib.urlopen('\''https://ci.openshift.redhat.com/jenkins/job/pull-ci-openshift-machine-api-operator-master-e2e/716/api/json'\''))['\''result'\'']'
+ result=SUCCESS
+ cat
++ date +%s
+ cat /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/builds/716/log
+ cp artifacts/generated/avc_denials.log artifacts/generated/containers.log artifacts/generated/dmesg.log artifacts/generated/docker.config artifacts/generated/docker.info artifacts/generated/filesystem.info artifacts/generated/installed_packages.log artifacts/generated/pid1.journal gcs/artifacts/generated/
+ cp artifacts/journals/dnsmasq.service artifacts/journals/docker.service artifacts/journals/systemd-journald.service gcs/artifacts/journals/
+ cp -r 'artifacts/gathered/*' gcs/artifacts/
cp: cannot stat ‘artifacts/gathered/*’: No such file or directory
++ export status=FAILURE
++ status=FAILURE
+ exit 0
[workspace] $ /bin/bash /tmp/jenkins5167898857619189381.sh
########## STARTING STAGE: PUSH THE ARTIFACTS AND METADATA ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ mktemp
+ script=/tmp/tmp.O7bm8Z9vJ4
+ cat
+ chmod +x /tmp/tmp.O7bm8Z9vJ4
+ scp -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.O7bm8Z9vJ4 openshiftdevel:/tmp/tmp.O7bm8Z9vJ4
+ ssh -F /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 300 /tmp/tmp.O7bm8Z9vJ4"'
+ cd /home/origin
+ trap 'exit 0' EXIT
+ [[ -n {"type":"presubmit","job":"pull-ci-openshift-machine-api-operator-master-e2e","buildid":"1169987167749935104","prowjobid":"2695ec19-d0b6-11e9-a06a-0a58ac108d5e","refs":{"org":"openshift","repo":"machine-api-operator","repo_link":"https://github.com/openshift/machine-api-operator","base_ref":"master","base_sha":"474e14e4965a8c5e6788417c851ccc7fad1acb3a","base_link":"https://github.com/openshift/machine-api-operator/commit/474e14e4965a8c5e6788417c851ccc7fad1acb3a","pulls":[{"number":389,"author":"sadasu","sha":"229c7ea627e98ef3b7c1927a25352d366fea7023","link":"https://github.com/openshift/machine-api-operator/pull/389","commit_link":"https://github.com/openshift/machine-api-operator/pull/389/commits/229c7ea627e98ef3b7c1927a25352d366fea7023","author_link":"https://github.com/sadasu"}]}} ]]
++ jq --compact-output '.buildid |= "716"'
+ JOB_SPEC='{"type":"presubmit","job":"pull-ci-openshift-machine-api-operator-master-e2e","buildid":"716","prowjobid":"2695ec19-d0b6-11e9-a06a-0a58ac108d5e","refs":{"org":"openshift","repo":"machine-api-operator","repo_link":"https://github.com/openshift/machine-api-operator","base_ref":"master","base_sha":"474e14e4965a8c5e6788417c851ccc7fad1acb3a","base_link":"https://github.com/openshift/machine-api-operator/commit/474e14e4965a8c5e6788417c851ccc7fad1acb3a","pulls":[{"number":389,"author":"sadasu","sha":"229c7ea627e98ef3b7c1927a25352d366fea7023","link":"https://github.com/openshift/machine-api-operator/pull/389","commit_link":"https://github.com/openshift/machine-api-operator/pull/389/commits/229c7ea627e98ef3b7c1927a25352d366fea7023","author_link":"https://github.com/sadasu"}]}}'
+ docker run -e 'JOB_SPEC={"type":"presubmit","job":"pull-ci-openshift-machine-api-operator-master-e2e","buildid":"716","prowjobid":"2695ec19-d0b6-11e9-a06a-0a58ac108d5e","refs":{"org":"openshift","repo":"machine-api-operator","repo_link":"https://github.com/openshift/machine-api-operator","base_ref":"master","base_sha":"474e14e4965a8c5e6788417c851ccc7fad1acb3a","base_link":"https://github.com/openshift/machine-api-operator/commit/474e14e4965a8c5e6788417c851ccc7fad1acb3a","pulls":[{"number":389,"author":"sadasu","sha":"229c7ea627e98ef3b7c1927a25352d366fea7023","link":"https://github.com/openshift/machine-api-operator/pull/389","commit_link":"https://github.com/openshift/machine-api-operator/pull/389/commits/229c7ea627e98ef3b7c1927a25352d366fea7023","author_link":"https://github.com/sadasu"}]}}' -v /data:/data:z registry.svc.ci.openshift.org/ci/gcsupload:latest --dry-run=false --gcs-path=gs://origin-ci-test --gcs-credentials-file=/data/credentials.json --path-strategy=single --default-org=openshift --default-repo=origin '/data/gcs/*'
Unable to find image 'registry.svc.ci.openshift.org/ci/gcsupload:latest' locally
Trying to pull repository registry.svc.ci.openshift.org/ci/gcsupload ... 
latest: Pulling from registry.svc.ci.openshift.org/ci/gcsupload
a073c86ecf9e: Already exists
cc3fc741b1a9: Already exists
822bed51ba40: Pulling fs layer
85cea451eec0: Pulling fs layer
85cea451eec0: Verifying Checksum
85cea451eec0: Download complete
822bed51ba40: Download complete
822bed51ba40: Pull complete
85cea451eec0: Pull complete
Digest: sha256:03aad50d7ec631ee07c12ac2ba679bd48c7781f7d5754f9e0dcc4e7260e35208
Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/gcsupload:latest
{"component":"gcsupload","file":"prow/gcsupload/run.go:107","func":"k8s.io/test-infra/prow/gcsupload.Options.assembleTargets","level":"warning","msg":"Encountered error in resolving items to upload for /data/gcs/*: stat /data/gcs/*: no such file or directory","time":"2019-09-06T15:13:16Z"}
{"component":"gcsupload","dest":"pr-logs/directory/pull-ci-openshift-machine-api-operator-master-e2e/716.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-06T15:13:16Z"}
{"component":"gcsupload","dest":"pr-logs/directory/pull-ci-openshift-machine-api-operator-master-e2e/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-06T15:13:16Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/389/pull-ci-openshift-machine-api-operator-master-e2e/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-06T15:13:16Z"}
{"component":"gcsupload","dest":"pr-logs/directory/pull-ci-openshift-machine-api-operator-master-e2e/716.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-06T15:13:17Z"}
{"component":"gcsupload","dest":"pr-logs/pull/openshift_machine-api-operator/389/pull-ci-openshift-machine-api-operator-master-e2e/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-06T15:13:17Z"}
{"component":"gcsupload","dest":"pr-logs/directory/pull-ci-openshift-machine-api-operator-master-e2e/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-06T15:13:17Z"}
{"component":"gcsupload","file":"prow/gcsupload/run.go:65","func":"k8s.io/test-infra/prow/gcsupload.Options.Run","level":"info","msg":"Finished upload to GCS","time":"2019-09-06T15:13:17Z"}
+ exit 0
+ set +o xtrace
########## FINISHED STAGE: SUCCESS: PUSH THE ARTIFACTS AND METADATA [00h 00m 06s] ##########
[workspace] $ /bin/bash /tmp/jenkins2649458443145145662.sh
########## STARTING STAGE: DEPROVISION CLOUD RESOURCES ##########
+ [[ -s /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate ]]
+ source /var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/activate
++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed
++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin
++ unset PYTHON_HOME
++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config
+ oct deprovision

PLAYBOOK: main.yml *************************************************************
4 plays in /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml

PLAY [ensure we have the parameters necessary to deprovision virtual hosts] ****

TASK [ensure all required variables are set] ***********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:9
skipping: [localhost] => (item=origin_ci_inventory_dir)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 11:13:18.727101", 
    "item": "origin_ci_inventory_dir", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}
skipping: [localhost] => (item=origin_ci_aws_region)  => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 11:13:18.729657", 
    "item": "origin_ci_aws_region", 
    "skip_reason": "Conditional check failed", 
    "skipped": true
}

PLAY [deprovision virtual hosts in EC2] ****************************************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

TASK [deprovision a virtual EC2 host] ******************************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:28
included: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml for localhost

TASK [update the SSH configuration to remove AWS EC2 specifics] ****************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:2
ok: [localhost] => {
    "changed": false, 
    "generated_timestamp": "2019-09-06 11:13:19.545389", 
    "msg": ""
}

TASK [rename EC2 instance for termination reaper] ******************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:8
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-09-06 11:13:20.212821", 
    "msg": "Tags {'Name': 'oct-terminate'} created for resource i-06550787d42cc325e."
}

TASK [tear down the EC2 instance] **********************************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:15
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-09-06 11:13:21.285786", 
    "instance_ids": [
        "i-06550787d42cc325e"
    ], 
    "instances": [
        {
            "ami_launch_index": "0", 
            "architecture": "x86_64", 
            "block_device_mapping": {
                "/dev/sda1": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-03d9d644224906960"
                }, 
                "/dev/sdb": {
                    "delete_on_termination": true, 
                    "status": "attached", 
                    "volume_id": "vol-0bbfd421d51201f8f"
                }
            }, 
            "dns_name": "ec2-52-200-5-193.compute-1.amazonaws.com", 
            "ebs_optimized": false, 
            "groups": {
                "sg-7e73221a": "default"
            }, 
            "hypervisor": "xen", 
            "id": "i-06550787d42cc325e", 
            "image_id": "ami-0b77b87a37c3e662c", 
            "instance_type": "m4.xlarge", 
            "kernel": null, 
            "key_name": "libra", 
            "launch_time": "2019-09-06T14:54:32.000Z", 
            "placement": "us-east-1c", 
            "private_dns_name": "ip-172-18-28-208.ec2.internal", 
            "private_ip": "172.18.28.208", 
            "public_dns_name": "ec2-52-200-5-193.compute-1.amazonaws.com", 
            "public_ip": "52.200.5.193", 
            "ramdisk": null, 
            "region": "us-east-1", 
            "root_device_name": "/dev/sda1", 
            "root_device_type": "ebs", 
            "state": "running", 
            "state_code": 16, 
            "tags": {
                "Name": "oct-terminate", 
                "openshift_etcd": "", 
                "openshift_master": "", 
                "openshift_node": ""
            }, 
            "tenancy": "default", 
            "virtualization_type": "hvm"
        }
    ], 
    "tagged_instances": []
}

TASK [remove the serialized host variables] ************************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:22
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-09-06 11:13:21.523102", 
    "path": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory/host_vars/172.18.28.208.yml", 
    "state": "absent"
}

PLAY [deprovision virtual hosts locally manged by Vagrant] *********************

TASK [Gathering Facts] *********************************************************
ok: [localhost]

PLAY [clean up local configuration for deprovisioned instances] ****************

TASK [remove inventory configuration directory] ********************************
task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:61
changed: [localhost] => {
    "changed": true, 
    "generated_timestamp": "2019-09-06 11:13:22.014626", 
    "path": "/var/lib/jenkins/jobs/pull-ci-openshift-machine-api-operator-master-e2e/workspace/.config/origin-ci-tool/inventory", 
    "state": "absent"
}

PLAY RECAP *********************************************************************
localhost                  : ok=8    changed=4    unreachable=0    failed=0   

+ set +o xtrace
########## FINISHED STAGE: SUCCESS: DEPROVISION CLOUD RESOURCES [00h 00m 05s] ##########
Archiving artifacts
Recording test results
[WS-CLEANUP] Deleting project workspace...
[WS-CLEANUP] Deferred wipeout is used...
[WS-CLEANUP] done
Finished: SUCCESS