Started by user OpenShift CI Robot [EnvInject] - Loading node environment variables. Building in workspace /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace [WS-CLEANUP] Deleting project workspace... [WS-CLEANUP] Deferred wipeout is used... [workspace] $ /bin/bash /tmp/jenkins5148882511808962107.sh ########## STARTING STAGE: INSTALL THE ORIGIN-CI-TOOL ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] ++ readlink /var/lib/jenkins/origin-ci-tool/latest + latest=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed + touch /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed + cp /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin/activate /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate + cat + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config + mkdir -p /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config + rm -rf /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool + oct configure ansible-client verbosity 2 Option verbosity updated to be 2. + oct configure aws-client keypair_name libra Option keypair_name updated to be libra. + oct configure aws-client private_key_path /var/lib/jenkins/.ssh/devenv.pem Option private_key_path updated to be /var/lib/jenkins/.ssh/devenv.pem. + set +o xtrace ########## FINISHED STAGE: SUCCESS: INSTALL THE ORIGIN-CI-TOOL [00h 00m 02s] ########## [workspace] $ /bin/bash /tmp/jenkins4166248208436772377.sh ########## STARTING STAGE: PROVISION CLOUD RESOURCES ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config + oct provision remote all-in-one --os rhel --stage base --provider aws --discrete-ssh-config --name ci-kubernetes-aws-actuator_975 PLAYBOOK: aws-up.yml *********************************************************** 2 plays in /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/aws-up.yml PLAY [ensure we have the parameters necessary to bring up the AWS EC2 instance] *** TASK [ensure all required variables are set] *********************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/aws-up.yml:9 skipping: [localhost] => (item=origin_ci_inventory_dir) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.805858", "item": "origin_ci_inventory_dir", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_keypair_name) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.809592", "item": "origin_ci_aws_keypair_name", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_private_key_path) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.812721", "item": "origin_ci_aws_private_key_path", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_region) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.816009", "item": "origin_ci_aws_region", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_ami_tags) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.819092", "item": "origin_ci_aws_ami_tags", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_instance_name) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.822206", "item": "origin_ci_aws_instance_name", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_master_instance_type) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.825299", "item": "origin_ci_aws_master_instance_type", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_identifying_tag_key) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.828395", "item": "origin_ci_aws_identifying_tag_key", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_hostname) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.832629", "item": "origin_ci_aws_hostname", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_ssh_config_strategy) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.835681", "item": "origin_ci_ssh_config_strategy", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=openshift_schedulable) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.838741", "item": "openshift_schedulable", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=openshift_node_labels) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.842495", "item": "openshift_node_labels", "skip_reason": "Conditional check failed", "skipped": true } TASK [ensure all required variables are set] *********************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/aws-up.yml:27 skipping: [localhost] => (item=origin_ci_aws_master_subnet) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.878928", "item": "origin_ci_aws_master_subnet", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_etcd_security_group) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.883813", "item": "origin_ci_aws_etcd_security_group", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_node_security_group) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.888289", "item": "origin_ci_aws_node_security_group", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_master_security_group) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.893693", "item": "origin_ci_aws_master_security_group", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_master_external_elb_security_group) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.897929", "item": "origin_ci_aws_master_external_elb_security_group", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_master_internal_elb_security_group) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.903387", "item": "origin_ci_aws_master_internal_elb_security_group", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_router_security_group) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.907639", "item": "origin_ci_aws_router_security_group", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_router_elb_security_group) => { "changed": false, "generated_timestamp": "2019-09-04 10:22:23.913811", "item": "origin_ci_aws_router_elb_security_group", "skip_reason": "Conditional check failed", "skipped": true } PLAY [provision an AWS EC2 instance] ******************************************* TASK [Gathering Facts] ********************************************************* ok: [localhost] TASK [inventory : initialize the inventory directory] ************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:2 ok: [localhost] => { "changed": false, "generated_timestamp": "2019-09-04 10:22:24.779117", "gid": 995, "group": "jenkins", "mode": "0755", "owner": "jenkins", "path": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory", "secontext": "system_u:object_r:var_lib_t:s0", "size": 6, "state": "directory", "uid": 997 } TASK [inventory : add the nested group mapping] ******************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:7 changed: [localhost] => { "changed": true, "checksum": "18aaee00994df38cc3a63b635893175235331a9c", "dest": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/nested_group_mappings", "generated_timestamp": "2019-09-04 10:22:25.253814", "gid": 995, "group": "jenkins", "md5sum": "b30c3226ea63efa3ff9c5e346c14a16e", "mode": "0644", "owner": "jenkins", "secontext": "system_u:object_r:var_lib_t:s0", "size": 93, "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1567606945.02-196076740591496/source", "state": "file", "uid": 997 } TASK [inventory : initialize the OSEv3 group variables directory] ************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:12 changed: [localhost] => { "changed": true, "generated_timestamp": "2019-09-04 10:22:25.426260", "gid": 995, "group": "jenkins", "mode": "0755", "owner": "jenkins", "path": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/group_vars/OSEv3", "secontext": "system_u:object_r:var_lib_t:s0", "size": 6, "state": "directory", "uid": 997 } TASK [inventory : initialize the host variables directory] ********************* task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:17 changed: [localhost] => { "changed": true, "generated_timestamp": "2019-09-04 10:22:25.594564", "gid": 995, "group": "jenkins", "mode": "0755", "owner": "jenkins", "path": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/host_vars", "secontext": "system_u:object_r:var_lib_t:s0", "size": 6, "state": "directory", "uid": 997 } TASK [inventory : add the default Origin installation configuration] *********** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/inventory/tasks/main.yml:22 changed: [localhost] => { "changed": true, "checksum": "4c06ba508f055c20f13426e8587342e8765a7b66", "dest": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/group_vars/OSEv3/general.yml", "generated_timestamp": "2019-09-04 10:22:25.889838", "gid": 995, "group": "jenkins", "md5sum": "8aec71c75f7d512b278ae7c6f2959b12", "mode": "0644", "owner": "jenkins", "secontext": "system_u:object_r:var_lib_t:s0", "size": 331, "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1567606945.75-261165542778792/source", "state": "file", "uid": 997 } TASK [aws-up : determine if we are inside AWS EC2] ***************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:2 changed: [localhost] => { "changed": true, "cmd": [ "curl", "-s", "http://instance-data.ec2.internal" ], "delta": "0:00:00.015980", "end": "2019-09-04 10:22:26.120720", "failed": false, "failed_when_result": false, "generated_timestamp": "2019-09-04 10:22:26.138536", "rc": 0, "start": "2019-09-04 10:22:26.104740", "stderr": [], "stdout": [ "1.0", "2007-01-19", "2007-03-01", "2007-08-29", "2007-10-10", "2007-12-15", "2008-02-01", "2008-09-01", "2009-04-04", "2011-01-01", "2011-05-01", "2012-01-12", "2014-02-25", "2014-11-05", "2015-10-20", "2016-04-19", "2016-06-30", "2016-09-02", "2018-03-28", "2018-08-17", "2018-09-24", "latest" ], "warnings": [ "Consider using get_url or uri module rather than running curl" ] } [WARNING]: Consider using get_url or uri module rather than running curl TASK [aws-up : configure EC2 parameters for inventory when controlling from inside EC2] *** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:7 ok: [localhost] => { "ansible_facts": { "origin_ci_aws_destination_variable": "private_dns_name", "origin_ci_aws_host_address_variable": "private_ip", "origin_ci_aws_vpc_destination_variable": "private_ip_address" }, "changed": false, "generated_timestamp": "2019-09-04 10:22:26.178488" } TASK [aws-up : determine where to put the AWS API cache] *********************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:14 ok: [localhost] => { "ansible_facts": { "origin_ci_aws_cache_dir": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ec2_cache" }, "changed": false, "generated_timestamp": "2019-09-04 10:22:26.215956" } TASK [aws-up : ensure we have a place to put the AWS API cache] **************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:18 changed: [localhost] => { "changed": true, "generated_timestamp": "2019-09-04 10:22:26.389646", "gid": 995, "group": "jenkins", "mode": "0755", "owner": "jenkins", "path": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ec2_cache", "secontext": "system_u:object_r:var_lib_t:s0", "size": 6, "state": "directory", "uid": 997 } TASK [aws-up : place the EC2 dynamic inventory script] ************************* task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:23 changed: [localhost] => { "changed": true, "checksum": "625b8af723189db3b96ba0026d0f997a0025bc47", "dest": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/ec2.py", "generated_timestamp": "2019-09-04 10:22:26.705556", "gid": 995, "group": "jenkins", "md5sum": "cac06c14065dac74904232b89d4ba24c", "mode": "0755", "owner": "jenkins", "secontext": "system_u:object_r:var_lib_t:s0", "size": 63725, "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1567606946.56-98750286791152/source", "state": "file", "uid": 997 } TASK [aws-up : place the EC2 dynamic inventory configuration] ****************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:29 changed: [localhost] => { "changed": true, "checksum": "22430909e04b5d9e24cd874ea095965d5f51740d", "dest": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/ec2.ini", "generated_timestamp": "2019-09-04 10:22:27.028532", "gid": 995, "group": "jenkins", "md5sum": "b108b2ee7da40628c3761fde3a05fc12", "mode": "0644", "owner": "jenkins", "secontext": "system_u:object_r:var_lib_t:s0", "size": 375, "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1567606946.74-175299818913787/source", "state": "file", "uid": 997 } TASK [aws-up : place the EC2 tag to group mappings] **************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:34 changed: [localhost] => { "changed": true, "checksum": "b4205a33dc73f62bd4f77f35d045cf8e09ae62b0", "dest": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/tag_to_group_mappings", "generated_timestamp": "2019-09-04 10:22:27.350222", "gid": 995, "group": "jenkins", "md5sum": "bc3a567a1b6f342e1005182efc1b66be", "mode": "0644", "owner": "jenkins", "secontext": "system_u:object_r:var_lib_t:s0", "size": 287, "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1567606947.2-151829364871283/source", "state": "file", "uid": 997 } TASK [aws-up : list available AMIs] ******************************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:40 ok: [localhost] => { "changed": false, "generated_timestamp": "2019-09-04 10:22:30.132919", "results": [ { "ami_id": "ami-04f9b88b6b0571f20", "architecture": "x86_64", "block_device_mapping": { "/dev/sda1": { "delete_on_termination": true, "encrypted": false, "size": 75, "snapshot_id": "snap-0655d2d962c590c8c", "volume_type": "gp2" }, "/dev/sdb": { "delete_on_termination": true, "encrypted": false, "size": 50, "snapshot_id": "snap-0d86ae865b17f4def", "volume_type": "gp2" } }, "creationDate": "2018-06-26T12:22:31.000Z", "description": "OpenShift Origin development AMI on rhel at the base stage.", "hypervisor": "xen", "is_public": false, "location": "531415883065/ami_build_origin_int_rhel_base_758", "name": "ami_build_origin_int_rhel_base_758", "owner_id": "531415883065", "platform": null, "root_device_name": "/dev/sda1", "root_device_type": "ebs", "state": "available", "tags": { "Name": "ami_build_origin_int_rhel_base_758", "image_stage": "base", "operating_system": "rhel", "ready": "yes" }, "virtualization_type": "hvm" }, { "ami_id": "ami-0b77b87a37c3e662c", "architecture": "x86_64", "block_device_mapping": { "/dev/sda1": { "delete_on_termination": true, "encrypted": false, "size": 75, "snapshot_id": "snap-02ec23d4818f2747e", "volume_type": "gp2" }, "/dev/sdb": { "delete_on_termination": true, "encrypted": false, "size": 50, "snapshot_id": "snap-0d8726e441d4ca329", "volume_type": "gp2" } }, "creationDate": "2018-06-26T22:18:53.000Z", "description": "OpenShift Origin development AMI on rhel at the base stage.", "hypervisor": "xen", "is_public": false, "location": "531415883065/ami_build_origin_int_rhel_base_760", "name": "ami_build_origin_int_rhel_base_760", "owner_id": "531415883065", "platform": null, "root_device_name": "/dev/sda1", "root_device_type": "ebs", "state": "available", "tags": { "Name": "ami_build_origin_int_rhel_base_760", "image_stage": "base", "operating_system": "rhel", "ready": "yes" }, "virtualization_type": "hvm" } ] } TASK [aws-up : choose appropriate AMIs for use] ******************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:49 ok: [localhost] => (item={u'ami_id': u'ami-04f9b88b6b0571f20', u'root_device_type': u'ebs', u'description': u'OpenShift Origin development AMI on rhel at the base stage.', u'tags': {u'ready': u'yes', u'image_stage': u'base', u'Name': u'ami_build_origin_int_rhel_base_758', u'operating_system': u'rhel'}, u'hypervisor': u'xen', u'block_device_mapping': {u'/dev/sdb': {u'encrypted': False, u'snapshot_id': u'snap-0d86ae865b17f4def', u'delete_on_termination': True, u'volume_type': u'gp2', u'size': 50}, u'/dev/sda1': {u'encrypted': False, u'snapshot_id': u'snap-0655d2d962c590c8c', u'delete_on_termination': True, u'volume_type': u'gp2', u'size': 75}}, u'architecture': u'x86_64', u'owner_id': u'531415883065', u'platform': None, u'state': u'available', u'location': u'531415883065/ami_build_origin_int_rhel_base_758', u'is_public': False, u'creationDate': u'2018-06-26T12:22:31.000Z', u'root_device_name': u'/dev/sda1', u'virtualization_type': u'hvm', u'name': u'ami_build_origin_int_rhel_base_758'}) => { "ansible_facts": { "origin_ci_aws_ami_id_candidate": "ami-04f9b88b6b0571f20" }, "changed": false, "generated_timestamp": "2019-09-04 10:22:30.180938", "item": { "ami_id": "ami-04f9b88b6b0571f20", "architecture": "x86_64", "block_device_mapping": { "/dev/sda1": { "delete_on_termination": true, "encrypted": false, "size": 75, "snapshot_id": "snap-0655d2d962c590c8c", "volume_type": "gp2" }, "/dev/sdb": { "delete_on_termination": true, "encrypted": false, "size": 50, "snapshot_id": "snap-0d86ae865b17f4def", "volume_type": "gp2" } }, "creationDate": "2018-06-26T12:22:31.000Z", "description": "OpenShift Origin development AMI on rhel at the base stage.", "hypervisor": "xen", "is_public": false, "location": "531415883065/ami_build_origin_int_rhel_base_758", "name": "ami_build_origin_int_rhel_base_758", "owner_id": "531415883065", "platform": null, "root_device_name": "/dev/sda1", "root_device_type": "ebs", "state": "available", "tags": { "Name": "ami_build_origin_int_rhel_base_758", "image_stage": "base", "operating_system": "rhel", "ready": "yes" }, "virtualization_type": "hvm" } } ok: [localhost] => (item={u'ami_id': u'ami-0b77b87a37c3e662c', u'root_device_type': u'ebs', u'description': u'OpenShift Origin development AMI on rhel at the base stage.', u'tags': {u'ready': u'yes', u'image_stage': u'base', u'Name': u'ami_build_origin_int_rhel_base_760', u'operating_system': u'rhel'}, u'hypervisor': u'xen', u'block_device_mapping': {u'/dev/sdb': {u'encrypted': False, u'snapshot_id': u'snap-0d8726e441d4ca329', u'delete_on_termination': True, u'volume_type': u'gp2', u'size': 50}, u'/dev/sda1': {u'encrypted': False, u'snapshot_id': u'snap-02ec23d4818f2747e', u'delete_on_termination': True, u'volume_type': u'gp2', u'size': 75}}, u'architecture': u'x86_64', u'owner_id': u'531415883065', u'platform': None, u'state': u'available', u'location': u'531415883065/ami_build_origin_int_rhel_base_760', u'is_public': False, u'creationDate': u'2018-06-26T22:18:53.000Z', u'root_device_name': u'/dev/sda1', u'virtualization_type': u'hvm', u'name': u'ami_build_origin_int_rhel_base_760'}) => { "ansible_facts": { "origin_ci_aws_ami_id_candidate": "ami-0b77b87a37c3e662c" }, "changed": false, "generated_timestamp": "2019-09-04 10:22:30.187822", "item": { "ami_id": "ami-0b77b87a37c3e662c", "architecture": "x86_64", "block_device_mapping": { "/dev/sda1": { "delete_on_termination": true, "encrypted": false, "size": 75, "snapshot_id": "snap-02ec23d4818f2747e", "volume_type": "gp2" }, "/dev/sdb": { "delete_on_termination": true, "encrypted": false, "size": 50, "snapshot_id": "snap-0d8726e441d4ca329", "volume_type": "gp2" } }, "creationDate": "2018-06-26T22:18:53.000Z", "description": "OpenShift Origin development AMI on rhel at the base stage.", "hypervisor": "xen", "is_public": false, "location": "531415883065/ami_build_origin_int_rhel_base_760", "name": "ami_build_origin_int_rhel_base_760", "owner_id": "531415883065", "platform": null, "root_device_name": "/dev/sda1", "root_device_type": "ebs", "state": "available", "tags": { "Name": "ami_build_origin_int_rhel_base_760", "image_stage": "base", "operating_system": "rhel", "ready": "yes" }, "virtualization_type": "hvm" } } TASK [aws-up : determine which AMI to use] ************************************* task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:55 ok: [localhost] => { "ansible_facts": { "origin_ci_aws_ami_id": "ami-0b77b87a37c3e662c" }, "changed": false, "generated_timestamp": "2019-09-04 10:22:30.231582" } TASK [aws-up : determine which subnets are available] ************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:60 ok: [localhost] => { "changed": false, "generated_timestamp": "2019-09-04 10:22:31.172257", "subnets": [ { "availability_zone": "us-east-1c", "available_ip_address_count": 4064, "cidr_block": "172.18.16.0/20", "default_for_az": "false", "id": "subnet-8bdb5ac2", "map_public_ip_on_launch": "true", "state": "available", "tags": { "Name": "devenv-subnet-2", "origin_ci_aws_cluster_component": "master_subnet" }, "vpc_id": "vpc-69705d0c" }, { "availability_zone": "us-east-1d", "available_ip_address_count": 3967, "cidr_block": "172.18.0.0/20", "default_for_az": "false", "id": "subnet-cf57c596", "map_public_ip_on_launch": "true", "state": "available", "tags": { "Name": "devenv-subnet-1", "origin_ci_aws_cluster_component": "master_subnet" }, "vpc_id": "vpc-69705d0c" } ] } TASK [aws-up : determine which subnets to use for the master] ****************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:67 ok: [localhost] => { "ansible_facts": { "origin_ci_aws_master_subnet_ids": [ "subnet-8bdb5ac2", "subnet-cf57c596" ] }, "changed": false, "generated_timestamp": "2019-09-04 10:22:31.212317" } TASK [aws-up : determine which security groups are available] ****************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:72 ok: [localhost] => { "changed": false, "generated_timestamp": "2019-09-04 10:22:32.003235", "security_groups": [ { "description": "default VPC security group", "group_id": "sg-7e73221a", "group_name": "default", "ip_permissions": [ { "ip_protocol": "-1", "ip_ranges": [], "ipv6_ranges": [], "prefix_list_ids": [], "user_id_group_pairs": [ { "group_id": "sg-7e73221a", "user_id": "531415883065" } ] }, { "from_port": 53, "ip_protocol": "tcp", "ip_ranges": [ { "cidr_ip": "119.254.120.64/26" }, { "cidr_ip": "209.132.176.0/20" }, { "cidr_ip": "209.132.186.34/32" }, { "cidr_ip": "213.175.37.10/32" }, { "cidr_ip": "213.175.37.9/32" }, { "cidr_ip": "38.140.108.0/24" }, { "cidr_ip": "38.99.12.232/29" }, { "cidr_ip": "4.14.33.72/30" }, { "cidr_ip": "4.14.35.88/29" }, { "cidr_ip": "50.227.40.96/29" }, { "cidr_ip": "62.40.79.66/32" }, { "cidr_ip": "66.187.224.0/20" }, { "cidr_ip": "66.187.239.0/24" } ], "ipv6_ranges": [], "prefix_list_ids": [], "to_port": 8444, "user_id_group_pairs": [] }, { "from_port": 80, "ip_protocol": "tcp", "ip_ranges": [ { "cidr_ip": "107.20.219.35/32" }, { "cidr_ip": "108.166.48.153/32" }, { "cidr_ip": "212.199.177.64/27" }, { "cidr_ip": "212.72.208.162/32" }, { "cidr_ip": "54.241.19.245/32" }, { "cidr_ip": "97.65.119.184/29" } ], "ipv6_ranges": [], "prefix_list_ids": [], "to_port": 443, "user_id_group_pairs": [] }, { "from_port": 22, "ip_protocol": "tcp", "ip_ranges": [ { "cidr_ip": "0.0.0.0/0" } ], "ipv6_ranges": [], "prefix_list_ids": [], "to_port": 22, "user_id_group_pairs": [] }, { "from_port": 53, "ip_protocol": "udp", "ip_ranges": [ { "cidr_ip": "209.132.176.0/20" }, { "cidr_ip": "66.187.224.0/20" }, { "cidr_ip": "66.187.239.0/24" } ], "ipv6_ranges": [], "prefix_list_ids": [], "to_port": 53, "user_id_group_pairs": [] }, { "from_port": 0, "ip_protocol": "udp", "ip_ranges": [], "ipv6_ranges": [], "prefix_list_ids": [], "to_port": 65535, "user_id_group_pairs": [ { "group_id": "sg-0d1a5377", "user_id": "531415883065" }, { "group_id": "sg-5875023f", "user_id": "531415883065" }, { "group_id": "sg-7e73221a", "user_id": "531415883065" }, { "group_id": "sg-e1760186", "user_id": "531415883065" } ] }, { "from_port": 3389, "ip_protocol": "tcp", "ip_ranges": [ { "cidr_ip": "0.0.0.0/0" } ], "ipv6_ranges": [], "prefix_list_ids": [], "to_port": 3389, "user_id_group_pairs": [] }, { "from_port": -1, "ip_protocol": "icmp", "ip_ranges": [ { "cidr_ip": "0.0.0.0/0" } ], "ipv6_ranges": [], "prefix_list_ids": [], "to_port": -1, "user_id_group_pairs": [] } ], "ip_permissions_egress": [ { "ip_protocol": "-1", "ip_ranges": [ { "cidr_ip": "0.0.0.0/0" } ], "ipv6_ranges": [], "prefix_list_ids": [], "user_id_group_pairs": [] } ], "owner_id": "531415883065", "tags": { "Name": "devenv-vpc", "openshift_infra": "true", "origin_ci_aws_cluster_component": "master_security_group" }, "vpc_id": "vpc-69705d0c" } ] } TASK [aws-up : determine which security group to use] ************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:79 ok: [localhost] => { "ansible_facts": { "origin_ci_aws_master_security_group_ids": [ "sg-7e73221a" ] }, "changed": false, "generated_timestamp": "2019-09-04 10:22:32.045678" } TASK [aws-up : provision an AWS EC2 instance] ********************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:84 changed: [localhost] => { "changed": true, "generated_timestamp": "2019-09-04 10:22:49.730803", "instance_ids": [ "i-00cad1e0a8c636632" ], "instances": [ { "ami_launch_index": "0", "architecture": "x86_64", "block_device_mapping": { "/dev/sda1": { "delete_on_termination": true, "status": "attached", "volume_id": "vol-097be3317dc12d72a" }, "/dev/sdb": { "delete_on_termination": true, "status": "attached", "volume_id": "vol-02974abdb2ed88770" } }, "dns_name": "ec2-54-82-177-53.compute-1.amazonaws.com", "ebs_optimized": false, "groups": { "sg-7e73221a": "default" }, "hypervisor": "xen", "id": "i-00cad1e0a8c636632", "image_id": "ami-0b77b87a37c3e662c", "instance_type": "m4.xlarge", "kernel": null, "key_name": "libra", "launch_time": "2019-09-04T14:22:33.000Z", "placement": "us-east-1c", "private_dns_name": "ip-172-18-18-57.ec2.internal", "private_ip": "172.18.18.57", "public_dns_name": "ec2-54-82-177-53.compute-1.amazonaws.com", "public_ip": "54.82.177.53", "ramdisk": null, "region": "us-east-1", "root_device_name": "/dev/sda1", "root_device_type": "ebs", "state": "running", "state_code": 16, "tags": { "Name": "ci-kubernetes-aws-actuator_975", "openshift_etcd": "", "openshift_master": "", "openshift_node": "" }, "tenancy": "default", "virtualization_type": "hvm" } ], "tagged_instances": [] } TASK [aws-up : determine the host address] ************************************* task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:110 ok: [localhost] => { "ansible_facts": { "origin_ci_aws_host": "172.18.18.57" }, "changed": false, "generated_timestamp": "2019-09-04 10:22:49.769247" } TASK [aws-up : determine the default user to use for SSH] ********************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:114 skipping: [localhost] => { "changed": false, "generated_timestamp": "2019-09-04 10:22:49.804420", "skip_reason": "Conditional check failed", "skipped": true } TASK [aws-up : determine the default user to use for SSH] ********************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:119 ok: [localhost] => { "ansible_facts": { "origin_ci_aws_ssh_user": "origin" }, "changed": false, "generated_timestamp": "2019-09-04 10:22:49.841367" } TASK [aws-up : update variables for the host] ********************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:124 changed: [localhost] => { "changed": true, "checksum": "689f1cfca8c61e4d39bd67459c271cacbb142d47", "dest": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/host_vars/172.18.18.57.yml", "generated_timestamp": "2019-09-04 10:22:50.154872", "gid": 995, "group": "jenkins", "md5sum": "3aa481207acd0989b359ca60fa17db1f", "mode": "0644", "owner": "jenkins", "secontext": "system_u:object_r:var_lib_t:s0", "size": 745, "src": "/var/lib/jenkins/.ansible/tmp/ansible-tmp-1567606970.01-215320040661327/source", "state": "file", "uid": 997 } TASK [aws-up : determine where updated SSH configuration should go] ************ task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:141 ok: [localhost] => { "ansible_facts": { "origin_ci_ssh_config_files": [ "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config" ] }, "changed": false, "generated_timestamp": "2019-09-04 10:22:50.193579" } TASK [aws-up : determine where updated SSH configuration should go] ************ task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:146 skipping: [localhost] => { "changed": false, "generated_timestamp": "2019-09-04 10:22:50.227141", "skip_reason": "Conditional check failed", "skipped": true } TASK [aws-up : ensure the targeted SSH configuration file exists] ************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:151 changed: [localhost] => (item=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config) => { "changed": true, "dest": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config", "generated_timestamp": "2019-09-04 10:22:50.405728", "gid": 995, "group": "jenkins", "item": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config", "mode": "0644", "owner": "jenkins", "secontext": "system_u:object_r:var_lib_t:s0", "size": 0, "state": "file", "uid": 997 } TASK [aws-up : update the SSH configuration] *********************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:157 changed: [localhost] => (item=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config) => { "changed": true, "generated_timestamp": "2019-09-04 10:22:50.675342", "item": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config", "msg": "Block inserted" } TASK [aws-up : wait for SSH to be available] *********************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/provision/roles/aws-up/tasks/main.yml:175 ok: [localhost] => { "changed": false, "elapsed": 82, "generated_timestamp": "2019-09-04 10:24:13.040198", "path": null, "port": 22, "search_regex": null, "state": "started" } PLAY RECAP ********************************************************************* localhost : ok=28 changed=13 unreachable=0 failed=0 + set +o xtrace ########## FINISHED STAGE: SUCCESS: PROVISION CLOUD RESOURCES [00h 01m 50s] ########## [workspace] $ /bin/bash /tmp/jenkins6730636854678162810.sh ########## STARTING STAGE: FORWARD GCS CREDENTIALS TO REMOTE HOST ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config + (( i = 0 )) + (( i < 10 )) + scp -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.config/gcloud/gcs-publisher-credentials.json openshiftdevel:/data/credentials.json + break + set +o xtrace ########## FINISHED STAGE: SUCCESS: FORWARD GCS CREDENTIALS TO REMOTE HOST [00h 00m 02s] ########## [workspace] $ /bin/bash /tmp/jenkins2063465252345873252.sh ########## STARTING STAGE: FORWARD PARAMETERS TO THE REMOTE HOST ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod o+rw /etc/environment + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''JOB_SPEC={"type":"presubmit","job":"ci-kubernetes-aws-actuator","buildid":"1169254337591906304","prowjobid":"54633cdd-cf1f-11e9-ab71-0a58ac108d31","refs":{"org":"openshift","repo":"cluster-api-provider-aws","repo_link":"https://github.com/openshift/cluster-api-provider-aws","base_ref":"master","base_sha":"d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","base_link":"https://github.com/openshift/cluster-api-provider-aws/commit/d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","pulls":[{"number":258,"author":"ingvagabund","sha":"6fc8f0a260389bc1798d5bc6fea49264a324669f","link":"https://github.com/openshift/cluster-api-provider-aws/pull/258","commit_link":"https://github.com/openshift/cluster-api-provider-aws/pull/258/commits/6fc8f0a260389bc1798d5bc6fea49264a324669f","author_link":"https://github.com/ingvagabund"}]}}'\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''buildId='\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''BUILD_ID=1169254337591906304'\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''REPO_OWNER=openshift'\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''REPO_NAME=cluster-api-provider-aws'\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_BASE_REF=master'\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_BASE_SHA=d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41'\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_REFS=master:d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41,258:6fc8f0a260389bc1798d5bc6fea49264a324669f'\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_NUMBER=258'\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''PULL_PULL_SHA=6fc8f0a260389bc1798d5bc6fea49264a324669f'\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''JOB_SPEC={"type":"presubmit","job":"ci-kubernetes-aws-actuator","buildid":"1169254337591906304","prowjobid":"54633cdd-cf1f-11e9-ab71-0a58ac108d31","refs":{"org":"openshift","repo":"cluster-api-provider-aws","repo_link":"https://github.com/openshift/cluster-api-provider-aws","base_ref":"master","base_sha":"d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","base_link":"https://github.com/openshift/cluster-api-provider-aws/commit/d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","pulls":[{"number":258,"author":"ingvagabund","sha":"6fc8f0a260389bc1798d5bc6fea49264a324669f","link":"https://github.com/openshift/cluster-api-provider-aws/pull/258","commit_link":"https://github.com/openshift/cluster-api-provider-aws/pull/258/commits/6fc8f0a260389bc1798d5bc6fea49264a324669f","author_link":"https://github.com/ingvagabund"}]}}'\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''BUILD_NUMBER=975'\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''CLONEREFS_ARGS='\'' >> /etc/environment' + set +o xtrace ########## FINISHED STAGE: SUCCESS: FORWARD PARAMETERS TO THE REMOTE HOST [00h 00m 05s] ########## [workspace] $ /bin/bash /tmp/jenkins5470517220240942062.sh ########## STARTING STAGE: SYNC REPOSITORIES ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ mktemp + script=/tmp/tmp.XLke5RqBKB + cat + chmod +x /tmp/tmp.XLke5RqBKB + scp -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.XLke5RqBKB openshiftdevel:/tmp/tmp.XLke5RqBKB + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.XLke5RqBKB"' + cd /home/origin ++ jq --compact-output '.buildid |= "975"' + JOB_SPEC='{"type":"presubmit","job":"ci-kubernetes-aws-actuator","buildid":"975","prowjobid":"54633cdd-cf1f-11e9-ab71-0a58ac108d31","refs":{"org":"openshift","repo":"cluster-api-provider-aws","repo_link":"https://github.com/openshift/cluster-api-provider-aws","base_ref":"master","base_sha":"d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","base_link":"https://github.com/openshift/cluster-api-provider-aws/commit/d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","pulls":[{"number":258,"author":"ingvagabund","sha":"6fc8f0a260389bc1798d5bc6fea49264a324669f","link":"https://github.com/openshift/cluster-api-provider-aws/pull/258","commit_link":"https://github.com/openshift/cluster-api-provider-aws/pull/258/commits/6fc8f0a260389bc1798d5bc6fea49264a324669f","author_link":"https://github.com/ingvagabund"}]}}' + for image in ''\''registry.svc.ci.openshift.org/ci/clonerefs:latest'\''' ''\''registry.svc.ci.openshift.org/ci/initupload:latest'\''' + (( i = 0 )) + (( i < 5 )) + docker pull registry.svc.ci.openshift.org/ci/clonerefs:latest Trying to pull repository registry.svc.ci.openshift.org/ci/clonerefs ... latest: Pulling from registry.svc.ci.openshift.org/ci/clonerefs 1160f4abea84: Pulling fs layer be60dbe7622d: Pulling fs layer d26b76701841: Pulling fs layer 1b90cab916ea: Pulling fs layer 3a00cbb24bdb: Pulling fs layer 1b90cab916ea: Waiting 3a00cbb24bdb: Waiting be60dbe7622d: Verifying Checksum be60dbe7622d: Download complete 1160f4abea84: Download complete d26b76701841: Verifying Checksum d26b76701841: Download complete 3a00cbb24bdb: Verifying Checksum 3a00cbb24bdb: Download complete 1b90cab916ea: Verifying Checksum 1b90cab916ea: Download complete 1160f4abea84: Pull complete be60dbe7622d: Pull complete d26b76701841: Pull complete 1b90cab916ea: Pull complete 3a00cbb24bdb: Pull complete Digest: sha256:d68e1c6c2de5c1167a79b24d5ba4f909349ca7a44fb634e214bdadc2c8b010cd Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/clonerefs:latest + break + for image in ''\''registry.svc.ci.openshift.org/ci/clonerefs:latest'\''' ''\''registry.svc.ci.openshift.org/ci/initupload:latest'\''' + (( i = 0 )) + (( i < 5 )) + docker pull registry.svc.ci.openshift.org/ci/initupload:latest Trying to pull repository registry.svc.ci.openshift.org/ci/initupload ... latest: Pulling from registry.svc.ci.openshift.org/ci/initupload a073c86ecf9e: Pulling fs layer cc3fc741b1a9: Pulling fs layer 8f72556ef119: Pulling fs layer 8e5b170ec95b: Pulling fs layer 8e5b170ec95b: Waiting a073c86ecf9e: Download complete cc3fc741b1a9: Verifying Checksum cc3fc741b1a9: Download complete 8e5b170ec95b: Verifying Checksum 8e5b170ec95b: Download complete 8f72556ef119: Verifying Checksum 8f72556ef119: Download complete a073c86ecf9e: Pull complete cc3fc741b1a9: Pull complete 8f72556ef119: Pull complete 8e5b170ec95b: Pull complete Digest: sha256:e651a6455ada7c070c439eddcd753e2e2ac1fb934c4f2a526c37a4674c8eaee4 Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/initupload:latest + break + clonerefs_args= + docker run -v /data:/data:z registry.svc.ci.openshift.org/ci/clonerefs:latest --src-root=/data --log=/data/clone.json --repo=openshift,cluster-api-provider-aws=master:d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41,258:6fc8f0a260389bc1798d5bc6fea49264a324669f {"component":"clonerefs","file":"prow/pod-utils/clone/clone.go:34","func":"k8s.io/test-infra/prow/pod-utils/clone.Run","level":"info","msg":"Cloning refs","refs":{"org":"openshift","repo":"cluster-api-provider-aws","base_ref":"master","base_sha":"d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","pulls":[{"number":258,"author":"","sha":"6fc8f0a260389bc1798d5bc6fea49264a324669f"}]},"time":"2019-09-04T14:25:32Z"} {"command":"mkdir -p /data/src/github.com/openshift/cluster-api-provider-aws","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-04T14:25:32Z"} {"command":"git init","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Initialized empty Git repository in /data/src/github.com/openshift/cluster-api-provider-aws/.git/\n","time":"2019-09-04T14:25:32Z"} {"command":"git config user.name ci-robot","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-04T14:25:32Z"} {"command":"git config user.email ci-robot@k8s.io","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-04T14:25:32Z"} {"command":"git fetch https://github.com/openshift/cluster-api-provider-aws.git --tags --prune","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"From https://github.com/openshift/cluster-api-provider-aws\n * branch HEAD -\u003e FETCH_HEAD\n * [new tag] v0.1.0 -\u003e v0.1.0\n * [new tag] v0.2.0 -\u003e v0.2.0\n","time":"2019-09-04T14:25:36Z"} {"command":"git fetch https://github.com/openshift/cluster-api-provider-aws.git master","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"From https://github.com/openshift/cluster-api-provider-aws\n * branch master -\u003e FETCH_HEAD\n","time":"2019-09-04T14:25:36Z"} {"command":"git checkout d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Note: checking out 'd4644b21c9dbeeb79215ad42e8d7f932dd3a5f41'.\n\nYou are in 'detached HEAD' state. You can look around, make experimental\nchanges and commit them, and you can discard any commits you make in this\nstate without impacting any branches by performing another checkout.\n\nIf you want to create a new branch to retain commits you create, you may\ndo so (now or later) by using -b with the checkout command again. Example:\n\n git checkout -b \u003cnew-branch-name\u003e\n\nHEAD is now at d4644b21... Merge pull request #256 from mgugino-upstream-stage/fix-invalid-delete\n","time":"2019-09-04T14:25:36Z"} {"command":"git branch --force master d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-04T14:25:36Z"} {"command":"git checkout master","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Switched to branch 'master'\n","time":"2019-09-04T14:25:37Z"} {"command":"git fetch https://github.com/openshift/cluster-api-provider-aws.git pull/258/head","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"From https://github.com/openshift/cluster-api-provider-aws\n * branch refs/pull/258/head -\u003e FETCH_HEAD\n","time":"2019-09-04T14:25:37Z"} {"command":"git merge --no-ff 6fc8f0a260389bc1798d5bc6fea49264a324669f","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"Merge made by the 'recursive' strategy.\n Makefile | 3 ++-\n pkg/apis/awsproviderconfig/v1beta1/zz_generated.deepcopy.go | 2 +-\n pkg/client/mock/client_generated.go | 3 ++-\n 3 files changed, 5 insertions(+), 3 deletions(-)\n","time":"2019-09-04T14:25:37Z"} {"command":"git submodule update --init --recursive","component":"clonerefs","error":null,"file":"prow/pod-utils/clone/clone.go:42","func":"k8s.io/test-infra/prow/pod-utils/clone.Run.func1","level":"info","msg":"Ran command","output":"","time":"2019-09-04T14:25:37Z"} {"component":"clonerefs","file":"prow/cmd/clonerefs/main.go:43","func":"main.main","level":"info","msg":"Finished cloning refs","time":"2019-09-04T14:25:37Z"} + docker run -e 'JOB_SPEC={"type":"presubmit","job":"ci-kubernetes-aws-actuator","buildid":"975","prowjobid":"54633cdd-cf1f-11e9-ab71-0a58ac108d31","refs":{"org":"openshift","repo":"cluster-api-provider-aws","repo_link":"https://github.com/openshift/cluster-api-provider-aws","base_ref":"master","base_sha":"d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","base_link":"https://github.com/openshift/cluster-api-provider-aws/commit/d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","pulls":[{"number":258,"author":"ingvagabund","sha":"6fc8f0a260389bc1798d5bc6fea49264a324669f","link":"https://github.com/openshift/cluster-api-provider-aws/pull/258","commit_link":"https://github.com/openshift/cluster-api-provider-aws/pull/258/commits/6fc8f0a260389bc1798d5bc6fea49264a324669f","author_link":"https://github.com/ingvagabund"}]}}' -v /data:/data:z registry.svc.ci.openshift.org/ci/initupload:latest --clone-log=/data/clone.json --dry-run=false --gcs-path=gs://origin-ci-test --gcs-credentials-file=/data/credentials.json --path-strategy=single --default-org=openshift --default-repo=origin {"component":"initupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/started.json","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:25:39Z"} {"component":"initupload","dest":"pr-logs/directory/ci-kubernetes-aws-actuator/975.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:25:39Z"} {"component":"initupload","dest":"pr-logs/directory/ci-kubernetes-aws-actuator/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:25:39Z"} {"component":"initupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:25:39Z"} {"component":"initupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/clone-log.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:25:39Z"} {"component":"initupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/clone-records.json","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:25:39Z"} {"component":"initupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/clone-log.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:25:40Z"} {"component":"initupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/started.json","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:25:40Z"} {"component":"initupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:25:40Z"} {"component":"initupload","dest":"pr-logs/directory/ci-kubernetes-aws-actuator/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:25:40Z"} {"component":"initupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/clone-records.json","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:25:40Z"} {"component":"initupload","dest":"pr-logs/directory/ci-kubernetes-aws-actuator/975.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:25:40Z"} {"component":"initupload","file":"prow/gcsupload/run.go:65","func":"k8s.io/test-infra/prow/gcsupload.Options.Run","level":"info","msg":"Finished upload to GCS","time":"2019-09-04T14:25:40Z"} + sudo chmod -R a+rwX /data + sudo chown -R origin:origin-git /data + set +o xtrace ########## FINISHED STAGE: SUCCESS: SYNC REPOSITORIES [00h 01m 22s] ########## [workspace] $ /bin/bash /tmp/jenkins8236251764415298432.sh ########## STARTING STAGE: FORWARD PARAMETERS TO THE REMOTE HOST ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo chmod o+rw /etc/environment + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''JOB_NAME=ci-kubernetes-aws-actuator'\'' >> /etc/environment' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'echo '\''BUILD_NUMBER=975'\'' >> /etc/environment' + set +o xtrace ########## FINISHED STAGE: SUCCESS: FORWARD PARAMETERS TO THE REMOTE HOST [00h 00m 01s] ########## [workspace] $ /bin/bash /tmp/jenkins1990493949780136937.sh ########## STARTING STAGE: UPLOAD THE DEFAULT AWS CREDENTIASL ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'mkdir -p ~/.aws' + scp -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.aws/credentials 'openshiftdevel:~/.aws' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'mkdir -p ~/.ssh' + scp -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.ssh/devenv.pem 'openshiftdevel:~/.ssh/devenv.pem' + set +o xtrace ########## FINISHED STAGE: SUCCESS: UPLOAD THE DEFAULT AWS CREDENTIASL [00h 00m 02s] ########## [workspace] $ /bin/bash /tmp/jenkins4702004125050890374.sh ########## STARTING STAGE: INSTALL MINIKUBE ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ mktemp + script=/tmp/tmp.kmpZdVRdxK + cat + chmod +x /tmp/tmp.kmpZdVRdxK + scp -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.kmpZdVRdxK openshiftdevel:/tmp/tmp.kmpZdVRdxK + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.kmpZdVRdxK"' + cd /home/origin + curl -Lo minikube https://storage.googleapis.com/minikube/releases/v0.30.0/minikube-linux-amd64 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 53 40.3M 53 21.7M 0 0 17.3M 0 0:00:02 0:00:01 0:00:01 17.4M 100 40.3M 100 40.3M 0 0 28.1M 0 0:00:01 0:00:01 --:--:-- 28.1M + chmod +x minikube + sudo mv minikube /usr/bin/ + curl -Lo kubectl https://storage.googleapis.com/kubernetes-release/release/v1.10.0/bin/linux/amd64/kubectl % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 51.7M 100 51.7M 0 0 68.7M 0 --:--:-- --:--:-- --:--:-- 68.8M + chmod +x kubectl + sudo mv kubectl /usr/bin/ + curl -Lo crictl-v1.12.0-linux-amd64.tar.gz https://github.com/kubernetes-sigs/cri-tools/releases/download/v1.12.0/crictl-v1.12.0-linux-amd64.tar.gz % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 623 0 623 0 0 3099 0 --:--:-- --:--:-- --:--:-- 3099 100 7237k 100 7237k 0 0 23.4M 0 --:--:-- --:--:-- --:--:-- 23.4M + tar -xvf crictl-v1.12.0-linux-amd64.tar.gz crictl + sudo mv crictl /usr/bin/ + set +o xtrace ########## FINISHED STAGE: SUCCESS: INSTALL MINIKUBE [00h 00m 04s] ########## [workspace] $ /bin/bash /tmp/jenkins4383539577637293953.sh ########## STARTING STAGE: DEPLOY KUBERNETES ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ mktemp + script=/tmp/tmp.kG8zMBzG3N + cat + chmod +x /tmp/tmp.kG8zMBzG3N + scp -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.kG8zMBzG3N openshiftdevel:/tmp/tmp.kG8zMBzG3N + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.kG8zMBzG3N"' + cd /home/origin + sudo setenforce 0 + sudo minikube start --vm-driver=none --extra-config=kubelet.cgroup-driver=systemd --kubernetes-version v1.12.0 --v 5 There is a newer version of minikube available (v1.3.1). Download it here: https://github.com/kubernetes/minikube/releases/tag/v1.3.1 To disable this notification, run the following: minikube config set WantUpdateNotification false Starting local Kubernetes v1.12.0 cluster... Starting VM... Creating CA: /root/.minikube/certs/ca.pem Creating client certificate: /root/.minikube/certs/cert.pem Getting VM IP address... Moving files into cluster... Downloading kubeadm v1.12.0 Downloading kubelet v1.12.0 Finished Downloading kubeadm v1.12.0 Finished Downloading kubelet v1.12.0 Setting up certs... Connecting to cluster... Setting up kubeconfig... Starting cluster components... Kubectl is now configured to use the cluster. =================== WARNING: IT IS RECOMMENDED NOT TO RUN THE NONE DRIVER ON PERSONAL WORKSTATIONS The 'none' driver will run an insecure kubernetes apiserver as root that may leave the host vulnerable to CSRF attacks When using the none driver, the kubectl config and credentials generated will be root owned and will appear in the root home directory. You will need to move the files to the appropriate location and then set the correct permissions. An example of this is below: sudo mv /root/.kube $HOME/.kube # this will write over any previous configuration sudo chown -R $USER $HOME/.kube sudo chgrp -R $USER $HOME/.kube sudo mv /root/.minikube $HOME/.minikube # this will write over any previous configuration sudo chown -R $USER $HOME/.minikube sudo chgrp -R $USER $HOME/.minikube This can also be done automatically by setting the env var CHANGE_MINIKUBE_NONE_USER=true Loading cached images from config file. + set +o xtrace ########## FINISHED STAGE: SUCCESS: DEPLOY KUBERNETES [00h 01m 03s] ########## [workspace] $ /bin/bash /tmp/jenkins707550653227925038.sh ########## STARTING STAGE: INSTALL GO 1.10.1 ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ mktemp + script=/tmp/tmp.Vcz4jNnyHz + cat + chmod +x /tmp/tmp.Vcz4jNnyHz + scp -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.Vcz4jNnyHz openshiftdevel:/tmp/tmp.Vcz4jNnyHz + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.Vcz4jNnyHz"' + cd /home/origin + mkdir -p /home/origin/bin + curl -sL -o /home/origin/bin/gimme https://raw.githubusercontent.com/travis-ci/gimme/master/gimme + chmod +x /home/origin/bin/gimme + gimme 1.10.1 unset GOOS; unset GOARCH; export GOROOT='/home/origin/.gimme/versions/go1.10.1.linux.amd64'; export PATH="/home/origin/.gimme/versions/go1.10.1.linux.amd64/bin:${PATH}"; go version >&2; export GIMME_ENV="/home/origin/.gimme/envs/go1.10.1.env" + set +o xtrace ########## FINISHED STAGE: SUCCESS: INSTALL GO 1.10.1 [00h 00m 07s] ########## [workspace] $ /bin/bash /tmp/jenkins6010614290113494595.sh ########## STARTING STAGE: BUILD THE MACHINE CONTROLLER ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ mktemp + script=/tmp/tmp.7wVavYEBGa + cat + chmod +x /tmp/tmp.7wVavYEBGa + scp -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.7wVavYEBGa openshiftdevel:/tmp/tmp.7wVavYEBGa + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.7wVavYEBGa"' + cd /home/origin + export GOPATH=/data + GOPATH=/data + cd /data/src/github.com/openshift/cluster-api-provider-aws + source /home/origin/.gimme/envs/go1.10.1.env ++ unset GOOS ++ unset GOARCH ++ export GOROOT=/home/origin/.gimme/versions/go1.10.1.linux.amd64 ++ GOROOT=/home/origin/.gimme/versions/go1.10.1.linux.amd64 ++ export PATH=/home/origin/.gimme/versions/go1.10.1.linux.amd64/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/home/origin/.local/bin:/home/origin/bin ++ PATH=/home/origin/.gimme/versions/go1.10.1.linux.amd64/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/home/origin/.local/bin:/home/origin/bin ++ go version go version go1.10.1 linux/amd64 + go get -u github.com/openshift/imagebuilder/cmd/imagebuilder + sudo mv /data/bin/imagebuilder /usr/bin + sed -i 's/FROM registry\.svc\.ci\.openshift\.org\/openshift\/origin-v4\.0:base/FROM docker\.io\/gofed\/base:baseci/' Dockerfile + sudo make images NO_DOCKER=1 imagebuilder -t "origin-aws-machine-controllers:v0.2.0-173-gb8c169e" -t "origin-aws-machine-controllers:latest" ./ --> Image registry.svc.ci.openshift.org/openshift/release:golang-1.12 was not found, pulling ... --> Pulled 0/2 layers, 12% complete --> Pulled 1/2 layers, 56% complete --> Pulled 2/2 layers, 100% complete --> Extracting --> FROM registry.svc.ci.openshift.org/openshift/release:golang-1.12 as builder --> WORKDIR /go/src/sigs.k8s.io/cluster-api-provider-aws --> COPY . . --> RUN unset VERSION && GOPROXY=off NO_DOCKER=1 make build go build -o "bin/machine-controller-manager" \ -ldflags "-X sigs.k8s.io/cluster-api-provider-aws/pkg/version.Raw=v0.2.0-173-gb8c169e -extldflags "-static"" "sigs.k8s.io/cluster-api-provider-aws/cmd/manager" go build -o bin/manager -ldflags '-extldflags "-static"' \ "sigs.k8s.io/cluster-api-provider-aws/vendor/github.com/openshift/cluster-api/cmd/manager" --> Image docker.io/gofed/base:baseci was not found, pulling ... --> Pulled 1/2 layers, 81% complete --> Pulled 2/2 layers, 100% complete --> Extracting --> FROM docker.io/gofed/base:baseci as 1 --> RUN INSTALL_PKGS=" openssh " && yum install -y $INSTALL_PKGS && rpm -V $INSTALL_PKGS && yum clean all Loaded plugins: fastestmirror, ovl Determining fastest mirrors * base: mirrors.advancedhosters.com * extras: mirrors.advancedhosters.com * updates: mirrors.advancedhosters.com Resolving Dependencies --> Running transaction check ---> Package openssh.x86_64 0:7.4p1-16.el7 will be installed --> Processing Dependency: libfipscheck.so.1()(64bit) for package: openssh-7.4p1-16.el7.x86_64 --> Running transaction check ---> Package fipscheck-lib.x86_64 0:1.4.1-6.el7 will be installed --> Processing Dependency: /usr/bin/fipscheck for package: fipscheck-lib-1.4.1-6.el7.x86_64 --> Running transaction check ---> Package fipscheck.x86_64 0:1.4.1-6.el7 will be installed --> Finished Dependency Resolution Dependencies Resolved ================================================================================ Package Arch Version Repository Size ================================================================================ Installing: openssh x86_64 7.4p1-16.el7 base 510 k Installing for dependencies: fipscheck x86_64 1.4.1-6.el7 base 21 k fipscheck-lib x86_64 1.4.1-6.el7 base 11 k Transaction Summary ================================================================================ Install 1 Package (+2 Dependent packages) Total download size: 542 k Installed size: 2.0 M Downloading packages: -------------------------------------------------------------------------------- Total 683 kB/s | 542 kB 00:00 Running transaction check Running transaction test Transaction test succeeded Running transaction Installing : fipscheck-1.4.1-6.el7.x86_64 1/3 Installing : fipscheck-lib-1.4.1-6.el7.x86_64 2/3 Installing : openssh-7.4p1-16.el7.x86_64 3/3 Verifying : fipscheck-lib-1.4.1-6.el7.x86_64 1/3 Verifying : fipscheck-1.4.1-6.el7.x86_64 2/3 Verifying : openssh-7.4p1-16.el7.x86_64 3/3 Installed: openssh.x86_64 0:7.4p1-16.el7 Dependency Installed: fipscheck.x86_64 0:1.4.1-6.el7 fipscheck-lib.x86_64 0:1.4.1-6.el7 Complete! Loaded plugins: fastestmirror, ovl Cleaning repos: base cbs-paas7-openshift-multiarch-el7-build extras updates Cleaning up list of fastest mirrors --> COPY --from=builder /go/src/sigs.k8s.io/cluster-api-provider-aws/bin/manager / --> COPY --from=builder /go/src/sigs.k8s.io/cluster-api-provider-aws/bin/machine-controller-manager / --> Committing changes to origin-aws-machine-controllers:v0.2.0-173-gb8c169e ... --> Tagged as origin-aws-machine-controllers:latest --> Done + set +o xtrace ########## FINISHED STAGE: SUCCESS: BUILD THE MACHINE CONTROLLER [00h 01m 56s] ########## [workspace] $ /bin/bash /tmp/jenkins6148646408611005647.sh ########## STARTING STAGE: CREATE CLUSTER RESOURCES ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ mktemp + script=/tmp/tmp.PSf7psfbst + cat + chmod +x /tmp/tmp.PSf7psfbst + scp -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.PSf7psfbst openshiftdevel:/tmp/tmp.PSf7psfbst + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.PSf7psfbst"' + cd /home/origin + export GOPATH=/data + GOPATH=/data + cd /data/src/github.com/openshift/cluster-api-provider-aws/ + sudo pip install awscli Collecting awscli Downloading https://files.pythonhosted.org/packages/3e/02/5b191b339489e492f3f02dbf1f3b9b586872e22183e22e99f7eda076fc89/awscli-1.16.231-py2.py3-none-any.whl (2.0MB) Collecting colorama<=0.3.9,>=0.2.5 (from awscli) Downloading https://files.pythonhosted.org/packages/db/c8/7dcf9dbcb22429512708fe3a547f8b6101c0d02137acbd892505aee57adf/colorama-0.3.9-py2.py3-none-any.whl Collecting docutils<0.16,>=0.10 (from awscli) Downloading https://files.pythonhosted.org/packages/3a/dc/bf2b15d1fa15a6f7a9e77a61b74ecbbae7258558fcda8ffc9a6638a6b327/docutils-0.15.2-py2-none-any.whl (548kB) Requirement already satisfied (use --upgrade to upgrade): PyYAML<=5.2,>=3.10; python_version != "2.6" in /usr/lib64/python2.7/site-packages (from awscli) Collecting rsa<=3.5.0,>=3.1.2 (from awscli) Downloading https://files.pythonhosted.org/packages/e1/ae/baedc9cb175552e95f3395c43055a6a5e125ae4d48a1d7a924baca83e92e/rsa-3.4.2-py2.py3-none-any.whl (46kB) Collecting s3transfer<0.3.0,>=0.2.0 (from awscli) Downloading https://files.pythonhosted.org/packages/16/8a/1fc3dba0c4923c2a76e1ff0d52b305c44606da63f718d14d3231e21c51b0/s3transfer-0.2.1-py2.py3-none-any.whl (70kB) Collecting botocore==1.12.221 (from awscli) Downloading https://files.pythonhosted.org/packages/1f/1d/53e32d29f441d36faa31790776603f8e20573f5583bac00e656e26ae1b69/botocore-1.12.221-py2.py3-none-any.whl (5.7MB) Requirement already satisfied (use --upgrade to upgrade): pyasn1>=0.1.3 in /usr/lib/python2.7/site-packages (from rsa<=3.5.0,>=3.1.2->awscli) Collecting futures<4.0.0,>=2.2.0; python_version == "2.6" or python_version == "2.7" (from s3transfer<0.3.0,>=0.2.0->awscli) Downloading https://files.pythonhosted.org/packages/d8/a6/f46ae3f1da0cd4361c344888f59ec2f5785e69c872e175a748ef6071cdb5/futures-3.3.0-py2-none-any.whl Requirement already satisfied (use --upgrade to upgrade): jmespath<1.0.0,>=0.7.1 in /usr/lib/python2.7/site-packages (from botocore==1.12.221->awscli) Collecting python-dateutil<3.0.0,>=2.1; python_version >= "2.7" (from botocore==1.12.221->awscli) Downloading https://files.pythonhosted.org/packages/41/17/c62faccbfbd163c7f57f3844689e3a78bae1f403648a6afb1d0866d87fbb/python_dateutil-2.8.0-py2.py3-none-any.whl (226kB) Collecting urllib3<1.26,>=1.20; python_version == "2.7" (from botocore==1.12.221->awscli) Downloading https://files.pythonhosted.org/packages/e6/60/247f23a7121ae632d62811ba7f273d0e58972d75e58a94d329d51550a47d/urllib3-1.25.3-py2.py3-none-any.whl (150kB) Requirement already satisfied (use --upgrade to upgrade): six>=1.5 in /usr/lib/python2.7/site-packages (from python-dateutil<3.0.0,>=2.1; python_version >= "2.7"->botocore==1.12.221->awscli) Installing collected packages: colorama, docutils, rsa, futures, python-dateutil, urllib3, botocore, s3transfer, awscli Found existing installation: python-dateutil 1.5 Uninstalling python-dateutil-1.5: Successfully uninstalled python-dateutil-1.5 Found existing installation: urllib3 1.10.2 Uninstalling urllib3-1.10.2: Successfully uninstalled urllib3-1.10.2 Successfully installed awscli-1.16.231 botocore-1.12.221 colorama-0.3.9 docutils-0.15.2 futures-3.3.0 python-dateutil-2.8.0 rsa-3.4.2 s3transfer-0.2.1 urllib3-1.25.3 You are using pip version 8.1.2, however version 19.2.3 is available. You should consider upgrading via the 'pip install --upgrade pip' command. + curl https://releases.hashicorp.com/terraform/0.11.8/terraform_0.11.8_linux_amd64.zip -o terraform_0.11.8_linux_amd64.zip % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 17.0M 100 17.0M 0 0 49.3M 0 --:--:-- --:--:-- --:--:-- 49.2M + unzip terraform_0.11.8_linux_amd64.zip Archive: terraform_0.11.8_linux_amd64.zip inflating: terraform + sudo cp ./terraform /usr/bin/. /data/src/github.com/openshift/cluster-api-provider-aws/hack /data/src/github.com/openshift/cluster-api-provider-aws + export CLUSTER_ID=pr-975-258 + CLUSTER_ID=pr-975-258 + pushd hack + set +x + AWS_DEFAULT_REGION=us-east-1 + ENVIRONMENT_ID=pr-975-258 + ./aws-provision.sh install /data/src/github.com/openshift/cluster-api-provider-aws/hack/prebuild /data/src/github.com/openshift/cluster-api-provider-aws/hack *** starting terraform with TF_VAR_environment_id=pr-975-258 Initializing modules... - module.vpc Found version 1.66.0 of terraform-aws-modules/vpc/aws on registry.terraform.io Getting source "terraform-aws-modules/vpc/aws" Initializing provider plugins... - Checking for available provider plugins on https://releases.hashicorp.com... - Downloading plugin for provider "aws" (2.26.0)... The following providers do not have any version constraints in configuration, so the latest version was installed. To prevent automatic upgrades to new major versions that may contain breaking changes, it is recommended to add version = "..." constraints to the corresponding provider blocks in configuration, with the constraint strings suggested below. * provider.aws: version = "~> 2.26" Terraform has been successfully initialized! Refreshing Terraform state in-memory prior to plan... The refreshed state will be used to calculate this plan, but will not be persisted to local or remote state storage.  data.aws_region.current: Refreshing state... data.aws_availability_zones.azs: Refreshing state... ------------------------------------------------------------------------ An execution plan has been generated and is shown below. Resource actions are indicated with the following symbols: + create  Terraform will perform the following actions:  + aws_iam_instance_profile.test_profile  id: <computed> arn: <computed> create_date: <computed> name: "pr-975-258-worker-profile" path: "/" role: "pr-975-258-role" roles.#: <computed> unique_id: <computed>   + aws_iam_role.role  id: <computed> arn: <computed> assume_role_policy: "{\n \"Version\": \"2012-10-17\",\n \"Statement\": [\n {\n \"Action\": \"sts:AssumeRole\",\n \"Principal\": {\n \"Service\": \"ec2.amazonaws.com\"\n },\n \"Effect\": \"Allow\",\n \"Sid\": \"\"\n }\n ]\n}\n" create_date: <computed> force_detach_policies: "false" max_session_duration: "3600" name: "pr-975-258-role" path: "/" unique_id: <computed>   + aws_security_group.cluster_default  id: <computed> arn: <computed> description: "pr-975-258 default security policy" egress.#: <computed> ingress.#: <computed> name: "pr-975-258-default" owner_id: <computed> revoke_rules_on_delete: "false" tags.%: "1" tags.Name: "pr-975-258-worker-sg" vpc_id: "${module.vpc.vpc_id}"   + aws_security_group_rule.allow_all  id: <computed> cidr_blocks.#: "1" cidr_blocks.0: "0.0.0.0/0" description: "SSH" from_port: "22" protocol: "tcp" security_group_id: "${aws_security_group.cluster_default.id}" self: "false" source_security_group_id: <computed> to_port: "22" type: "ingress"   + aws_security_group_rule.default_egress  id: <computed> cidr_blocks.#: "1" cidr_blocks.0: "0.0.0.0/0" from_port: "0" protocol: "-1" security_group_id: "${aws_security_group.cluster_default.id}" self: "false" source_security_group_id: <computed> to_port: "0" type: "egress"   + aws_security_group_rule.default_ingress  id: <computed> cidr_blocks.#: "1" cidr_blocks.0: "0.0.0.0/0" from_port: "0" protocol: "-1" security_group_id: "${aws_security_group.cluster_default.id}" self: "false" source_security_group_id: <computed> to_port: "0" type: "ingress"   + module.vpc.aws_internet_gateway.this  id: <computed> owner_id: <computed> tags.%: "3" tags.Environment: "dev" tags.Name: "vpc-pr-975-258" tags.Owner: "jenkins" vpc_id: "${local.vpc_id}"   + module.vpc.aws_route.public_internet_gateway  id: <computed> destination_cidr_block: "0.0.0.0/0" destination_prefix_list_id: <computed> egress_only_gateway_id: <computed> gateway_id: "${aws_internet_gateway.this.id}" instance_id: <computed> instance_owner_id: <computed> nat_gateway_id: <computed> network_interface_id: <computed> origin: <computed> route_table_id: "${aws_route_table.public.id}" state: <computed>   + module.vpc.aws_route_table.private[0]  id: <computed> owner_id: <computed> propagating_vgws.#: <computed> route.#: <computed> tags.%: "3" tags.Environment: "dev" tags.Name: "vpc-pr-975-258-private-us-east-1a" tags.Owner: "jenkins" vpc_id: "${local.vpc_id}"   + module.vpc.aws_route_table.private[1]  id: <computed> owner_id: <computed> propagating_vgws.#: <computed> route.#: <computed> tags.%: "3" tags.Environment: "dev" tags.Name: "vpc-pr-975-258-private-us-east-1b" tags.Owner: "jenkins" vpc_id: "${local.vpc_id}"   + module.vpc.aws_route_table.private[2]  id: <computed> owner_id: <computed> propagating_vgws.#: <computed> route.#: <computed> tags.%: "3" tags.Environment: "dev" tags.Name: "vpc-pr-975-258-private-us-east-1c" tags.Owner: "jenkins" vpc_id: "${local.vpc_id}"   + module.vpc.aws_route_table.public  id: <computed> owner_id: <computed> propagating_vgws.#: <computed> route.#: <computed> tags.%: "3" tags.Environment: "dev" tags.Name: "vpc-pr-975-258-public" tags.Owner: "jenkins" vpc_id: "${local.vpc_id}"   + module.vpc.aws_route_table_association.private[0]  id: <computed> route_table_id: "${element(aws_route_table.private.*.id, (var.single_nat_gateway ? 0 : count.index))}" subnet_id: "${element(aws_subnet.private.*.id, count.index)}"   + module.vpc.aws_route_table_association.private[1]  id: <computed> route_table_id: "${element(aws_route_table.private.*.id, (var.single_nat_gateway ? 0 : count.index))}" subnet_id: "${element(aws_subnet.private.*.id, count.index)}"   + module.vpc.aws_route_table_association.private[2]  id: <computed> route_table_id: "${element(aws_route_table.private.*.id, (var.single_nat_gateway ? 0 : count.index))}" subnet_id: "${element(aws_subnet.private.*.id, count.index)}"   + module.vpc.aws_route_table_association.public[0]  id: <computed> route_table_id: "${aws_route_table.public.id}" subnet_id: "${element(aws_subnet.public.*.id, count.index)}"   + module.vpc.aws_route_table_association.public[1]  id: <computed> route_table_id: "${aws_route_table.public.id}" subnet_id: "${element(aws_subnet.public.*.id, count.index)}"   + module.vpc.aws_route_table_association.public[2]  id: <computed> route_table_id: "${aws_route_table.public.id}" subnet_id: "${element(aws_subnet.public.*.id, count.index)}"   + module.vpc.aws_subnet.private[0]  id: <computed> arn: <computed> assign_ipv6_address_on_creation: "false" availability_zone: "us-east-1a" availability_zone_id: <computed> cidr_block: "10.0.1.0/24" ipv6_cidr_block: <computed> ipv6_cidr_block_association_id: <computed> map_public_ip_on_launch: "false" owner_id: <computed> tags.%: "3" tags.Environment: "dev" tags.Name: "pr-975-258" tags.Owner: "jenkins" vpc_id: "${local.vpc_id}"   + module.vpc.aws_subnet.private[1]  id: <computed> arn: <computed> assign_ipv6_address_on_creation: "false" availability_zone: "us-east-1b" availability_zone_id: <computed> cidr_block: "10.0.2.0/24" ipv6_cidr_block: <computed> ipv6_cidr_block_association_id: <computed> map_public_ip_on_launch: "false" owner_id: <computed> tags.%: "3" tags.Environment: "dev" tags.Name: "pr-975-258" tags.Owner: "jenkins" vpc_id: "${local.vpc_id}"   + module.vpc.aws_subnet.private[2]  id: <computed> arn: <computed> assign_ipv6_address_on_creation: "false" availability_zone: "us-east-1c" availability_zone_id: <computed> cidr_block: "10.0.3.0/24" ipv6_cidr_block: <computed> ipv6_cidr_block_association_id: <computed> map_public_ip_on_launch: "false" owner_id: <computed> tags.%: "3" tags.Environment: "dev" tags.Name: "pr-975-258" tags.Owner: "jenkins" vpc_id: "${local.vpc_id}"   + module.vpc.aws_subnet.public[0]  id: <computed> arn: <computed> assign_ipv6_address_on_creation: "false" availability_zone: "us-east-1a" availability_zone_id: <computed> cidr_block: "10.0.101.0/24" ipv6_cidr_block: <computed> ipv6_cidr_block_association_id: <computed> map_public_ip_on_launch: "true" owner_id: <computed> tags.%: "3" tags.Environment: "dev" tags.Name: "pr-975-258-worker-foo" tags.Owner: "jenkins" vpc_id: "${local.vpc_id}"   + module.vpc.aws_subnet.public[1]  id: <computed> arn: <computed> assign_ipv6_address_on_creation: "false" availability_zone: "us-east-1b" availability_zone_id: <computed> cidr_block: "10.0.102.0/24" ipv6_cidr_block: <computed> ipv6_cidr_block_association_id: <computed> map_public_ip_on_launch: "true" owner_id: <computed> tags.%: "3" tags.Environment: "dev" tags.Name: "pr-975-258-worker-foo" tags.Owner: "jenkins" vpc_id: "${local.vpc_id}"   + module.vpc.aws_subnet.public[2]  id: <computed> arn: <computed> assign_ipv6_address_on_creation: "false" availability_zone: "us-east-1c" availability_zone_id: <computed> cidr_block: "10.0.103.0/24" ipv6_cidr_block: <computed> ipv6_cidr_block_association_id: <computed> map_public_ip_on_launch: "true" owner_id: <computed> tags.%: "3" tags.Environment: "dev" tags.Name: "pr-975-258-worker-foo" tags.Owner: "jenkins" vpc_id: "${local.vpc_id}"   + module.vpc.aws_vpc.this  id: <computed> arn: <computed> assign_generated_ipv6_cidr_block: "false" cidr_block: "10.0.0.0/16" default_network_acl_id: <computed> default_route_table_id: <computed> default_security_group_id: <computed> dhcp_options_id: <computed> enable_classiclink: <computed> enable_classiclink_dns_support: <computed> enable_dns_hostnames: "true" enable_dns_support: "true" instance_tenancy: "default" ipv6_association_id: <computed> ipv6_cidr_block: <computed> main_route_table_id: <computed> owner_id: <computed> tags.%: "3" tags.Environment: "dev" tags.Name: "vpc-pr-975-258" tags.Owner: "jenkins"   Plan: 25 to add, 0 to change, 0 to destroy. aws_iam_role.role: Creating... arn: "" => "<computed>" assume_role_policy: "" => "{\n \"Version\": \"2012-10-17\",\n \"Statement\": [\n {\n \"Action\": \"sts:AssumeRole\",\n \"Principal\": {\n \"Service\": \"ec2.amazonaws.com\"\n },\n \"Effect\": \"Allow\",\n \"Sid\": \"\"\n }\n ]\n}\n" create_date: "" => "<computed>" force_detach_policies: "" => "false" max_session_duration: "" => "3600" name: "" => "pr-975-258-role" path: "" => "/" unique_id: "" => "<computed>" module.vpc.aws_vpc.this: Creating... arn: "" => "<computed>" assign_generated_ipv6_cidr_block: "" => "false" cidr_block: "" => "10.0.0.0/16" default_network_acl_id: "" => "<computed>" default_route_table_id: "" => "<computed>" default_security_group_id: "" => "<computed>" dhcp_options_id: "" => "<computed>" enable_classiclink: "" => "<computed>" enable_classiclink_dns_support: "" => "<computed>" enable_dns_hostnames: "" => "true" enable_dns_support: "" => "true" instance_tenancy: "" => "default" ipv6_association_id: "" => "<computed>" ipv6_cidr_block: "" => "<computed>" main_route_table_id: "" => "<computed>" owner_id: "" => "<computed>" tags.%: "" => "3" tags.Environment: "" => "dev" tags.Name: "" => "vpc-pr-975-258" tags.Owner: "" => "jenkins" aws_iam_role.role: Creation complete after 0s (ID: pr-975-258-role) aws_iam_instance_profile.test_profile: Creating... arn: "" => "<computed>" create_date: "" => "<computed>" name: "" => "pr-975-258-worker-profile" path: "" => "/" role: "" => "pr-975-258-role" roles.#: "" => "<computed>" unique_id: "" => "<computed>" aws_iam_instance_profile.test_profile: Creation complete after 0s (ID: pr-975-258-worker-profile) module.vpc.aws_vpc.this: Creation complete after 2s (ID: vpc-01b6178fc04d23888) aws_security_group.cluster_default: Creating... arn: "" => "<computed>" description: "" => "pr-975-258 default security policy" egress.#: "" => "<computed>" ingress.#: "" => "<computed>" name: "" => "pr-975-258-default" owner_id: "" => "<computed>" revoke_rules_on_delete: "" => "false" tags.%: "" => "1" tags.Name: "" => "pr-975-258-worker-sg" vpc_id: "" => "vpc-01b6178fc04d23888" module.vpc.aws_route_table.private[2]: Creating... owner_id: "" => "<computed>" propagating_vgws.#: "" => "<computed>" route.#: "" => "<computed>" tags.%: "" => "3" tags.Environment: "" => "dev" tags.Name: "" => "vpc-pr-975-258-private-us-east-1c" tags.Owner: "" => "jenkins" vpc_id: "" => "vpc-01b6178fc04d23888" module.vpc.aws_subnet.private[0]: Creating... arn: "" => "<computed>" assign_ipv6_address_on_creation: "" => "false" availability_zone: "" => "us-east-1a" availability_zone_id: "" => "<computed>" cidr_block: "" => "10.0.1.0/24" ipv6_cidr_block: "" => "<computed>" ipv6_cidr_block_association_id: "" => "<computed>" map_public_ip_on_launch: "" => "false" owner_id: "" => "<computed>" tags.%: "" => "3" tags.Environment: "" => "dev" tags.Name: "" => "pr-975-258" tags.Owner: "" => "jenkins" vpc_id: "" => "vpc-01b6178fc04d23888" module.vpc.aws_subnet.public[0]: Creating... arn: "" => "<computed>" assign_ipv6_address_on_creation: "" => "false" availability_zone: "" => "us-east-1a" availability_zone_id: "" => "<computed>" cidr_block: "" => "10.0.101.0/24" ipv6_cidr_block: "" => "<computed>" ipv6_cidr_block_association_id: "" => "<computed>" map_public_ip_on_launch: "" => "true" owner_id: "" => "<computed>" tags.%: "" => "3" tags.Environment: "" => "dev" tags.Name: "" => "pr-975-258-worker-foo" tags.Owner: "" => "jenkins" vpc_id: "" => "vpc-01b6178fc04d23888" module.vpc.aws_route_table.private[0]: Creating... owner_id: "" => "<computed>" propagating_vgws.#: "" => "<computed>" route.#: "" => "<computed>" tags.%: "" => "3" tags.Environment: "" => "dev" tags.Name: "" => "vpc-pr-975-258-private-us-east-1a" tags.Owner: "" => "jenkins" vpc_id: "" => "vpc-01b6178fc04d23888" module.vpc.aws_route_table.private[1]: Creating... owner_id: "" => "<computed>" propagating_vgws.#: "" => "<computed>" route.#: "" => "<computed>" tags.%: "" => "3" tags.Environment: "" => "dev" tags.Name: "" => "vpc-pr-975-258-private-us-east-1b" tags.Owner: "" => "jenkins" vpc_id: "" => "vpc-01b6178fc04d23888" module.vpc.aws_subnet.public[2]: Creating... arn: "" => "<computed>" assign_ipv6_address_on_creation: "" => "false" availability_zone: "" => "us-east-1c" availability_zone_id: "" => "<computed>" cidr_block: "" => "10.0.103.0/24" ipv6_cidr_block: "" => "<computed>" ipv6_cidr_block_association_id: "" => "<computed>" map_public_ip_on_launch: "" => "true" owner_id: "" => "<computed>" tags.%: "" => "3" tags.Environment: "" => "dev" tags.Name: "" => "pr-975-258-worker-foo" tags.Owner: "" => "jenkins" vpc_id: "" => "vpc-01b6178fc04d23888" module.vpc.aws_subnet.private[2]: Creating... arn: "" => "<computed>" assign_ipv6_address_on_creation: "" => "false" availability_zone: "" => "us-east-1c" availability_zone_id: "" => "<computed>" cidr_block: "" => "10.0.3.0/24" ipv6_cidr_block: "" => "<computed>" ipv6_cidr_block_association_id: "" => "<computed>" map_public_ip_on_launch: "" => "false" owner_id: "" => "<computed>" tags.%: "" => "3" tags.Environment: "" => "dev" tags.Name: "" => "pr-975-258" tags.Owner: "" => "jenkins" vpc_id: "" => "vpc-01b6178fc04d23888" module.vpc.aws_subnet.public[1]: Creating... arn: "" => "<computed>" assign_ipv6_address_on_creation: "" => "false" availability_zone: "" => "us-east-1b" availability_zone_id: "" => "<computed>" cidr_block: "" => "10.0.102.0/24" ipv6_cidr_block: "" => "<computed>" ipv6_cidr_block_association_id: "" => "<computed>" map_public_ip_on_launch: "" => "true" owner_id: "" => "<computed>" tags.%: "" => "3" tags.Environment: "" => "dev" tags.Name: "" => "pr-975-258-worker-foo" tags.Owner: "" => "jenkins" vpc_id: "" => "vpc-01b6178fc04d23888" module.vpc.aws_subnet.private[1]: Creating... arn: "" => "<computed>" assign_ipv6_address_on_creation: "" => "false" availability_zone: "" => "us-east-1b" availability_zone_id: "" => "<computed>" cidr_block: "" => "10.0.2.0/24" ipv6_cidr_block: "" => "<computed>" ipv6_cidr_block_association_id: "" => "<computed>" map_public_ip_on_launch: "" => "false" owner_id: "" => "<computed>" tags.%: "" => "3" tags.Environment: "" => "dev" tags.Name: "" => "pr-975-258" tags.Owner: "" => "jenkins" vpc_id: "" => "vpc-01b6178fc04d23888" module.vpc.aws_route_table.private[2]: Creation complete after 0s (ID: rtb-0eb95634bb4bdf242) module.vpc.aws_route_table.private[0]: Creation complete after 0s (ID: rtb-01c859705a5683857) module.vpc.aws_route_table.public: Creating... owner_id: "" => "<computed>" propagating_vgws.#: "" => "<computed>" route.#: "" => "<computed>" tags.%: "" => "3" tags.Environment: "" => "dev" tags.Name: "" => "vpc-pr-975-258-public" tags.Owner: "" => "jenkins" vpc_id: "" => "vpc-01b6178fc04d23888" module.vpc.aws_internet_gateway.this: Creating... owner_id: "" => "<computed>" tags.%: "0" => "3" tags.Environment: "" => "dev" tags.Name: "" => "vpc-pr-975-258" tags.Owner: "" => "jenkins" vpc_id: "" => "vpc-01b6178fc04d23888" module.vpc.aws_route_table.private[1]: Creation complete after 0s (ID: rtb-0cc4d8ad28295d807) aws_security_group.cluster_default: Creation complete after 0s (ID: sg-03e2fd2a04defa788) aws_security_group_rule.default_egress: Creating... cidr_blocks.#: "" => "1" cidr_blocks.0: "" => "0.0.0.0/0" from_port: "" => "0" protocol: "" => "-1" security_group_id: "" => "sg-03e2fd2a04defa788" self: "" => "false" source_security_group_id: "" => "<computed>" to_port: "" => "0" type: "" => "egress" aws_security_group_rule.default_ingress: Creating... cidr_blocks.#: "" => "1" cidr_blocks.0: "" => "0.0.0.0/0" from_port: "" => "0" protocol: "" => "-1" security_group_id: "" => "sg-03e2fd2a04defa788" self: "" => "false" source_security_group_id: "" => "<computed>" to_port: "" => "0" type: "" => "ingress" module.vpc.aws_route_table.public: Creation complete after 0s (ID: rtb-02d9caddda23f272f) aws_security_group_rule.allow_all: Creating... cidr_blocks.#: "" => "1" cidr_blocks.0: "" => "0.0.0.0/0" description: "" => "SSH" from_port: "" => "22" protocol: "" => "tcp" security_group_id: "" => "sg-03e2fd2a04defa788" self: "" => "false" source_security_group_id: "" => "<computed>" to_port: "" => "22" type: "" => "ingress" module.vpc.aws_subnet.private[1]: Creation complete after 0s (ID: subnet-0adce266c18f72255) module.vpc.aws_subnet.private[2]: Creation complete after 0s (ID: subnet-078025a95b8289428) module.vpc.aws_subnet.public[2]: Creation complete after 0s (ID: subnet-0508530a7b553cace) module.vpc.aws_internet_gateway.this: Creation complete after 0s (ID: igw-0bcc735e9b9964b0d) module.vpc.aws_subnet.public[0]: Creation complete after 0s (ID: subnet-079ba6af91da7e6e2) module.vpc.aws_route.public_internet_gateway: Creating... destination_cidr_block: "" => "0.0.0.0/0" destination_prefix_list_id: "" => "<computed>" egress_only_gateway_id: "" => "<computed>" gateway_id: "" => "igw-0bcc735e9b9964b0d" instance_id: "" => "<computed>" instance_owner_id: "" => "<computed>" nat_gateway_id: "" => "<computed>" network_interface_id: "" => "<computed>" origin: "" => "<computed>" route_table_id: "" => "rtb-02d9caddda23f272f" state: "" => "<computed>" module.vpc.aws_subnet.public[1]: Creation complete after 0s (ID: subnet-0df88ffea33bee0de) module.vpc.aws_subnet.private[0]: Creation complete after 0s (ID: subnet-0527838d60f1601f6) module.vpc.aws_route_table_association.public[1]: Creating... route_table_id: "" => "rtb-02d9caddda23f272f" subnet_id: "" => "subnet-0df88ffea33bee0de" module.vpc.aws_route_table_association.public[0]: Creating... route_table_id: "" => "rtb-02d9caddda23f272f" subnet_id: "" => "subnet-079ba6af91da7e6e2" module.vpc.aws_route_table_association.public[2]: Creating... route_table_id: "" => "rtb-02d9caddda23f272f" subnet_id: "" => "subnet-0508530a7b553cace" module.vpc.aws_route_table_association.private[2]: Creating... route_table_id: "" => "rtb-0eb95634bb4bdf242" subnet_id: "" => "subnet-078025a95b8289428" module.vpc.aws_route_table_association.private[0]: Creating... route_table_id: "" => "rtb-01c859705a5683857" subnet_id: "" => "subnet-0527838d60f1601f6" module.vpc.aws_route_table_association.private[1]: Creating... route_table_id: "" => "rtb-0cc4d8ad28295d807" subnet_id: "" => "subnet-0adce266c18f72255" aws_security_group_rule.default_egress: Creation complete after 0s (ID: sgrule-712216049) module.vpc.aws_route_table_association.public[1]: Creation complete after 0s (ID: rtbassoc-08e10268bb436cde3) module.vpc.aws_route.public_internet_gateway: Creation complete after 0s (ID: r-rtb-02d9caddda23f272f1080289494) module.vpc.aws_route_table_association.private[0]: Creation complete after 0s (ID: rtbassoc-072092ddd4cbe1d4e) module.vpc.aws_route_table_association.private[2]: Creation complete after 0s (ID: rtbassoc-0a1ec90e46fdec4b8) module.vpc.aws_route_table_association.public[2]: Creation complete after 0s (ID: rtbassoc-07032d65a9f137ac4) module.vpc.aws_route_table_association.private[1]: Creation complete after 0s (ID: rtbassoc-0c6024cd954c92135) module.vpc.aws_route_table_association.public[0]: Creation complete after 1s (ID: rtbassoc-0c286cad3e587bdd3) aws_security_group_rule.default_ingress: Creation complete after 1s (ID: sgrule-1217565267) aws_security_group_rule.allow_all: Creation complete after 1s (ID: sgrule-2442055330)  Apply complete! Resources: 25 added, 0 changed, 0 destroyed.  Outputs: vpc_id = vpc-01b6178fc04d23888 /data/src/github.com/openshift/cluster-api-provider-aws + popd + sudo cp /etc/ssl/certs/ca-bundle.crt /etc/ssl/certs/ca-certificates.crt + set +o xtrace ########## FINISHED STAGE: SUCCESS: CREATE CLUSTER RESOURCES [00h 00m 31s] ########## [workspace] $ /bin/bash /tmp/jenkins6718608391732824932.sh ########## STARTING STAGE: RUN E2E TESTS ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ mktemp + script=/tmp/tmp.aVWCHYXuLX + cat + chmod +x /tmp/tmp.aVWCHYXuLX + scp -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.aVWCHYXuLX openshiftdevel:/tmp/tmp.aVWCHYXuLX + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.aVWCHYXuLX"' + cd /home/origin + set +x + chmod 0600 /home/origin/.ssh/devenv.pem + export CLUSTER_ID=pr-975-258 + CLUSTER_ID=pr-975-258 + export ENVIRONMENT_ID=pr-975-258 + ENVIRONMENT_ID=pr-975-258 + export SSH_PK=/home/origin/.ssh/devenv.pem + SSH_PK=/home/origin/.ssh/devenv.pem ++ sudo kubectl get service kubernetes -o json ++ jq .spec.clusterIP --raw-output + export CLUSTER_IP=10.96.0.1 + CLUSTER_IP=10.96.0.1 + export KUBECONFIG=/etc/kubernetes/admin.conf + KUBECONFIG=/etc/kubernetes/admin.conf + sudo -E kubectl config set-cluster kubernetes --server=https://10.96.0.1:443 Cluster "kubernetes" set. + source /home/origin/.gimme/envs/go1.10.1.env ++ unset GOOS ++ unset GOARCH ++ export GOROOT=/home/origin/.gimme/versions/go1.10.1.linux.amd64 ++ GOROOT=/home/origin/.gimme/versions/go1.10.1.linux.amd64 ++ export PATH=/home/origin/.gimme/versions/go1.10.1.linux.amd64/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/home/origin/.local/bin:/home/origin/bin ++ PATH=/home/origin/.gimme/versions/go1.10.1.linux.amd64/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/home/origin/.local/bin:/home/origin/bin ++ go version go version go1.10.1 linux/amd64 + sudo cp /home/origin/.gimme/versions/go1.10.1.linux.amd64/bin/go /bin/go ++ git describe --always --abbrev=7 + sudo -E make k8s-e2e ACTUATOR_IMAGE=origin-aws-machine-controllers:v0.2.0-173-gb8c169e # KUBECONFIG and SSH_PK dirs needs to be mounted inside a container if tests are run in containers go test -timeout 30m \ -v sigs.k8s.io/cluster-api-provider-aws/test/machines \ -kubeconfig ${KUBECONFIG:-~/.kube/config} \ -ssh-key ${SSH_PK:-~/.ssh/id_rsa} \ -machine-controller-image ${ACTUATOR_IMAGE:-gcr.io/k8s-cluster-api/aws-machine-controller:0.0.1} \ -machine-manager-image ${ACTUATOR_IMAGE:-gcr.io/k8s-cluster-api/aws-machine-controller:0.0.1} \ -nodelink-controller-image $(docker run registry.svc.ci.openshift.org/origin/release:4.2 image machine-api-operator) \ -cluster-id ${ENVIRONMENT_ID:-""} \ -ginkgo.v \ -args -v 5 -logtostderr true Unable to find image 'registry.svc.ci.openshift.org/origin/release:4.2' locally Trying to pull repository registry.svc.ci.openshift.org/origin/release ... 4.2: Pulling from registry.svc.ci.openshift.org/origin/release c8d67acdb2ff: Pulling fs layer 79d11c1a86c4: Pulling fs layer ffd263ce0b34: Pulling fs layer 73aafda5b164: Pulling fs layer 471e39299fe4: Pulling fs layer d2ad566fdcfd: Pulling fs layer 73aafda5b164: Waiting 471e39299fe4: Waiting d2ad566fdcfd: Waiting 79d11c1a86c4: Verifying Checksum 79d11c1a86c4: Download complete ffd263ce0b34: Verifying Checksum ffd263ce0b34: Download complete 73aafda5b164: Verifying Checksum 73aafda5b164: Download complete 471e39299fe4: Verifying Checksum 471e39299fe4: Download complete d2ad566fdcfd: Verifying Checksum d2ad566fdcfd: Download complete c8d67acdb2ff: Verifying Checksum c8d67acdb2ff: Download complete c8d67acdb2ff: Pull complete 79d11c1a86c4: Pull complete ffd263ce0b34: Pull complete 73aafda5b164: Pull complete 471e39299fe4: Pull complete d2ad566fdcfd: Pull complete Digest: sha256:f733a35f525a020a690d84c83afc7a7a0c91c29f0128b72847407fa647ed5de7 Status: Downloaded newer image for registry.svc.ci.openshift.org/origin/release:4.2 === RUN TestCart Running Suite: Machine Suite ============================ Random Seed: 1567607418 Will run 3 of 3 specs [sigs.k8s.io] Machines AWS actuator Can create AWS instances /data/src/sigs.k8s.io/cluster-api-provider-aws/test/machines/machines_test.go:131 STEP: Creating "namespace-84a81477-cf20-11e9-b8bb-0a17b8b8d0c8" namespace STEP: Deploying cluster API stack components STEP: Deploying cluster CRD manifest I0904 14:30:27.329534 19288 framework.go:335] create.err: <nil> I0904 14:30:27.335838 19288 framework.go:344] get.err: <nil> STEP: Deploying machine CRD manifest I0904 14:30:32.341851 19288 framework.go:335] create.err: <nil> I0904 14:30:32.346661 19288 framework.go:344] get.err: <nil> STEP: Deploying machineset CRD manifest I0904 14:30:37.350940 19288 framework.go:335] create.err: <nil> I0904 14:30:37.354920 19288 framework.go:344] get.err: <nil> STEP: Deploying machinedeployment CRD manifest I0904 14:30:42.360183 19288 framework.go:335] create.err: <nil> I0904 14:30:42.366127 19288 framework.go:344] get.err: <nil> STEP: Deploying cluster role STEP: Deploying machine API controllers STEP: Waiting until cluster objects can be listed STEP: Cluster API stack deployed STEP: Creating "pr-975-258" cluster STEP: Creating "pr-975-258-machine-9ef1e7" machine STEP: Waiting for "pr-975-258-machine-9ef1e7" machine STEP: Verify machine's underlying instance is running I0904 14:31:17.418500 19288 machines.go:80] Waiting for instance to come up I0904 14:31:22.418526 19288 machines.go:80] Waiting for instance to come up I0904 14:31:27.418514 19288 machines.go:80] Waiting for instance to come up I0904 14:31:27.543913 19288 machines.go:88] Machine is running STEP: Checking subnet STEP: Checking availability zone STEP: Checking security groups STEP: Checking IAM role STEP: Checking tags STEP: Checking machine status STEP: Deleting "pr-975-258-machine-9ef1e7" machine STEP: Verify instance is terminated I0904 14:31:37.980673 19288 machines.go:50] Waiting for instance to be terminated STEP: Deleting "pr-975-258-machine-9ef1e7" machine STEP: Verify instance is terminated I0904 14:31:48.085993 19288 machines.go:50] Waiting for instance to be terminated STEP: Deleting machine API controllers I0904 14:31:53.193621 19288 framework.go:311] del.err: <nil> I0904 14:31:53.195253 19288 framework.go:323] get.err: deployments.apps "clusterapi-controllers" not found STEP: Deleting cluster role I0904 14:31:58.201667 19288 framework.go:311] del.err: <nil> I0904 14:31:58.204075 19288 framework.go:323] get.err: clusterrolebindings.rbac.authorization.k8s.io "manager-rolebinding" not found I0904 14:32:03.208472 19288 framework.go:311] del.err: <nil> I0904 14:32:03.210679 19288 framework.go:323] get.err: clusterroles.rbac.authorization.k8s.io "manager-role" not found time="2019-09-04T14:32:03Z" level=info msg="namespace-84a81477-cf20-11e9-b8bb-0a17b8b8d0c8: &v1.Namespace{TypeMeta:v1.TypeMeta{Kind:\"\", APIVersion:\"\"}, ObjectMeta:v1.ObjectMeta{Name:\"namespace-84a81477-cf20-11e9-b8bb-0a17b8b8d0c8\", GenerateName:\"\", Namespace:\"\", SelfLink:\"\", UID:\"\", ResourceVersion:\"\", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:\"\", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.NamespaceSpec{Finalizers:[]v1.FinalizerName(nil)}, Status:v1.NamespaceStatus{Phase:\"\"}}" source="machines_test.go:73" STEP: Destroying "namespace-84a81477-cf20-11e9-b8bb-0a17b8b8d0c8" namespace • [SLOW TEST:104.916 seconds] [sigs.k8s.io] Machines /data/src/sigs.k8s.io/cluster-api-provider-aws/vendor/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/framework/framework.go:373 AWS actuator /data/src/sigs.k8s.io/cluster-api-provider-aws/test/machines/machines_test.go:87 Can create AWS instances /data/src/sigs.k8s.io/cluster-api-provider-aws/test/machines/machines_test.go:131 ------------------------------ [sigs.k8s.io] Machines AWS actuator Can create EBS volumes /data/src/sigs.k8s.io/cluster-api-provider-aws/test/machines/machines_test.go:209 STEP: Creating "namespace-c3304f9a-cf20-11e9-b8bb-0a17b8b8d0c8" namespace STEP: Deploying cluster API stack components STEP: Deploying cluster CRD manifest I0904 14:32:08.221621 19288 framework.go:335] create.err: customresourcedefinitions.apiextensions.k8s.io "clusters.cluster.k8s.io" already exists STEP: Deploying machine CRD manifest I0904 14:32:13.225638 19288 framework.go:335] create.err: customresourcedefinitions.apiextensions.k8s.io "machines.machine.openshift.io" already exists STEP: Deploying machineset CRD manifest I0904 14:32:18.230712 19288 framework.go:335] create.err: customresourcedefinitions.apiextensions.k8s.io "machinesets.machine.openshift.io" already exists STEP: Deploying machinedeployment CRD manifest I0904 14:32:23.235793 19288 framework.go:335] create.err: customresourcedefinitions.apiextensions.k8s.io "machinedeployments.machine.openshift.io" already exists STEP: Deploying cluster role STEP: Deploying machine API controllers STEP: Waiting until cluster objects can be listed STEP: Cluster API stack deployed STEP: Creating "pr-975-258" cluster STEP: Creating "pr-975-258-machine-db0ee0" machine STEP: Waiting for "pr-975-258-machine-db0ee0" machine STEP: Verify machine's underlying instance is running I0904 14:32:58.267086 19288 machines.go:80] Waiting for instance to come up I0904 14:33:03.267100 19288 machines.go:80] Waiting for instance to come up I0904 14:33:08.267088 19288 machines.go:80] Waiting for instance to come up I0904 14:33:08.381360 19288 machines.go:88] Machine is running STEP: Checking EBS volume mount STEP: Checking EBS volume size STEP: Checking EBS volume type STEP: Checking only root volume get's modified STEP: Deleting "pr-975-258-machine-db0ee0" machine STEP: Verify instance is terminated I0904 14:33:18.680913 19288 machines.go:50] Waiting for instance to be terminated STEP: Deleting machine API controllers I0904 14:33:23.736154 19288 framework.go:311] del.err: <nil> I0904 14:33:23.738385 19288 framework.go:323] get.err: deployments.apps "clusterapi-controllers" not found STEP: Deleting cluster role I0904 14:33:28.742971 19288 framework.go:311] del.err: <nil> I0904 14:33:28.744421 19288 framework.go:323] get.err: clusterrolebindings.rbac.authorization.k8s.io "manager-rolebinding" not found I0904 14:33:33.748679 19288 framework.go:311] del.err: <nil> I0904 14:33:33.750726 19288 framework.go:323] get.err: clusterroles.rbac.authorization.k8s.io "manager-role" not found time="2019-09-04T14:33:33Z" level=info msg="namespace-c3304f9a-cf20-11e9-b8bb-0a17b8b8d0c8: &v1.Namespace{TypeMeta:v1.TypeMeta{Kind:\"\", APIVersion:\"\"}, ObjectMeta:v1.ObjectMeta{Name:\"namespace-c3304f9a-cf20-11e9-b8bb-0a17b8b8d0c8\", GenerateName:\"\", Namespace:\"\", SelfLink:\"\", UID:\"\", ResourceVersion:\"\", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:\"\", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.NamespaceSpec{Finalizers:[]v1.FinalizerName(nil)}, Status:v1.NamespaceStatus{Phase:\"\"}}" source="machines_test.go:73" STEP: Destroying "namespace-c3304f9a-cf20-11e9-b8bb-0a17b8b8d0c8" namespace • [SLOW TEST:90.540 seconds] [sigs.k8s.io] Machines /data/src/sigs.k8s.io/cluster-api-provider-aws/vendor/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/framework/framework.go:373 AWS actuator /data/src/sigs.k8s.io/cluster-api-provider-aws/test/machines/machines_test.go:87 Can create EBS volumes /data/src/sigs.k8s.io/cluster-api-provider-aws/test/machines/machines_test.go:209 ------------------------------ [sigs.k8s.io] Machines AWS actuator Can deploy compute nodes through machineset /data/src/sigs.k8s.io/cluster-api-provider-aws/test/machines/machines_test.go:245 STEP: Creating "namespace-f927ab3c-cf20-11e9-b8bb-0a17b8b8d0c8" namespace STEP: Deploying cluster API stack components STEP: Deploying cluster CRD manifest I0904 14:33:38.761065 19288 framework.go:335] create.err: customresourcedefinitions.apiextensions.k8s.io "clusters.cluster.k8s.io" already exists STEP: Deploying machine CRD manifest I0904 14:33:43.767532 19288 framework.go:335] create.err: customresourcedefinitions.apiextensions.k8s.io "machines.machine.openshift.io" already exists STEP: Deploying machineset CRD manifest I0904 14:33:48.772600 19288 framework.go:335] create.err: customresourcedefinitions.apiextensions.k8s.io "machinesets.machine.openshift.io" already exists STEP: Deploying machinedeployment CRD manifest I0904 14:33:53.777074 19288 framework.go:335] create.err: customresourcedefinitions.apiextensions.k8s.io "machinedeployments.machine.openshift.io" already exists STEP: Deploying cluster role STEP: Deploying machine API controllers STEP: Waiting until cluster objects can be listed STEP: Cluster API stack deployed STEP: Creating "pr-975-258" cluster STEP: Creating "pr-975-258-master-machine-14021e" machine STEP: Waiting for "pr-975-258-master-machine-14021e" machine STEP: Verify machine's underlying instance is running I0904 14:34:33.812904 19288 machines.go:80] Waiting for instance to come up I0904 14:34:38.812810 19288 machines.go:80] Waiting for instance to come up I0904 14:34:43.812907 19288 machines.go:80] Waiting for instance to come up I0904 14:34:43.921258 19288 machines.go:88] Machine is running STEP: Collecting master kubeconfig I0904 14:34:53.988805 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:35:58.196795 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:35:58.988828 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:35:59.274100 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:36:03.988833 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:36:04.362786 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:36:08.988839 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:36:09.298005 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:36:13.988829 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:36:14.490539 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:36:18.988835 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:36:19.330856 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:36:23.988838 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:36:24.270604 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:36:28.988824 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:36:29.267045 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:36:33.988833 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:36:34.265165 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:36:38.988835 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:36:39.274587 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:36:43.988828 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:36:44.268054 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:36:48.988831 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:36:49.262034 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:36:53.988831 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:36:54.264279 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:36:58.988826 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:36:59.271960 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:37:03.988827 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:37:04.265329 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:37:08.988832 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:37:09.271657 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:37:13.988832 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:37:14.277223 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:37:18.988824 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:37:19.289286 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:37:23.988829 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:37:24.267509 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:37:28.988833 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:37:29.292202 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:37:33.988824 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:37:34.275699 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:37:38.988816 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:37:39.300533 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:37:43.988834 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:37:44.304042 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:37:48.988823 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:37:49.293622 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:37:53.988835 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:37:54.299392 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:37:58.988823 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:37:59.301154 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:38:03.988825 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:38:04.266026 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:38:08.988826 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:38:09.325494 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:38:13.988816 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:38:14.295616 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:38:18.988830 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:38:19.271894 19288 machines.go:324] Unable to pull kubeconfig: failed to collect kubeconfig: Process exited with status 1, cat: /root/.kube/config: No such file or directory I0904 14:38:23.988821 19288 machines.go:316] Pulling kubeconfig from ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:38:24.308095 19288 machines.go:334] Master running on https://ec2-35-174-241-123.compute-1.amazonaws.com:8443 I0904 14:38:24.447686 19288 framework.go:377] Uploading "origin-aws-machine-controllers:v0.2.0-173-gb8c169e" to the master machine under "ec2-35-174-241-123.compute-1.amazonaws.com" I0904 14:39:16.695938 19288 framework.go:391] Warning: Permanently added 'ec2-35-174-241-123.compute-1.amazonaws.com,35.174.241.123' (ECDSA) to the list of known hosts. Loaded image: origin-aws-machine-controllers:v0.2.0-173-gb8c169e STEP: Creating "namespace-f927ab3c-cf20-11e9-b8bb-0a17b8b8d0c8" namespace STEP: Deploying cluster API stack components STEP: Deploying cluster CRD manifest I0904 14:39:21.724900 19288 framework.go:335] create.err: <nil> I0904 14:39:21.728080 19288 framework.go:344] get.err: <nil> STEP: Deploying machine CRD manifest I0904 14:39:26.733147 19288 framework.go:335] create.err: <nil> I0904 14:39:26.737335 19288 framework.go:344] get.err: <nil> STEP: Deploying machineset CRD manifest I0904 14:39:31.741725 19288 framework.go:335] create.err: <nil> I0904 14:39:33.060902 19288 framework.go:344] get.err: <nil> STEP: Deploying machinedeployment CRD manifest I0904 14:39:38.065915 19288 framework.go:335] create.err: <nil> I0904 14:39:38.068614 19288 framework.go:344] get.err: <nil> STEP: Deploying cluster role STEP: Deploying machine API controllers STEP: Waiting until cluster objects can be listed STEP: Cluster API stack deployed STEP: Deploy worker nodes through machineset STEP: Creating "pr-975-258" cluster workerMachineSet: &v1beta1.MachineSet{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pr-975-258-worker-machineset-e15efa", GenerateName:"pr-975-258-worker-machine-e15efa-", Namespace:"namespace-f927ab3c-cf20-11e9-b8bb-0a17b8b8d0c8", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"machine.openshift.io/cluster-api-cluster":"pr-975-258"}, Annotations:map[string]string{"machine.openshift.io/exclude-node-draining":""}, OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1beta1.MachineSetSpec{Replicas:(*int32)(0xc4205b88b0), MinReadySeconds:0, DeletePolicy:"", Selector:v1.LabelSelector{MatchLabels:map[string]string{"machine.openshift.io/cluster-api-cluster":"pr-975-258", "machine.openshift.io/cluster-api-machineset":"pr-975-258-worker-machineset-e15efa"}, MatchExpressions:[]v1.LabelSelectorRequirement(nil)}, Template:v1beta1.MachineTemplateSpec{ObjectMeta:v1.ObjectMeta{Name:"", GenerateName:"pr-975-258-worker-machine-e15efa-", Namespace:"", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"machine.openshift.io/cluster-api-machineset":"pr-975-258-worker-machineset-e15efa", "machine.openshift.io/cluster-api-cluster":"pr-975-258"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1beta1.MachineSpec{ObjectMeta:v1.ObjectMeta{Name:"", GenerateName:"", Namespace:"", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"node-role.kubernetes.io/compute":""}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Taints:[]v1.Taint(nil), ProviderSpec:v1beta1.ProviderSpec{Value:(*runtime.RawExtension)(0xc42064e810)}, ProviderID:(*string)(nil)}}}, Status:v1beta1.MachineSetStatus{Replicas:0, FullyLabeledReplicas:0, ReadyReplicas:0, AvailableReplicas:0, ObservedGeneration:0, ErrorReason:(*common.MachineSetStatusError)(nil), ErrorMessage:(*string)(nil)}} STEP: Creating "pr-975-258-worker-machineset-e15efa" machineset STEP: Verify machineset's underlying instances is running STEP: Waiting for "pr-975-258-worker-machineset-e15efa-b84cz" machine STEP: Verify machine's underlying instance is running I0904 14:40:23.365227 19288 machines.go:80] Waiting for instance to come up I0904 14:40:28.365284 19288 machines.go:80] Waiting for instance to come up I0904 14:40:28.508734 19288 machines.go:88] Machine is running STEP: Checking master and worker nodes are ready I0904 14:40:33.513113 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:40:38.514769 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:40:43.511611 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:40:48.512142 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:40:53.513058 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:40:58.512039 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:41:03.511674 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:41:08.511982 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:41:13.511652 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:41:18.511575 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:41:23.511877 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:41:28.512082 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:41:33.511994 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:41:38.511929 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:41:43.512141 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:41:48.511984 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:41:53.511956 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:41:58.512149 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:42:03.511837 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:42:08.511926 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:42:13.511779 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:42:18.511693 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:42:23.511873 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:42:28.512080 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:42:33.511924 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:42:38.511961 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:42:43.512021 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:42:48.511990 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:42:53.511794 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:42:58.512078 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:43:03.511576 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:43:08.512157 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:43:13.511599 19288 machines.go:226] Waiting for 2 nodes to come up, have 1 I0904 14:43:18.512057 19288 machines.go:239] Node "ip-10-0-101-140.ec2.internal" is ready I0904 14:43:18.512086 19288 machines.go:236] Node "ip-10-0-101-214.ec2.internal" not ready I0904 14:43:23.512552 19288 machines.go:239] Node "ip-10-0-101-140.ec2.internal" is ready I0904 14:43:23.512578 19288 machines.go:236] Node "ip-10-0-101-214.ec2.internal" not ready I0904 14:43:28.512295 19288 machines.go:239] Node "ip-10-0-101-140.ec2.internal" is ready I0904 14:43:28.512327 19288 machines.go:239] Node "ip-10-0-101-214.ec2.internal" is ready STEP: Checking compute node role and node linking time="2019-09-04T14:43:33Z" level=info msg="node \"ip-10-0-101-214.ec2.internal\" role set to 'node-role.kubernetes.io/compute'" source="machines_test.go:374" time="2019-09-04T14:43:33Z" level=info msg="node \"ip-10-0-101-214.ec2.internal\" is linked with \"pr-975-258-worker-machineset-e15efa-b84cz\" machine" source="machines_test.go:385" STEP: Destroying worker machines STEP: Get all "pr-975-258-worker-machineset-e15efa" machineset's machines STEP: Deleting "pr-975-258-worker-machineset-e15efa" machineset STEP: Waiting for all machines to be deleted STEP: Verify machine's underlying instance is not running I0904 14:43:38.527062 19288 machines.go:102] Waiting for instance to terminate STEP: Waiting for "pr-975-258-worker-machineset-e15efa-b84cz" machine object to be deleted I0904 14:53:38.675509 19288 machines.go:130] unable to wait for machine to get deleted: timed out waiting for the condition STEP: Destroying master machine STEP: Deleting "pr-975-258-master-machine-14021e" machine STEP: Verify instance is terminated I0904 14:53:53.685916 19288 machines.go:50] Waiting for instance to be terminated STEP: Get all "pr-975-258-worker-machineset-e15efa" machineset's machines STEP: Deleting "pr-975-258-master-machine-14021e" machine STEP: Verify instance is terminated I0904 14:54:33.800196 19288 machines.go:50] Waiting for instance to be terminated STEP: Deleting machine API controllers I0904 14:54:38.892675 19288 framework.go:311] del.err: <nil> I0904 14:54:38.897795 19288 framework.go:323] get.err: deployments.apps "clusterapi-controllers" not found STEP: Deleting cluster role I0904 14:54:43.901900 19288 framework.go:311] del.err: <nil> I0904 14:54:43.903239 19288 framework.go:323] get.err: clusterrolebindings.rbac.authorization.k8s.io "manager-rolebinding" not found I0904 14:54:48.907315 19288 framework.go:311] del.err: <nil> I0904 14:54:48.908805 19288 framework.go:323] get.err: clusterroles.rbac.authorization.k8s.io "manager-role" not found time="2019-09-04T14:54:48Z" level=info msg="namespace-f927ab3c-cf20-11e9-b8bb-0a17b8b8d0c8: &v1.Namespace{TypeMeta:v1.TypeMeta{Kind:\"\", APIVersion:\"\"}, ObjectMeta:v1.ObjectMeta{Name:\"namespace-f927ab3c-cf20-11e9-b8bb-0a17b8b8d0c8\", GenerateName:\"\", Namespace:\"\", SelfLink:\"\", UID:\"\", ResourceVersion:\"\", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:\"\", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.NamespaceSpec{Finalizers:[]v1.FinalizerName(nil)}, Status:v1.NamespaceStatus{Phase:\"\"}}" source="machines_test.go:73" STEP: Destroying "namespace-f927ab3c-cf20-11e9-b8bb-0a17b8b8d0c8" namespace • [SLOW TEST:1275.158 seconds] [sigs.k8s.io] Machines /data/src/sigs.k8s.io/cluster-api-provider-aws/vendor/github.com/openshift/cluster-api-actuator-pkg/pkg/e2e/framework/framework.go:373 AWS actuator /data/src/sigs.k8s.io/cluster-api-provider-aws/test/machines/machines_test.go:87 Can deploy compute nodes through machineset /data/src/sigs.k8s.io/cluster-api-provider-aws/test/machines/machines_test.go:245 ------------------------------ Ran 3 of 3 Specs in 1470.614 seconds SUCCESS! -- 3 Passed | 0 Failed | 0 Pending | 0 Skipped --- PASS: TestCart (1470.61s) PASS ok sigs.k8s.io/cluster-api-provider-aws/test/machines 1470.659s The aws instance was running and terminated + echo 'The aws instance was running and terminated' + set +o xtrace ########## FINISHED STAGE: SUCCESS: RUN E2E TESTS [00h 25m 22s] ########## [PostBuildScript] - Executing post build scripts. [workspace] $ /bin/bash /tmp/jenkins8102904811257147110.sh ########## STARTING STAGE: DOWNLOAD ARTIFACTS FROM THE REMOTE HOST ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config + trap 'exit 0' EXIT ++ pwd + ARTIFACT_DIR=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/gathered + rm -rf /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/gathered + mkdir -p /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/gathered + tree /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/gathered /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/gathered 0 directories, 0 files + exit 0 [workspace] $ /bin/bash /tmp/jenkins4343081540019051232.sh ########## STARTING STAGE: GENERATE ARTIFACTS FROM THE REMOTE HOST ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config + trap 'exit 0' EXIT ++ pwd + ARTIFACT_DIR=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/generated + rm -rf /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/generated + mkdir /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/generated + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo docker version && sudo docker info && sudo docker images && sudo docker ps -a 2>&1' WARNING: You're not using the default seccomp profile + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo cat /etc/sysconfig/docker /etc/sysconfig/docker-network /etc/sysconfig/docker-storage /etc/sysconfig/docker-storage-setup /etc/systemd/system/docker.service 2>&1' + true + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo find /var/lib/docker/containers -name *.log | sudo xargs tail -vn +1 2>&1' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo ausearch -m AVC -m SELINUX_ERR -m USER_AVC 2>&1' + true + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo df -T -h && sudo pvs && sudo vgs && sudo lvs && sudo findmnt --all 2>&1' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo yum list installed 2>&1' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl --dmesg --no-pager --all --lines=all 2>&1' + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel 'sudo journalctl _PID=1 --no-pager --all --lines=all 2>&1' + tree /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/generated /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/generated ├── avc_denials.log ├── containers.log ├── dmesg.log ├── docker.config ├── docker.info ├── filesystem.info ├── installed_packages.log └── pid1.journal 0 directories, 8 files + exit 0 [workspace] $ /bin/bash /tmp/jenkins2799588782348008507.sh ########## STARTING STAGE: FETCH SYSTEMD JOURNALS FROM THE REMOTE HOST ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config + trap 'exit 0' EXIT ++ pwd + ARTIFACT_DIR=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/journals + rm -rf /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/journals + mkdir /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/journals + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit docker.service --no-pager --all --lines=all + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit dnsmasq.service --no-pager --all --lines=all + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config openshiftdevel sudo journalctl --unit systemd-journald.service --no-pager --all --lines=all + tree /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/journals /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/artifacts/journals ├── dnsmasq.service ├── docker.service └── systemd-journald.service 0 directories, 3 files + exit 0 [workspace] $ /bin/bash /tmp/jenkins1064271138803010849.sh ########## STARTING STAGE: DEPROVISION RESOURCES ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ mktemp + script=/tmp/tmp.TDo2nlUM3V + cat + chmod +x /tmp/tmp.TDo2nlUM3V + scp -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.TDo2nlUM3V openshiftdevel:/tmp/tmp.TDo2nlUM3V + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 14400 /tmp/tmp.TDo2nlUM3V"' + cd /home/origin + cd /data/src/github.com/openshift/cluster-api-provider-aws/hack + [[ -e envs ]] + set +x + AWS_DEFAULT_REGION=us-east-1 + ./aws-provision.sh destroy /data/src/github.com/openshift/cluster-api-provider-aws/hack/prebuild /data/src/github.com/openshift/cluster-api-provider-aws/hack { "TerminatingInstances": [ { "InstanceId": "i-08a574885b8679b39", "CurrentState": { "Code": 32, "Name": "shutting-down" }, "PreviousState": { "Code": 16, "Name": "running" } } ] } data.aws_region.current: Refreshing state... aws_iam_role.role: Refreshing state... (ID: pr-975-258-role) data.aws_availability_zones.azs: Refreshing state... aws_vpc.this: Refreshing state... (ID: vpc-01b6178fc04d23888) aws_iam_instance_profile.test_profile: Refreshing state... (ID: pr-975-258-worker-profile) aws_security_group.cluster_default: Refreshing state... (ID: sg-03e2fd2a04defa788) aws_route_table.public: Refreshing state... (ID: rtb-02d9caddda23f272f) aws_route_table.private[1]: Refreshing state... (ID: rtb-0cc4d8ad28295d807) aws_route_table.private[2]: Refreshing state... (ID: rtb-0eb95634bb4bdf242) aws_route_table.private[0]: Refreshing state... (ID: rtb-01c859705a5683857) aws_subnet.public[1]: Refreshing state... (ID: subnet-0df88ffea33bee0de) aws_subnet.public[0]: Refreshing state... (ID: subnet-079ba6af91da7e6e2) aws_subnet.private[1]: Refreshing state... (ID: subnet-0adce266c18f72255) aws_subnet.public[2]: Refreshing state... (ID: subnet-0508530a7b553cace) aws_subnet.private[0]: Refreshing state... (ID: subnet-0527838d60f1601f6) aws_subnet.private[2]: Refreshing state... (ID: subnet-078025a95b8289428) aws_internet_gateway.this: Refreshing state... (ID: igw-0bcc735e9b9964b0d) aws_security_group_rule.default_ingress: Refreshing state... (ID: sgrule-1217565267) aws_security_group_rule.default_egress: Refreshing state... (ID: sgrule-712216049) aws_security_group_rule.allow_all: Refreshing state... (ID: sgrule-2442055330) aws_route_table_association.public[1]: Refreshing state... (ID: rtbassoc-08e10268bb436cde3) aws_route_table_association.public[0]: Refreshing state... (ID: rtbassoc-0c286cad3e587bdd3) aws_route_table_association.public[2]: Refreshing state... (ID: rtbassoc-07032d65a9f137ac4) aws_route.public_internet_gateway: Refreshing state... (ID: r-rtb-02d9caddda23f272f1080289494) aws_route_table_association.private[1]: Refreshing state... (ID: rtbassoc-0c6024cd954c92135) aws_route_table_association.private[0]: Refreshing state... (ID: rtbassoc-072092ddd4cbe1d4e) aws_route_table_association.private[2]: Refreshing state... (ID: rtbassoc-0a1ec90e46fdec4b8) aws_security_group_rule.default_egress: Destroying... (ID: sgrule-712216049) module.vpc.aws_route_table_association.public[2]: Destroying... (ID: rtbassoc-07032d65a9f137ac4) aws_iam_instance_profile.test_profile: Destroying... (ID: pr-975-258-worker-profile) module.vpc.aws_route_table_association.private[1]: Destroying... (ID: rtbassoc-0c6024cd954c92135) module.vpc.aws_route_table_association.public[1]: Destroying... (ID: rtbassoc-08e10268bb436cde3) module.vpc.aws_route.public_internet_gateway: Destroying... (ID: r-rtb-02d9caddda23f272f1080289494) module.vpc.aws_route_table_association.private[2]: Destroying... (ID: rtbassoc-0a1ec90e46fdec4b8) aws_security_group_rule.allow_all: Destroying... (ID: sgrule-2442055330) module.vpc.aws_route_table_association.public[0]: Destroying... (ID: rtbassoc-0c286cad3e587bdd3) aws_security_group_rule.default_ingress: Destroying... (ID: sgrule-1217565267) module.vpc.aws_route_table_association.public[0]: Destruction complete after 1s module.vpc.aws_route_table_association.private[1]: Destruction complete after 1s aws_iam_instance_profile.test_profile: Destruction complete after 1s module.vpc.aws_route_table_association.private[0]: Destroying... (ID: rtbassoc-072092ddd4cbe1d4e) module.vpc.aws_route_table_association.public[2]: Destruction complete after 1s module.vpc.aws_route_table_association.private[2]: Destruction complete after 1s aws_iam_role.role: Destroying... (ID: pr-975-258-role) aws_security_group_rule.default_egress: Destruction complete after 1s module.vpc.aws_route_table_association.public[1]: Destruction complete after 1s module.vpc.aws_route.public_internet_gateway: Destruction complete after 1s module.vpc.aws_subnet.public[1]: Destroying... (ID: subnet-0df88ffea33bee0de) module.vpc.aws_subnet.public[2]: Destroying... (ID: subnet-0508530a7b553cace) module.vpc.aws_subnet.public[0]: Destroying... (ID: subnet-079ba6af91da7e6e2) module.vpc.aws_route_table.public: Destroying... (ID: rtb-02d9caddda23f272f) module.vpc.aws_internet_gateway.this: Destroying... (ID: igw-0bcc735e9b9964b0d) module.vpc.aws_route_table_association.private[0]: Destruction complete after 0s module.vpc.aws_route_table.private[2]: Destroying... (ID: rtb-0eb95634bb4bdf242) module.vpc.aws_subnet.private[1]: Destroying... (ID: subnet-0adce266c18f72255) aws_iam_role.role: Destruction complete after 0s module.vpc.aws_route_table.private[0]: Destroying... (ID: rtb-01c859705a5683857) aws_security_group_rule.allow_all: Destruction complete after 1s module.vpc.aws_subnet.private[2]: Destroying... (ID: subnet-078025a95b8289428) module.vpc.aws_route_table.public: Destruction complete after 0s module.vpc.aws_route_table.private[1]: Destroying... (ID: rtb-0cc4d8ad28295d807) module.vpc.aws_route_table.private[2]: Destruction complete after 0s module.vpc.aws_subnet.private[0]: Destroying... (ID: subnet-0527838d60f1601f6) aws_security_group_rule.default_ingress: Destruction complete after 1s aws_security_group.cluster_default: Destroying... (ID: sg-03e2fd2a04defa788) module.vpc.aws_route_table.private[0]: Destruction complete after 0s module.vpc.aws_route_table.private[1]: Destruction complete after 0s module.vpc.aws_subnet.public[1]: Destruction complete after 0s module.vpc.aws_subnet.private[1]: Destruction complete after 0s module.vpc.aws_subnet.public[2]: Destruction complete after 0s module.vpc.aws_subnet.private[2]: Destruction complete after 0s module.vpc.aws_subnet.private[0]: Destruction complete after 1s module.vpc.aws_subnet.public.0: Still destroying... (ID: subnet-079ba6af91da7e6e2, 10s elapsed) module.vpc.aws_internet_gateway.this: Still destroying... (ID: igw-0bcc735e9b9964b0d, 10s elapsed) aws_security_group.cluster_default: Still destroying... (ID: sg-03e2fd2a04defa788, 10s elapsed) module.vpc.aws_internet_gateway.this: Destruction complete after 17s module.vpc.aws_subnet.public.0: Still destroying... (ID: subnet-079ba6af91da7e6e2, 20s elapsed) aws_security_group.cluster_default: Still destroying... (ID: sg-03e2fd2a04defa788, 20s elapsed) module.vpc.aws_subnet.public[0]: Destruction complete after 26s aws_security_group.cluster_default: Destruction complete after 26s module.vpc.aws_vpc.this: Destroying... (ID: vpc-01b6178fc04d23888) module.vpc.aws_vpc.this: Destruction complete after 1s  Destroy complete! Resources: 25 destroyed. + set +o xtrace ########## FINISHED STAGE: SUCCESS: DEPROVISION RESOURCES [00h 00m 33s] ########## [workspace] $ /bin/bash /tmp/jenkins3086990846189927780.sh ########## STARTING STAGE: ASSEMBLE GCS OUTPUT ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config + trap 'exit 0' EXIT + mkdir -p gcs/artifacts gcs/artifacts/generated gcs/artifacts/journals gcs/artifacts/gathered ++ python -c 'import json; import urllib; print json.load(urllib.urlopen('\''https://ci.openshift.redhat.com/jenkins/job/ci-kubernetes-aws-actuator/975/api/json'\''))['\''result'\'']' + result=SUCCESS + cat ++ date +%s + cat /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/builds/975/log + cp artifacts/generated/avc_denials.log artifacts/generated/containers.log artifacts/generated/dmesg.log artifacts/generated/docker.config artifacts/generated/docker.info artifacts/generated/filesystem.info artifacts/generated/installed_packages.log artifacts/generated/pid1.journal gcs/artifacts/generated/ + cp artifacts/journals/dnsmasq.service artifacts/journals/docker.service artifacts/journals/systemd-journald.service gcs/artifacts/journals/ ++ pwd + scp -F ./.config/origin-ci-tool/inventory/.ssh_config -r /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/gcs openshiftdevel:/data + scp -F ./.config/origin-ci-tool/inventory/.ssh_config /var/lib/jenkins/.config/gcloud/gcs-publisher-credentials.json openshiftdevel:/data/credentials.json + exit 0 [workspace] $ /bin/bash /tmp/jenkins8003300154911593544.sh ########## STARTING STAGE: PUSH THE ARTIFACTS AND METADATA ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ mktemp + script=/tmp/tmp.6t12DFh6xl + cat + chmod +x /tmp/tmp.6t12DFh6xl + scp -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config /tmp/tmp.6t12DFh6xl openshiftdevel:/tmp/tmp.6t12DFh6xl + ssh -F /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/.ssh_config -t openshiftdevel 'bash -l -c "timeout 300 /tmp/tmp.6t12DFh6xl"' + cd /home/origin + trap 'exit 0' EXIT + [[ -n {"type":"presubmit","job":"ci-kubernetes-aws-actuator","buildid":"1169254337591906304","prowjobid":"54633cdd-cf1f-11e9-ab71-0a58ac108d31","refs":{"org":"openshift","repo":"cluster-api-provider-aws","repo_link":"https://github.com/openshift/cluster-api-provider-aws","base_ref":"master","base_sha":"d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","base_link":"https://github.com/openshift/cluster-api-provider-aws/commit/d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","pulls":[{"number":258,"author":"ingvagabund","sha":"6fc8f0a260389bc1798d5bc6fea49264a324669f","link":"https://github.com/openshift/cluster-api-provider-aws/pull/258","commit_link":"https://github.com/openshift/cluster-api-provider-aws/pull/258/commits/6fc8f0a260389bc1798d5bc6fea49264a324669f","author_link":"https://github.com/ingvagabund"}]}} ]] ++ jq --compact-output '.buildid |= "975"' + JOB_SPEC='{"type":"presubmit","job":"ci-kubernetes-aws-actuator","buildid":"975","prowjobid":"54633cdd-cf1f-11e9-ab71-0a58ac108d31","refs":{"org":"openshift","repo":"cluster-api-provider-aws","repo_link":"https://github.com/openshift/cluster-api-provider-aws","base_ref":"master","base_sha":"d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","base_link":"https://github.com/openshift/cluster-api-provider-aws/commit/d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","pulls":[{"number":258,"author":"ingvagabund","sha":"6fc8f0a260389bc1798d5bc6fea49264a324669f","link":"https://github.com/openshift/cluster-api-provider-aws/pull/258","commit_link":"https://github.com/openshift/cluster-api-provider-aws/pull/258/commits/6fc8f0a260389bc1798d5bc6fea49264a324669f","author_link":"https://github.com/ingvagabund"}]}}' + sudo docker run -e 'JOB_SPEC={"type":"presubmit","job":"ci-kubernetes-aws-actuator","buildid":"975","prowjobid":"54633cdd-cf1f-11e9-ab71-0a58ac108d31","refs":{"org":"openshift","repo":"cluster-api-provider-aws","repo_link":"https://github.com/openshift/cluster-api-provider-aws","base_ref":"master","base_sha":"d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","base_link":"https://github.com/openshift/cluster-api-provider-aws/commit/d4644b21c9dbeeb79215ad42e8d7f932dd3a5f41","pulls":[{"number":258,"author":"ingvagabund","sha":"6fc8f0a260389bc1798d5bc6fea49264a324669f","link":"https://github.com/openshift/cluster-api-provider-aws/pull/258","commit_link":"https://github.com/openshift/cluster-api-provider-aws/pull/258/commits/6fc8f0a260389bc1798d5bc6fea49264a324669f","author_link":"https://github.com/ingvagabund"}]}}' -v /data:/data:z registry.svc.ci.openshift.org/ci/gcsupload:latest --dry-run=false --gcs-path=gs://origin-ci-test --gcs-credentials-file=/data/credentials.json /data/gcs/artifacts /data/gcs/build-log.txt /data/gcs/finished.json Unable to find image 'registry.svc.ci.openshift.org/ci/gcsupload:latest' locally Trying to pull repository registry.svc.ci.openshift.org/ci/gcsupload ... latest: Pulling from registry.svc.ci.openshift.org/ci/gcsupload a073c86ecf9e: Already exists cc3fc741b1a9: Already exists 822bed51ba40: Pulling fs layer 85cea451eec0: Pulling fs layer 85cea451eec0: Verifying Checksum 85cea451eec0: Download complete 822bed51ba40: Download complete 822bed51ba40: Pull complete 85cea451eec0: Pull complete Digest: sha256:03aad50d7ec631ee07c12ac2ba679bd48c7781f7d5754f9e0dcc4e7260e35208 Status: Downloaded newer image for registry.svc.ci.openshift.org/ci/gcsupload:latest {"component":"gcsupload","file":"prow/gcsupload/run.go:166","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts","level":"info","msg":"Gathering artifacts from artifact directory: /data/gcs/artifacts","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","file":"prow/gcsupload/run.go:184","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/avc_denials.log in artifact directory. Uploading as pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/avc_denials.log\n","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","file":"prow/gcsupload/run.go:184","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/containers.log in artifact directory. Uploading as pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/containers.log\n","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","file":"prow/gcsupload/run.go:184","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/dmesg.log in artifact directory. Uploading as pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/dmesg.log\n","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","file":"prow/gcsupload/run.go:184","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.config in artifact directory. Uploading as pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/docker.config\n","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","file":"prow/gcsupload/run.go:184","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/docker.info in artifact directory. Uploading as pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/docker.info\n","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","file":"prow/gcsupload/run.go:184","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/filesystem.info in artifact directory. Uploading as pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/filesystem.info\n","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","file":"prow/gcsupload/run.go:184","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/installed_packages.log in artifact directory. Uploading as pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/installed_packages.log\n","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","file":"prow/gcsupload/run.go:184","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/generated/pid1.journal in artifact directory. Uploading as pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/pid1.journal\n","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","file":"prow/gcsupload/run.go:184","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/journals/dnsmasq.service in artifact directory. Uploading as pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/journals/dnsmasq.service\n","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","file":"prow/gcsupload/run.go:184","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/journals/docker.service in artifact directory. Uploading as pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/journals/docker.service\n","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","file":"prow/gcsupload/run.go:184","func":"k8s.io/test-infra/prow/gcsupload.gatherArtifacts.func1","level":"info","msg":"Found /data/gcs/artifacts/journals/systemd-journald.service in artifact directory. Uploading as pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/journals/systemd-journald.service\n","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/containers.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/dmesg.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/docker.config","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/installed_packages.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/finished.json","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/avc_denials.log","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/docker.info","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/journals/dnsmasq.service","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/directory/ci-kubernetes-aws-actuator/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/journals/systemd-journald.service","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/directory/ci-kubernetes-aws-actuator/975.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/filesystem.info","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/pid1.journal","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/journals/docker.service","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/build-log.txt","file":"prow/pod-utils/gcs/upload.go:64","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload","level":"info","msg":"Queued for upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/finished.json","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/installed_packages.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/docker.config","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/journals/systemd-journald.service","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/filesystem.info","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/docker.info","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/directory/ci-kubernetes-aws-actuator/975.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/directory/ci-kubernetes-aws-actuator/latest-build.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/avc_denials.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/journals/dnsmasq.service","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/containers.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/dmesg.log","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/generated/pid1.journal","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/build-log.txt","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:51Z"} {"component":"gcsupload","dest":"pr-logs/pull/openshift_cluster-api-provider-aws/258/ci-kubernetes-aws-actuator/975/artifacts/journals/docker.service","file":"prow/pod-utils/gcs/upload.go:70","func":"k8s.io/test-infra/prow/pod-utils/gcs.upload.func1","level":"info","msg":"Finished upload","time":"2019-09-04T14:55:52Z"} {"component":"gcsupload","file":"prow/gcsupload/run.go:65","func":"k8s.io/test-infra/prow/gcsupload.Options.Run","level":"info","msg":"Finished upload to GCS","time":"2019-09-04T14:55:52Z"} + exit 0 + set +o xtrace ########## FINISHED STAGE: SUCCESS: PUSH THE ARTIFACTS AND METADATA [00h 00m 06s] ########## [workspace] $ /bin/bash /tmp/jenkins8404865748238474611.sh ########## STARTING STAGE: DEPROVISION CLOUD RESOURCES ########## + [[ -s /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ]] + source /var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/activate ++ export VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ VIRTUAL_ENV=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed ++ export PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ PATH=/var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/bin:/sbin:/usr/sbin:/bin:/usr/bin ++ unset PYTHON_HOME ++ export OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config ++ OCT_CONFIG_HOME=/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config + oct deprovision PLAYBOOK: main.yml ************************************************************* 4 plays in /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml PLAY [ensure we have the parameters necessary to deprovision virtual hosts] **** TASK [ensure all required variables are set] *********************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:9 skipping: [localhost] => (item=origin_ci_inventory_dir) => { "changed": false, "generated_timestamp": "2019-09-04 10:55:53.351956", "item": "origin_ci_inventory_dir", "skip_reason": "Conditional check failed", "skipped": true } skipping: [localhost] => (item=origin_ci_aws_region) => { "changed": false, "generated_timestamp": "2019-09-04 10:55:53.356553", "item": "origin_ci_aws_region", "skip_reason": "Conditional check failed", "skipped": true } PLAY [deprovision virtual hosts in EC2] **************************************** TASK [Gathering Facts] ********************************************************* ok: [localhost] TASK [deprovision a virtual EC2 host] ****************************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:28 included: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml for localhost TASK [update the SSH configuration to remove AWS EC2 specifics] **************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:2 ok: [localhost] => { "changed": false, "generated_timestamp": "2019-09-04 10:55:54.258798", "msg": "" } TASK [rename EC2 instance for termination reaper] ****************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:8 changed: [localhost] => { "changed": true, "generated_timestamp": "2019-09-04 10:55:54.943632", "msg": "Tags {'Name': 'oct-terminate'} created for resource i-00cad1e0a8c636632." } TASK [tear down the EC2 instance] ********************************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:15 changed: [localhost] => { "changed": true, "generated_timestamp": "2019-09-04 10:55:55.844880", "instance_ids": [ "i-00cad1e0a8c636632" ], "instances": [ { "ami_launch_index": "0", "architecture": "x86_64", "block_device_mapping": { "/dev/sda1": { "delete_on_termination": true, "status": "attached", "volume_id": "vol-097be3317dc12d72a" }, "/dev/sdb": { "delete_on_termination": true, "status": "attached", "volume_id": "vol-02974abdb2ed88770" } }, "dns_name": "ec2-54-82-177-53.compute-1.amazonaws.com", "ebs_optimized": false, "groups": { "sg-7e73221a": "default" }, "hypervisor": "xen", "id": "i-00cad1e0a8c636632", "image_id": "ami-0b77b87a37c3e662c", "instance_type": "m4.xlarge", "kernel": null, "key_name": "libra", "launch_time": "2019-09-04T14:22:33.000Z", "placement": "us-east-1c", "private_dns_name": "ip-172-18-18-57.ec2.internal", "private_ip": "172.18.18.57", "public_dns_name": "ec2-54-82-177-53.compute-1.amazonaws.com", "public_ip": "54.82.177.53", "ramdisk": null, "region": "us-east-1", "root_device_name": "/dev/sda1", "root_device_type": "ebs", "state": "running", "state_code": 16, "tags": { "Name": "oct-terminate", "openshift_etcd": "", "openshift_master": "", "openshift_node": "" }, "tenancy": "default", "virtualization_type": "hvm" } ], "tagged_instances": [] } TASK [remove the serialized host variables] ************************************ task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/roles/aws-down/tasks/main.yml:22 changed: [localhost] => { "changed": true, "generated_timestamp": "2019-09-04 10:55:56.104093", "path": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory/host_vars/172.18.18.57.yml", "state": "absent" } PLAY [deprovision virtual hosts locally manged by Vagrant] ********************* TASK [Gathering Facts] ********************************************************* ok: [localhost] PLAY [clean up local configuration for deprovisioned instances] **************** TASK [remove inventory configuration directory] ******************************** task path: /var/lib/jenkins/origin-ci-tool/2b40f3e11aadb569dc9c0c9fb90e7273658ce6ed/lib/python2.7/site-packages/oct/ansible/oct/playbooks/deprovision/main.yml:61 changed: [localhost] => { "changed": true, "generated_timestamp": "2019-09-04 10:55:56.646193", "path": "/var/lib/jenkins/jobs/ci-kubernetes-aws-actuator/workspace/.config/origin-ci-tool/inventory", "state": "absent" } PLAY RECAP ********************************************************************* localhost : ok=8 changed=4 unreachable=0 failed=0 + set +o xtrace ########## FINISHED STAGE: SUCCESS: DEPROVISION CLOUD RESOURCES [00h 00m 04s] ########## Archiving artifacts Recording test results [WS-CLEANUP] Deleting project workspace... [WS-CLEANUP] Deferred wipeout is used... [WS-CLEANUP] done Finished: SUCCESS