Console Output
Started by upstream project "test_pull_request_origin_aggregated_logging" build number 85
originally caused by:
Started by remote host 50.17.198.52
[EnvInject] - Loading node environment variables.
Building in workspace /var/lib/jenkins/jobs/test-origin-aggregated-logging/workspace
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties content
OS_ROOT=/data/src/github.com/openshift/origin
INSTANCE_TYPE=c4.xlarge
GITHUB_REPO=openshift
OS=rhel7
TESTNAME=logging
[EnvInject] - Variables injected successfully.
[workspace] $ /bin/sh -xe /tmp/hudson8039332160490712258.sh
+ false
+ unset GOPATH
+ REPO_NAME=origin-aggregated-logging
+ rm -rf origin-aggregated-logging
+ vagrant origin-local-checkout --replace --repo origin-aggregated-logging -b master
You don't seem to have the GOPATH environment variable set on your system.
See: 'go help gopath' for more details about GOPATH.
Waiting for the cloning process to finish
Cloning origin-aggregated-logging ...
Submodule 'deployer/common' (https://github.com/openshift/origin-integration-common) registered for path 'deployer/common'
Submodule 'kibana-proxy' (https://github.com/fabric8io/openshift-auth-proxy.git) registered for path 'kibana-proxy'
Cloning into 'deployer/common'...
Submodule path 'deployer/common': checked out '45bf993212cdcbab5cbce3b3fab74a72b851402e'
Cloning into 'kibana-proxy'...
Submodule path 'kibana-proxy': checked out '118dfb40f7a8082d370ba7f4805255c9ec7c8178'
Origin repositories cloned into /var/lib/jenkins/jobs/test-origin-aggregated-logging/workspace
+ pushd origin-aggregated-logging
~/jobs/test-origin-aggregated-logging/workspace/origin-aggregated-logging ~/jobs/test-origin-aggregated-logging/workspace
+ git checkout master
Already on 'master'
+ popd
~/jobs/test-origin-aggregated-logging/workspace
+ '[' -n '' ']'
+ vagrant origin-local-checkout --replace
You don't seem to have the GOPATH environment variable set on your system.
See: 'go help gopath' for more details about GOPATH.
Waiting for the cloning process to finish
Checking repo integrity for /var/lib/jenkins/jobs/test-origin-aggregated-logging/workspace/origin
~/jobs/test-origin-aggregated-logging/workspace/origin ~/jobs/test-origin-aggregated-logging/workspace
# On branch master
# Untracked files:
# (use "git add <file>..." to include in what will be committed)
#
# artifacts/
nothing added to commit but untracked files present (use "git add" to track)
~/jobs/test-origin-aggregated-logging/workspace
Replacing: /var/lib/jenkins/jobs/test-origin-aggregated-logging/workspace/origin
~/jobs/test-origin-aggregated-logging/workspace/origin ~/jobs/test-origin-aggregated-logging/workspace
Already on 'master'
HEAD is now at 596d795 Merge pull request #14507 from stevekuznetsov/skuznets/update-script-dir
Removing .vagrant-openshift.json
Removing .vagrant/
Removing artifacts/
fatal: branch name required
~/jobs/test-origin-aggregated-logging/workspace
Origin repositories cloned into /var/lib/jenkins/jobs/test-origin-aggregated-logging/workspace
+ pushd origin
~/jobs/test-origin-aggregated-logging/workspace/origin ~/jobs/test-origin-aggregated-logging/workspace
+ INSTANCE_NAME=origin_logging-rhel7-1641
+ GIT_URL=https://github.com/openshift/origin-aggregated-logging
++ echo https://github.com/openshift/origin-aggregated-logging
++ sed s,https://,,
+ OAL_LOCAL_PATH=github.com/openshift/origin-aggregated-logging
+ OS_O_A_L_DIR=/data/src/github.com/openshift/origin-aggregated-logging
+ sort
+ env
_=/bin/env
BRANCH=master
BUILD_CAUSE=UPSTREAMTRIGGER
BUILD_CAUSE_UPSTREAMTRIGGER=true
BUILD_DISPLAY_NAME=#1641
BUILD_ID=1641
BUILD_NUMBER=1641
BUILD_TAG=jenkins-test-origin-aggregated-logging-1641
BUILD_URL=https://ci.openshift.redhat.com/jenkins/job/test-origin-aggregated-logging/1641/
EXECUTOR_NUMBER=61
GITHUB_REPO=openshift
HOME=/var/lib/jenkins
HUDSON_COOKIE=699668d3-18e8-48d8-b989-c837c3259599
HUDSON_HOME=/var/lib/jenkins
HUDSON_SERVER_COOKIE=ec11f8b2841c966f
HUDSON_URL=https://ci.openshift.redhat.com/jenkins/
INSTANCE_TYPE=c4.xlarge
JENKINS_HOME=/var/lib/jenkins
JENKINS_SERVER_COOKIE=ec11f8b2841c966f
JENKINS_URL=https://ci.openshift.redhat.com/jenkins/
JOB_BASE_NAME=test-origin-aggregated-logging
JOB_DISPLAY_URL=https://ci.openshift.redhat.com/jenkins/job/test-origin-aggregated-logging/display/redirect
JOB_NAME=test-origin-aggregated-logging
JOB_URL=https://ci.openshift.redhat.com/jenkins/job/test-origin-aggregated-logging/
LANG=en_US.UTF-8
LOGNAME=jenkins
MERGE=false
MERGE_SEVERITY=none
NLSPATH=/usr/dt/lib/nls/msg/%L/%N.cat
NODE_LABELS=master
NODE_NAME=master
OLDPWD=/var/lib/jenkins/jobs/test-origin-aggregated-logging/workspace
OPENSHIFT_ANSIBLE_TARGET_BRANCH=master
ORIGIN_AGGREGATED_LOGGING_PULL_ID=461
ORIGIN_AGGREGATED_LOGGING_TARGET_BRANCH=master
OS_ANSIBLE_BRANCH=master
OS_ANSIBLE_REPO=https://github.com/openshift/openshift-ansible
OS=rhel7
OS_ROOT=/data/src/github.com/openshift/origin
PATH=/sbin:/usr/sbin:/bin:/usr/bin
PWD=/var/lib/jenkins/jobs/test-origin-aggregated-logging/workspace/origin
ROOT_BUILD_CAUSE=REMOTECAUSE
ROOT_BUILD_CAUSE_REMOTECAUSE=true
RUN_CHANGES_DISPLAY_URL=https://ci.openshift.redhat.com/jenkins/job/test-origin-aggregated-logging/1641/display/redirect?page=changes
RUN_DISPLAY_URL=https://ci.openshift.redhat.com/jenkins/job/test-origin-aggregated-logging/1641/display/redirect
SHELL=/bin/bash
SHLVL=3
TESTNAME=logging
TEST_PERF=false
USER=jenkins
WORKSPACE=/var/lib/jenkins/jobs/test-origin-aggregated-logging/workspace
XFILESEARCHPATH=/usr/dt/app-defaults/%L/Dt
+ vagrant origin-init --stage inst --os rhel7 --instance-type c4.xlarge origin_logging-rhel7-1641
Reading AWS credentials from /var/lib/jenkins/.awscred
Searching devenv-rhel7_* for latest base AMI (required_name_tag=)
Found: ami-83a1fc95 (devenv-rhel7_6323)
++ seq 0 2
+ for i in '$(seq 0 2)'
+ vagrant up --provider aws
Bringing machine 'openshiftdev' up with 'aws' provider...
==> openshiftdev: Warning! The AWS provider doesn't support any of the Vagrant
==> openshiftdev: high-level network configurations (`config.vm.network`). They
==> openshiftdev: will be silently ignored.
==> openshiftdev: Warning! You're launching this instance into a VPC without an
==> openshiftdev: elastic IP. Please verify you're properly connected to a VPN so
==> openshiftdev: you can access this machine, otherwise Vagrant will not be able
==> openshiftdev: to SSH into it.
==> openshiftdev: Launching an instance with the following settings...
==> openshiftdev: -- Type: c4.xlarge
==> openshiftdev: -- AMI: ami-83a1fc95
==> openshiftdev: -- Region: us-east-1
==> openshiftdev: -- Keypair: libra
==> openshiftdev: -- Subnet ID: subnet-cf57c596
==> openshiftdev: -- User Data: yes
==> openshiftdev: -- User Data:
==> openshiftdev: # cloud-config
==> openshiftdev:
==> openshiftdev: growpart:
==> openshiftdev: mode: auto
==> openshiftdev: devices: ['/']
==> openshiftdev: runcmd:
==> openshiftdev: - [ sh, -xc, "sed -i s/^Defaults.*requiretty/#Defaults requiretty/g /etc/sudoers"]
==> openshiftdev:
==> openshiftdev: -- Block Device Mapping: [{"DeviceName"=>"/dev/sda1", "Ebs.VolumeSize"=>25, "Ebs.VolumeType"=>"gp2"}, {"DeviceName"=>"/dev/sdb", "Ebs.VolumeSize"=>35, "Ebs.VolumeType"=>"gp2"}]
==> openshiftdev: -- Terminate On Shutdown: false
==> openshiftdev: -- Monitoring: false
==> openshiftdev: -- EBS optimized: false
==> openshiftdev: -- Assigning a public IP address in a VPC: false
/var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/middlewares/expects.rb:6:in `response_call': The instance ID 'i-00bd1adc91a71f8d3' does not exist (Fog::Compute::AWS::NotFound)
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/middlewares/response_parser.rb:8:in `response_call'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/connection.rb:389:in `response'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/connection.rb:253:in `request'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/middlewares/idempotent.rb:26:in `error_call'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/middlewares/base.rb:10:in `error_call'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/middlewares/base.rb:10:in `error_call'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/connection.rb:273:in `rescue in request'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/connection.rb:221:in `request'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/middlewares/idempotent.rb:26:in `error_call'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/middlewares/base.rb:10:in `error_call'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/middlewares/base.rb:10:in `error_call'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/connection.rb:273:in `rescue in request'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/connection.rb:221:in `request'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/middlewares/idempotent.rb:26:in `error_call'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/middlewares/base.rb:10:in `error_call'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/middlewares/base.rb:10:in `error_call'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/connection.rb:273:in `rescue in request'
from /var/lib/jenkins/.vagrant.d/gems/gems/excon-0.49.0/lib/excon/connection.rb:221:in `request'
from /var/lib/jenkins/.vagrant.d/gems/gems/fog-xml-0.1.1/lib/fog/xml/sax_parser_connection.rb:37:in `request'
from /var/lib/jenkins/.vagrant.d/gems/gems/fog-xml-0.1.1/lib/fog/xml/connection.rb:7:in `request'
from /var/lib/jenkins/.vagrant.d/gems/gems/fog-1.26.0/lib/fog/aws/compute.rb:522:in `_request'
from /var/lib/jenkins/.vagrant.d/gems/gems/fog-1.26.0/lib/fog/aws/compute.rb:517:in `request'
from /var/lib/jenkins/.vagrant.d/gems/gems/fog-1.26.0/lib/fog/aws/requests/compute/create_tags.rb:31:in `create_tags'
from /var/lib/jenkins/.vagrant.d/gems/gems/fog-1.26.0/lib/fog/aws/models/compute/servers.rb:167:in `block in save_many'
from /var/lib/jenkins/.vagrant.d/gems/gems/fog-1.26.0/lib/fog/aws/models/compute/servers.rb:161:in `map'
from /var/lib/jenkins/.vagrant.d/gems/gems/fog-1.26.0/lib/fog/aws/models/compute/servers.rb:161:in `save_many'
from /var/lib/jenkins/.vagrant.d/gems/gems/fog-1.26.0/lib/fog/aws/models/compute/server.rb:201:in `save'
from /var/lib/jenkins/.vagrant.d/gems/gems/fog-core-1.43.0/lib/fog/core/collection.rb:51:in `create'
from /var/lib/jenkins/.vagrant.d/gems/gems/vagrant-aws-0.6.0/lib/vagrant-aws/action/run_instance.rb:102:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:34:in `call'
from /var/lib/jenkins/.vagrant.d/gems/gems/vagrant-openshift-3.0.9/lib/vagrant-openshift/hooks/find_ami.rb:37:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:34:in `call'
from /var/lib/jenkins/.vagrant.d/gems/gems/vagrant-aws-0.6.0/lib/vagrant-aws/action/elb_register_instance.rb:16:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:34:in `call'
from /var/lib/jenkins/.vagrant.d/gems/gems/vagrant-aws-0.6.0/lib/vagrant-aws/action/warn_networks.rb:14:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:34:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/builtin/synced_folders.rb:86:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:34:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/builtin/provision.rb:80:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:34:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:95:in `block in finalize_action'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:34:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:34:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/builder.rb:116:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/runner.rb:66:in `block in run'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/util/busy.rb:19:in `busy'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/runner.rb:66:in `run'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/builtin/call.rb:53:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:34:in `call'
from /var/lib/jenkins/.vagrant.d/gems/gems/vagrant-aws-0.6.0/lib/vagrant-aws/action/connect_aws.rb:43:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:34:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/builtin/config_validate.rb:25:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:34:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/builtin/handle_box.rb:56:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/warden.rb:34:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/builder.rb:116:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/runner.rb:66:in `block in run'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/util/busy.rb:19:in `busy'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/action/runner.rb:66:in `run'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/machine.rb:214:in `action_raw'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/machine.rb:191:in `block in action'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/environment.rb:516:in `lock'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/machine.rb:178:in `call'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/machine.rb:178:in `action'
from /opt/vagrant/embedded/gems/gems/vagrant-1.7.4/lib/vagrant/batch_action.rb:82:in `block (2 levels) in run'
+ echo ''\''vagrant up'\'' failed - retrying'
'vagrant up' failed - retrying
+ vagrant destroy -f
==> openshiftdev: Instance is not created. Please run `vagrant up` first.
+ for i in '$(seq 0 2)'
+ vagrant up --provider aws
Bringing machine 'openshiftdev' up with 'aws' provider...
==> openshiftdev: Warning! The AWS provider doesn't support any of the Vagrant
==> openshiftdev: high-level network configurations (`config.vm.network`). They
==> openshiftdev: will be silently ignored.
==> openshiftdev: Warning! You're launching this instance into a VPC without an
==> openshiftdev: elastic IP. Please verify you're properly connected to a VPN so
==> openshiftdev: you can access this machine, otherwise Vagrant will not be able
==> openshiftdev: to SSH into it.
==> openshiftdev: Launching an instance with the following settings...
==> openshiftdev: -- Type: c4.xlarge
==> openshiftdev: -- AMI: ami-83a1fc95
==> openshiftdev: -- Region: us-east-1
==> openshiftdev: -- Keypair: libra
==> openshiftdev: -- Subnet ID: subnet-cf57c596
==> openshiftdev: -- User Data: yes
==> openshiftdev: -- User Data:
==> openshiftdev: # cloud-config
==> openshiftdev:
==> openshiftdev: growpart:
==> openshiftdev: mode: auto
==> openshiftdev: devices: ['/']
==> openshiftdev: runcmd:
==> openshiftdev: - [ sh, -xc, "sed -i s/^Defaults.*requiretty/#Defaults requiretty/g /etc/sudoers"]
==> openshiftdev:
==> openshiftdev: -- Block Device Mapping: [{"DeviceName"=>"/dev/sda1", "Ebs.VolumeSize"=>25, "Ebs.VolumeType"=>"gp2"}, {"DeviceName"=>"/dev/sdb", "Ebs.VolumeSize"=>35, "Ebs.VolumeType"=>"gp2"}]
==> openshiftdev: -- Terminate On Shutdown: false
==> openshiftdev: -- Monitoring: false
==> openshiftdev: -- EBS optimized: false
==> openshiftdev: -- Assigning a public IP address in a VPC: false
==> openshiftdev: Waiting for instance to become "ready"...
==> openshiftdev: Waiting for SSH to become available...
==> openshiftdev: Machine is booted and ready for use!
==> openshiftdev: Running provisioner: setup (shell)...
openshiftdev: Running: /tmp/vagrant-shell20170608-23701-lhkrh5.sh
==> openshiftdev: Host: ec2-52-200-44-42.compute-1.amazonaws.com
+ break
+ vagrant sync-origin-aggregated-logging -c -s
Running ssh/sudo command 'rm -rf /data/src/github.com/openshift/origin-aggregated-logging-bare;
' with timeout 14400. Attempt #0
Running ssh/sudo command 'mkdir -p /ec2-user/.ssh;
mv /tmp/file20170608-24794-wj0m9d /ec2-user/.ssh/config &&
chown ec2-user:ec2-user /ec2-user/.ssh/config &&
chmod 0600 /ec2-user/.ssh/config' with timeout 14400. Attempt #0
Running ssh/sudo command 'mkdir -p /data/src/github.com/openshift/' with timeout 14400. Attempt #0
Running ssh/sudo command 'mkdir -p /data/src/github.com/openshift/builder && chown -R ec2-user:ec2-user /data/src/github.com/openshift/' with timeout 14400. Attempt #0
Running ssh/sudo command 'set -e
rm -fr /data/src/github.com/openshift/origin-aggregated-logging-bare;
if [ ! -d /data/src/github.com/openshift/origin-aggregated-logging-bare ]; then
git clone --quiet --bare https://github.com/openshift/origin-aggregated-logging.git /data/src/github.com/openshift/origin-aggregated-logging-bare >/dev/null
fi
' with timeout 14400. Attempt #0
Synchronizing local sources
Synchronizing [origin-aggregated-logging@master] from origin-aggregated-logging...
Warning: Permanently added '52.200.44.42' (ECDSA) to the list of known hosts.
Running ssh/sudo command 'set -e
if [ -d /data/src/github.com/openshift/origin-aggregated-logging-bare ]; then
rm -rf /data/src/github.com/openshift/origin-aggregated-logging
echo 'Cloning origin-aggregated-logging ...'
git clone --quiet --recurse-submodules /data/src/github.com/openshift/origin-aggregated-logging-bare /data/src/github.com/openshift/origin-aggregated-logging
else
MISSING_REPO+='origin-aggregated-logging-bare'
fi
if [ -n "$MISSING_REPO" ]; then
echo 'Missing required upstream repositories:'
echo $MISSING_REPO
echo 'To fix, execute command: vagrant clone-upstream-repos'
fi
' with timeout 14400. Attempt #0
Cloning origin-aggregated-logging ...
Submodule 'deployer/common' (https://github.com/openshift/origin-integration-common) registered for path 'deployer/common'
Submodule 'kibana-proxy' (https://github.com/fabric8io/openshift-auth-proxy.git) registered for path 'kibana-proxy'
Cloning into 'deployer/common'...
Submodule path 'deployer/common': checked out '45bf993212cdcbab5cbce3b3fab74a72b851402e'
Cloning into 'kibana-proxy'...
Submodule path 'kibana-proxy': checked out '118dfb40f7a8082d370ba7f4805255c9ec7c8178'
+ vagrant ssh -c 'if [ ! -d /tmp/openshift ] ; then mkdir /tmp/openshift ; fi ; sudo chmod 777 /tmp/openshift'
+ for image in openshift/base-centos7 centos:centos7 openshift/origin-logging-elasticsearch openshift/origin-logging-fluentd openshift/origin-logging-curator openshift/origin-logging-kibana
+ echo pulling image openshift/base-centos7 ...
pulling image openshift/base-centos7 ...
+ vagrant ssh -c 'docker pull openshift/base-centos7' -- -n
Using default tag: latest
Trying to pull repository docker.io/openshift/base-centos7 ...
latest: Pulling from docker.io/openshift/base-centos7
45a2e645736c: Pulling fs layer
734fb161cf89: Pulling fs layer
78efc9e155c4: Pulling fs layer
8a3400b7e31a: Pulling fs layer
8a3400b7e31a: Waiting
734fb161cf89: Verifying Checksum
734fb161cf89: Download complete
8a3400b7e31a: Download complete
45a2e645736c: Download complete
78efc9e155c4: Verifying Checksum
78efc9e155c4: Download complete
45a2e645736c: Pull complete
734fb161cf89: Pull complete
78efc9e155c4: Pull complete
8a3400b7e31a: Pull complete
Digest: sha256:aea292a3bddba020cde0ee83e6a45807931eb607c164ec6a3674f67039d8cd7c
+ echo done with openshift/base-centos7
done with openshift/base-centos7
+ for image in openshift/base-centos7 centos:centos7 openshift/origin-logging-elasticsearch openshift/origin-logging-fluentd openshift/origin-logging-curator openshift/origin-logging-kibana
+ echo pulling image centos:centos7 ...
pulling image centos:centos7 ...
+ vagrant ssh -c 'docker pull centos:centos7' -- -n
Trying to pull repository docker.io/library/centos ...
centos7: Pulling from docker.io/library/centos
Digest: sha256:aebf12af704307dfa0079b3babdca8d7e8ff6564696882bcb5d11f1d461f9ee9
+ echo done with centos:centos7
done with centos:centos7
+ for image in openshift/base-centos7 centos:centos7 openshift/origin-logging-elasticsearch openshift/origin-logging-fluentd openshift/origin-logging-curator openshift/origin-logging-kibana
+ echo pulling image openshift/origin-logging-elasticsearch ...
pulling image openshift/origin-logging-elasticsearch ...
+ vagrant ssh -c 'docker pull openshift/origin-logging-elasticsearch' -- -n
Using default tag: latest
Trying to pull repository docker.io/openshift/origin-logging-elasticsearch ...
latest: Pulling from docker.io/openshift/origin-logging-elasticsearch
d5e46245fe40: Already exists
3a9c76d9e103: Pulling fs layer
68dc0402a412: Pulling fs layer
a3e21278d73b: Pulling fs layer
1cb971d1904a: Pulling fs layer
7f787b4c1aa4: Pulling fs layer
eb9125dce9a5: Pulling fs layer
09427247134d: Pulling fs layer
3ac7e779da16: Pulling fs layer
e68089779cf0: Pulling fs layer
f77449eb85d6: Pulling fs layer
eb9125dce9a5: Waiting
09427247134d: Waiting
3ac7e779da16: Waiting
e68089779cf0: Waiting
f77449eb85d6: Waiting
1cb971d1904a: Waiting
7f787b4c1aa4: Waiting
a3e21278d73b: Verifying Checksum
a3e21278d73b: Download complete
3a9c76d9e103: Verifying Checksum
1cb971d1904a: Verifying Checksum
1cb971d1904a: Download complete
eb9125dce9a5: Verifying Checksum
eb9125dce9a5: Download complete
7f787b4c1aa4: Download complete
09427247134d: Verifying Checksum
09427247134d: Download complete
e68089779cf0: Download complete
f77449eb85d6: Verifying Checksum
f77449eb85d6: Download complete
3ac7e779da16: Verifying Checksum
3ac7e779da16: Download complete
68dc0402a412: Download complete
3a9c76d9e103: Pull complete
68dc0402a412: Pull complete
a3e21278d73b: Pull complete
1cb971d1904a: Pull complete
7f787b4c1aa4: Pull complete
eb9125dce9a5: Pull complete
09427247134d: Pull complete
3ac7e779da16: Pull complete
e68089779cf0: Pull complete
f77449eb85d6: Pull complete
Digest: sha256:273c11b82929c5f946ddb48bd49dbe1e77ba40d9c15275dc7619fd07b9c8ec54
+ echo done with openshift/origin-logging-elasticsearch
done with openshift/origin-logging-elasticsearch
+ for image in openshift/base-centos7 centos:centos7 openshift/origin-logging-elasticsearch openshift/origin-logging-fluentd openshift/origin-logging-curator openshift/origin-logging-kibana
+ echo pulling image openshift/origin-logging-fluentd ...
pulling image openshift/origin-logging-fluentd ...
+ vagrant ssh -c 'docker pull openshift/origin-logging-fluentd' -- -n
Using default tag: latest
Trying to pull repository docker.io/openshift/origin-logging-fluentd ...
latest: Pulling from docker.io/openshift/origin-logging-fluentd
d5e46245fe40: Already exists
6318bc7c7f58: Pulling fs layer
3b55aa2d4d05: Pulling fs layer
814e0506d29d: Pulling fs layer
ed25a891441c: Pulling fs layer
63d9bd56de92: Pulling fs layer
ed25a891441c: Waiting
63d9bd56de92: Waiting
814e0506d29d: Verifying Checksum
814e0506d29d: Download complete
ed25a891441c: Verifying Checksum
ed25a891441c: Download complete
63d9bd56de92: Verifying Checksum
63d9bd56de92: Download complete
3b55aa2d4d05: Verifying Checksum
3b55aa2d4d05: Download complete
6318bc7c7f58: Download complete
6318bc7c7f58: Pull complete
3b55aa2d4d05: Pull complete
814e0506d29d: Pull complete
ed25a891441c: Pull complete
63d9bd56de92: Pull complete
Digest: sha256:b55877bf5f5624c0111688db16bac54ed7b64291114323fb010d9e8e630b1c89
+ echo done with openshift/origin-logging-fluentd
done with openshift/origin-logging-fluentd
+ for image in openshift/base-centos7 centos:centos7 openshift/origin-logging-elasticsearch openshift/origin-logging-fluentd openshift/origin-logging-curator openshift/origin-logging-kibana
+ echo pulling image openshift/origin-logging-curator ...
pulling image openshift/origin-logging-curator ...
+ vagrant ssh -c 'docker pull openshift/origin-logging-curator' -- -n
Using default tag: latest
Trying to pull repository docker.io/openshift/origin-logging-curator ...
latest: Pulling from docker.io/openshift/origin-logging-curator
d5e46245fe40: Already exists
b84a40433153: Pulling fs layer
b797c0ddd9bc: Pulling fs layer
b84a40433153: Download complete
b797c0ddd9bc: Verifying Checksum
b797c0ddd9bc: Download complete
b84a40433153: Pull complete
b797c0ddd9bc: Pull complete
Digest: sha256:c09a1f9ef6f2ca23b4d3dca5f4a2c0f5bb76ed65351d9295d7d3344e5bce3f89
+ echo done with openshift/origin-logging-curator
done with openshift/origin-logging-curator
+ for image in openshift/base-centos7 centos:centos7 openshift/origin-logging-elasticsearch openshift/origin-logging-fluentd openshift/origin-logging-curator openshift/origin-logging-kibana
+ echo pulling image openshift/origin-logging-kibana ...
pulling image openshift/origin-logging-kibana ...
+ vagrant ssh -c 'docker pull openshift/origin-logging-kibana' -- -n
Using default tag: latest
Trying to pull repository docker.io/openshift/origin-logging-kibana ...
latest: Pulling from docker.io/openshift/origin-logging-kibana
45a2e645736c: Already exists
734fb161cf89: Already exists
78efc9e155c4: Already exists
8a3400b7e31a: Already exists
d24cd3f22cc6: Pulling fs layer
b64de6f5fdbe: Pulling fs layer
d443465a8496: Pulling fs layer
cb8fa38c6d97: Pulling fs layer
9879c78ef10f: Pulling fs layer
3f29d3f59ef3: Pulling fs layer
cb8fa38c6d97: Waiting
3f29d3f59ef3: Waiting
9879c78ef10f: Waiting
d443465a8496: Verifying Checksum
d443465a8496: Download complete
d24cd3f22cc6: Verifying Checksum
d24cd3f22cc6: Download complete
cb8fa38c6d97: Verifying Checksum
cb8fa38c6d97: Download complete
9879c78ef10f: Verifying Checksum
9879c78ef10f: Download complete
d24cd3f22cc6: Pull complete
b64de6f5fdbe: Verifying Checksum
b64de6f5fdbe: Download complete
3f29d3f59ef3: Verifying Checksum
3f29d3f59ef3: Download complete
b64de6f5fdbe: Pull complete
d443465a8496: Pull complete
cb8fa38c6d97: Pull complete
9879c78ef10f: Pull complete
3f29d3f59ef3: Pull complete
Digest: sha256:3b95068a8514b13780c047f955e99d481af5597bf2b214501030e7be2f4d6dc1
+ echo done with openshift/origin-logging-kibana
done with openshift/origin-logging-kibana
+ vagrant test-origin-aggregated-logging -d --env GIT_URL=https://github.com/openshift/origin-aggregated-logging --env GIT_BRANCH=master --env O_A_L_DIR=/data/src/github.com/openshift/origin-aggregated-logging --env OS_ROOT=/data/src/github.com/openshift/origin --env ENABLE_OPS_CLUSTER=true --env USE_LOCAL_SOURCE=true --env TEST_PERF=false --env VERBOSE=1 --env OS_ANSIBLE_REPO=https://github.com/openshift/openshift-ansible --env OS_ANSIBLE_BRANCH=master
***************************************************
Running GIT_URL=https://github.com/openshift/origin-aggregated-logging GIT_BRANCH=master O_A_L_DIR=/data/src/github.com/openshift/origin-aggregated-logging OS_ROOT=/data/src/github.com/openshift/origin ENABLE_OPS_CLUSTER=true USE_LOCAL_SOURCE=true TEST_PERF=false VERBOSE=1 OS_ANSIBLE_REPO=https://github.com/openshift/openshift-ansible OS_ANSIBLE_BRANCH=master ./logging.sh...
/data/src/github.com/openshift/origin /data/src/github.com/openshift/origin-aggregated-logging/hack/testing
/data/src/github.com/openshift/origin-aggregated-logging/hack/testing
/data/src/github.com/openshift/origin-aggregated-logging /data/src/github.com/openshift/origin-aggregated-logging/hack/testing
/data/src/github.com/openshift/origin-aggregated-logging/hack/testing
Loaded plugins: amazon-id, rhui-lb, search-disabled-repos
Metadata Cache Created
Loaded plugins: amazon-id, rhui-lb, search-disabled-repos
Resolving Dependencies
--> Running transaction check
---> Package ansible.noarch 0:2.3.0.0-3.el7 will be installed
--> Processing Dependency: sshpass for package: ansible-2.3.0.0-3.el7.noarch
--> Processing Dependency: python-paramiko for package: ansible-2.3.0.0-3.el7.noarch
--> Processing Dependency: python-keyczar for package: ansible-2.3.0.0-3.el7.noarch
--> Processing Dependency: python-httplib2 for package: ansible-2.3.0.0-3.el7.noarch
--> Processing Dependency: python-crypto for package: ansible-2.3.0.0-3.el7.noarch
---> Package python2-pip.noarch 0:8.1.2-5.el7 will be installed
---> Package python2-ruamel-yaml.x86_64 0:0.12.14-9.el7 will be installed
--> Processing Dependency: python2-typing for package: python2-ruamel-yaml-0.12.14-9.el7.x86_64
--> Processing Dependency: python2-ruamel-ordereddict for package: python2-ruamel-yaml-0.12.14-9.el7.x86_64
--> Running transaction check
---> Package python-httplib2.noarch 0:0.9.1-2.el7aos will be installed
---> Package python-keyczar.noarch 0:0.71c-2.el7aos will be installed
--> Processing Dependency: python-pyasn1 for package: python-keyczar-0.71c-2.el7aos.noarch
---> Package python-paramiko.noarch 0:2.1.1-1.el7 will be installed
--> Processing Dependency: python-cryptography for package: python-paramiko-2.1.1-1.el7.noarch
---> Package python2-crypto.x86_64 0:2.6.1-13.el7 will be installed
--> Processing Dependency: libtomcrypt.so.0()(64bit) for package: python2-crypto-2.6.1-13.el7.x86_64
---> Package python2-ruamel-ordereddict.x86_64 0:0.4.9-3.el7 will be installed
---> Package python2-typing.noarch 0:3.5.2.2-3.el7 will be installed
---> Package sshpass.x86_64 0:1.06-1.el7 will be installed
--> Running transaction check
---> Package libtomcrypt.x86_64 0:1.17-23.el7 will be installed
--> Processing Dependency: libtommath >= 0.42.0 for package: libtomcrypt-1.17-23.el7.x86_64
--> Processing Dependency: libtommath.so.0()(64bit) for package: libtomcrypt-1.17-23.el7.x86_64
---> Package python2-cryptography.x86_64 0:1.3.1-3.el7 will be installed
--> Processing Dependency: python-idna >= 2.0 for package: python2-cryptography-1.3.1-3.el7.x86_64
--> Processing Dependency: python-cffi >= 1.4.1 for package: python2-cryptography-1.3.1-3.el7.x86_64
--> Processing Dependency: python-ipaddress for package: python2-cryptography-1.3.1-3.el7.x86_64
--> Processing Dependency: python-enum34 for package: python2-cryptography-1.3.1-3.el7.x86_64
---> Package python2-pyasn1.noarch 0:0.1.9-7.el7 will be installed
--> Running transaction check
---> Package libtommath.x86_64 0:0.42.0-4.el7 will be installed
---> Package python-cffi.x86_64 0:1.6.0-5.el7 will be installed
--> Processing Dependency: python-pycparser for package: python-cffi-1.6.0-5.el7.x86_64
---> Package python-enum34.noarch 0:1.0.4-1.el7 will be installed
---> Package python-idna.noarch 0:2.0-1.el7 will be installed
---> Package python-ipaddress.noarch 0:1.0.16-2.el7 will be installed
--> Running transaction check
---> Package python-pycparser.noarch 0:2.14-1.el7 will be installed
--> Processing Dependency: python-ply for package: python-pycparser-2.14-1.el7.noarch
--> Running transaction check
---> Package python-ply.noarch 0:3.4-10.el7 will be installed
--> Finished Dependency Resolution
Dependencies Resolved
================================================================================
Package Arch Version Repository Size
================================================================================
Installing:
ansible noarch 2.3.0.0-3.el7 epel 5.7 M
python2-pip noarch 8.1.2-5.el7 epel 1.7 M
python2-ruamel-yaml x86_64 0.12.14-9.el7 li 245 k
Installing for dependencies:
libtomcrypt x86_64 1.17-23.el7 epel 224 k
libtommath x86_64 0.42.0-4.el7 epel 35 k
python-cffi x86_64 1.6.0-5.el7 oso-rhui-rhel-server-releases 218 k
python-enum34 noarch 1.0.4-1.el7 oso-rhui-rhel-server-releases 52 k
python-httplib2 noarch 0.9.1-2.el7aos li 115 k
python-idna noarch 2.0-1.el7 oso-rhui-rhel-server-releases 92 k
python-ipaddress noarch 1.0.16-2.el7 oso-rhui-rhel-server-releases 34 k
python-keyczar noarch 0.71c-2.el7aos rhel-7-server-ose-3.1-rpms 217 k
python-paramiko noarch 2.1.1-1.el7 rhel-7-server-ose-3.4-rpms 266 k
python-ply noarch 3.4-10.el7 oso-rhui-rhel-server-releases 123 k
python-pycparser noarch 2.14-1.el7 oso-rhui-rhel-server-releases 105 k
python2-crypto x86_64 2.6.1-13.el7 epel 476 k
python2-cryptography x86_64 1.3.1-3.el7 oso-rhui-rhel-server-releases 471 k
python2-pyasn1 noarch 0.1.9-7.el7 oso-rhui-rhel-server-releases 100 k
python2-ruamel-ordereddict
x86_64 0.4.9-3.el7 li 38 k
python2-typing noarch 3.5.2.2-3.el7 epel 39 k
sshpass x86_64 1.06-1.el7 epel 21 k
Transaction Summary
================================================================================
Install 3 Packages (+17 Dependent packages)
Total download size: 10 M
Installed size: 47 M
Downloading packages:
--------------------------------------------------------------------------------
Total 5.3 MB/s | 10 MB 00:01
Running transaction check
Running transaction test
Transaction test succeeded
Running transaction
Installing : python2-pyasn1-0.1.9-7.el7.noarch 1/20
Installing : sshpass-1.06-1.el7.x86_64 2/20
Installing : libtommath-0.42.0-4.el7.x86_64 3/20
Installing : libtomcrypt-1.17-23.el7.x86_64 4/20
Installing : python2-crypto-2.6.1-13.el7.x86_64 5/20
Installing : python-keyczar-0.71c-2.el7aos.noarch 6/20
Installing : python-enum34-1.0.4-1.el7.noarch 7/20
Installing : python-ply-3.4-10.el7.noarch 8/20
Installing : python-pycparser-2.14-1.el7.noarch 9/20
Installing : python-cffi-1.6.0-5.el7.x86_64 10/20
Installing : python-httplib2-0.9.1-2.el7aos.noarch 11/20
Installing : python-idna-2.0-1.el7.noarch 12/20
Installing : python2-ruamel-ordereddict-0.4.9-3.el7.x86_64 13/20
Installing : python2-typing-3.5.2.2-3.el7.noarch 14/20
Installing : python-ipaddress-1.0.16-2.el7.noarch 15/20
Installing : python2-cryptography-1.3.1-3.el7.x86_64 16/20
Installing : python-paramiko-2.1.1-1.el7.noarch 17/20
Installing : ansible-2.3.0.0-3.el7.noarch 18/20
Installing : python2-ruamel-yaml-0.12.14-9.el7.x86_64 19/20
Installing : python2-pip-8.1.2-5.el7.noarch 20/20
Verifying : python-pycparser-2.14-1.el7.noarch 1/20
Verifying : python-ipaddress-1.0.16-2.el7.noarch 2/20
Verifying : ansible-2.3.0.0-3.el7.noarch 3/20
Verifying : python2-typing-3.5.2.2-3.el7.noarch 4/20
Verifying : python2-pip-8.1.2-5.el7.noarch 5/20
Verifying : python2-pyasn1-0.1.9-7.el7.noarch 6/20
Verifying : libtomcrypt-1.17-23.el7.x86_64 7/20
Verifying : python-cffi-1.6.0-5.el7.x86_64 8/20
Verifying : python2-ruamel-yaml-0.12.14-9.el7.x86_64 9/20
Verifying : python2-ruamel-ordereddict-0.4.9-3.el7.x86_64 10/20
Verifying : python-idna-2.0-1.el7.noarch 11/20
Verifying : python-httplib2-0.9.1-2.el7aos.noarch 12/20
Verifying : python-ply-3.4-10.el7.noarch 13/20
Verifying : python-enum34-1.0.4-1.el7.noarch 14/20
Verifying : python-keyczar-0.71c-2.el7aos.noarch 15/20
Verifying : libtommath-0.42.0-4.el7.x86_64 16/20
Verifying : sshpass-1.06-1.el7.x86_64 17/20
Verifying : python2-cryptography-1.3.1-3.el7.x86_64 18/20
Verifying : python-paramiko-2.1.1-1.el7.noarch 19/20
Verifying : python2-crypto-2.6.1-13.el7.x86_64 20/20
Installed:
ansible.noarch 0:2.3.0.0-3.el7 python2-pip.noarch 0:8.1.2-5.el7
python2-ruamel-yaml.x86_64 0:0.12.14-9.el7
Dependency Installed:
libtomcrypt.x86_64 0:1.17-23.el7
libtommath.x86_64 0:0.42.0-4.el7
python-cffi.x86_64 0:1.6.0-5.el7
python-enum34.noarch 0:1.0.4-1.el7
python-httplib2.noarch 0:0.9.1-2.el7aos
python-idna.noarch 0:2.0-1.el7
python-ipaddress.noarch 0:1.0.16-2.el7
python-keyczar.noarch 0:0.71c-2.el7aos
python-paramiko.noarch 0:2.1.1-1.el7
python-ply.noarch 0:3.4-10.el7
python-pycparser.noarch 0:2.14-1.el7
python2-crypto.x86_64 0:2.6.1-13.el7
python2-cryptography.x86_64 0:1.3.1-3.el7
python2-pyasn1.noarch 0:0.1.9-7.el7
python2-ruamel-ordereddict.x86_64 0:0.4.9-3.el7
python2-typing.noarch 0:3.5.2.2-3.el7
sshpass.x86_64 0:1.06-1.el7
Complete!
Cloning into '/tmp/tmp.vkVyOKV2MD/openhift-ansible'...
Copying oc from path to /usr/local/bin for use by openshift-ansible
Copying oc from path to /usr/bin for use by openshift-ansible
Copying oadm from path to /usr/local/bin for use by openshift-ansible
Copying oadm from path to /usr/bin for use by openshift-ansible
[INFO] Starting logging tests at Thu Jun 8 17:44:38 EDT 2017
Generated new key pair as /tmp/openshift/origin-aggregated-logging/openshift.local.config/master/serviceaccounts.public.key and /tmp/openshift/origin-aggregated-logging/openshift.local.config/master/serviceaccounts.private.key
Generating node credentials ...
Created node config for 172.18.1.243 in /tmp/openshift/origin-aggregated-logging/openshift.local.config/node-172.18.1.243
Wrote master config to: /tmp/openshift/origin-aggregated-logging/openshift.local.config/master/master-config.yaml
Running hack/lib/start.sh:352: executing 'oc get --raw /healthz --config='/tmp/openshift/origin-aggregated-logging/openshift.local.config/master/admin.kubeconfig'' expecting any result and text 'ok'; re-trying every 0.25s until completion or 80.000s...
SUCCESS after 6.314s: hack/lib/start.sh:352: executing 'oc get --raw /healthz --config='/tmp/openshift/origin-aggregated-logging/openshift.local.config/master/admin.kubeconfig'' expecting any result and text 'ok'; re-trying every 0.25s until completion or 80.000s
Standard output from the command:
ok
Standard error from the command:
The connection to the server 172.18.1.243:8443 was refused - did you specify the right host or port?
... repeated 5 times
Error from server (Forbidden): User "system:admin" cannot "get" on "/healthz"
... repeated 6 times
Running hack/lib/start.sh:353: executing 'oc get --raw https://172.18.1.243:10250/healthz --config='/tmp/openshift/origin-aggregated-logging/openshift.local.config/master/admin.kubeconfig'' expecting any result and text 'ok'; re-trying every 0.5s until completion or 120.000s...
SUCCESS after 0.243s: hack/lib/start.sh:353: executing 'oc get --raw https://172.18.1.243:10250/healthz --config='/tmp/openshift/origin-aggregated-logging/openshift.local.config/master/admin.kubeconfig'' expecting any result and text 'ok'; re-trying every 0.5s until completion or 120.000s
Standard output from the command:
ok
There was no error output from the command.
Running hack/lib/start.sh:354: executing 'oc get --raw /healthz/ready --config='/tmp/openshift/origin-aggregated-logging/openshift.local.config/master/admin.kubeconfig'' expecting any result and text 'ok'; re-trying every 0.25s until completion or 80.000s...
SUCCESS after 0.793s: hack/lib/start.sh:354: executing 'oc get --raw /healthz/ready --config='/tmp/openshift/origin-aggregated-logging/openshift.local.config/master/admin.kubeconfig'' expecting any result and text 'ok'; re-trying every 0.25s until completion or 80.000s
Standard output from the command:
ok
Standard error from the command:
Error from server (InternalError): an error on the server ("") has prevented the request from succeeding
Running hack/lib/start.sh:355: executing 'oc get service kubernetes --namespace default --config='/tmp/openshift/origin-aggregated-logging/openshift.local.config/master/admin.kubeconfig'' expecting success; re-trying every 0.25s until completion or 160.000s...
SUCCESS after 0.456s: hack/lib/start.sh:355: executing 'oc get service kubernetes --namespace default --config='/tmp/openshift/origin-aggregated-logging/openshift.local.config/master/admin.kubeconfig'' expecting success; re-trying every 0.25s until completion or 160.000s
Standard output from the command:
NAME CLUSTER-IP EXTERNAL-IP PORT(S) AGE
kubernetes 172.30.0.1 <none> 443/TCP,53/UDP,53/TCP 5s
There was no error output from the command.
Running hack/lib/start.sh:356: executing 'oc get --raw /api/v1/nodes/172.18.1.243 --config='/tmp/openshift/origin-aggregated-logging/openshift.local.config/master/admin.kubeconfig'' expecting success; re-trying every 0.25s until completion or 80.000s...
SUCCESS after 0.258s: hack/lib/start.sh:356: executing 'oc get --raw /api/v1/nodes/172.18.1.243 --config='/tmp/openshift/origin-aggregated-logging/openshift.local.config/master/admin.kubeconfig'' expecting success; re-trying every 0.25s until completion or 80.000s
Standard output from the command:
{"kind":"Node","apiVersion":"v1","metadata":{"name":"172.18.1.243","selfLink":"/api/v1/nodes/172.18.1.243","uid":"b84a670e-4c93-11e7-88b4-0e982e1d6fc6","resourceVersion":"295","creationTimestamp":"2017-06-08T21:44:59Z","labels":{"beta.kubernetes.io/arch":"amd64","beta.kubernetes.io/os":"linux","kubernetes.io/hostname":"172.18.1.243"},"annotations":{"volumes.kubernetes.io/controller-managed-attach-detach":"true"}},"spec":{"externalID":"172.18.1.243","providerID":"aws:////i-02d5d2cd5d73ee5e0"},"status":{"capacity":{"cpu":"4","memory":"7231688Ki","pods":"40"},"allocatable":{"cpu":"4","memory":"7129288Ki","pods":"40"},"conditions":[{"type":"OutOfDisk","status":"False","lastHeartbeatTime":"2017-06-08T21:44:59Z","lastTransitionTime":"2017-06-08T21:44:59Z","reason":"KubeletHasSufficientDisk","message":"kubelet has sufficient disk space available"},{"type":"MemoryPressure","status":"False","lastHeartbeatTime":"2017-06-08T21:44:59Z","lastTransitionTime":"2017-06-08T21:44:59Z","reason":"KubeletHasSufficientMemory","message":"kubelet has sufficient memory available"},{"type":"DiskPressure","status":"False","lastHeartbeatTime":"2017-06-08T21:44:59Z","lastTransitionTime":"2017-06-08T21:44:59Z","reason":"KubeletHasNoDiskPressure","message":"kubelet has no disk pressure"},{"type":"Ready","status":"True","lastHeartbeatTime":"2017-06-08T21:44:59Z","lastTransitionTime":"2017-06-08T21:44:59Z","reason":"KubeletReady","message":"kubelet is posting ready status"}],"addresses":[{"type":"LegacyHostIP","address":"172.18.1.243"},{"type":"InternalIP","address":"172.18.1.243"},{"type":"Hostname","address":"172.18.1.243"}],"daemonEndpoints":{"kubeletEndpoint":{"Port":10250}},"nodeInfo":{"machineID":"f9370ed252a14f73b014c1301a9b6d1b","systemUUID":"EC2C74B5-291C-E37B-AE6B-3CADABD8C74B","bootID":"fe112b5c-8928-4ad8-ae62-60c67392ec78","kernelVersion":"3.10.0-327.22.2.el7.x86_64","osImage":"Red Hat Enterprise Linux Server 7.3 (Maipo)","containerRuntimeVersion":"docker://1.12.6","kubeletVersion":"v1.6.1+5115d708d7","kubeProxyVersion":"v1.6.1+5115d708d7","operatingSystem":"linux","architecture":"amd64"},"images":[{"names":["openshift/origin-federation:6acabdc","openshift/origin-federation:latest"],"sizeBytes":1205885664},{"names":["openshift/origin-docker-registry:6acabdc","openshift/origin-docker-registry:latest"],"sizeBytes":1100164272},{"names":["openshift/origin-gitserver:6acabdc","openshift/origin-gitserver:latest"],"sizeBytes":1086520226},{"names":["openshift/openvswitch:6acabdc","openshift/openvswitch:latest"],"sizeBytes":1053403667},{"names":["openshift/node:6acabdc","openshift/node:latest"],"sizeBytes":1051721928},{"names":["openshift/origin-keepalived-ipfailover:6acabdc","openshift/origin-keepalived-ipfailover:latest"],"sizeBytes":1028529711},{"names":["openshift/origin-haproxy-router:6acabdc","openshift/origin-haproxy-router:latest"],"sizeBytes":1022758742},{"names":["openshift/origin:6acabdc","openshift/origin:latest"],"sizeBytes":1001728427},{"names":["openshift/origin-f5-router:6acabdc","openshift/origin-f5-router:latest"],"sizeBytes":1001728427},{"names":["openshift/origin-sti-builder:6acabdc","openshift/origin-sti-builder:latest"],"sizeBytes":1001728427},{"names":["openshift/origin-recycler:6acabdc","openshift/origin-recycler:latest"],"sizeBytes":1001728427},{"names":["openshift/origin-deployer:6acabdc","openshift/origin-deployer:latest"],"sizeBytes":1001728427},{"names":["openshift/origin-docker-builder:6acabdc","openshift/origin-docker-builder:latest"],"sizeBytes":1001728427},{"names":["openshift/origin-cluster-capacity:6acabdc","openshift/origin-cluster-capacity:latest"],"sizeBytes":962455026},{"names":["rhel7.1:latest"],"sizeBytes":765301508},{"names":["openshift/dind-master:latest"],"sizeBytes":731456758},{"names":["openshift/dind-node:latest"],"sizeBytes":731453034},{"names":["\u003cnone\u003e@\u003cnone\u003e","\u003cnone\u003e:\u003cnone\u003e"],"sizeBytes":709532011},{"names":["docker.io/openshift/origin-logging-kibana@sha256:3b95068a8514b13780c047f955e99d481af5597bf2b214501030e7be2f4d6dc1","docker.io/openshift/origin-logging-kibana:latest"],"sizeBytes":682851503},{"names":["openshift/dind:latest"],"sizeBytes":640650210},{"names":["docker.io/openshift/origin-logging-elasticsearch@sha256:273c11b82929c5f946ddb48bd49dbe1e77ba40d9c15275dc7619fd07b9c8ec54","docker.io/openshift/origin-logging-elasticsearch:latest"],"sizeBytes":425433879},{"names":["docker.io/openshift/base-centos7@sha256:aea292a3bddba020cde0ee83e6a45807931eb607c164ec6a3674f67039d8cd7c","docker.io/openshift/base-centos7:latest"],"sizeBytes":383049978},{"names":["rhel7.2:latest"],"sizeBytes":377493597},{"names":["openshift/origin-egress-router:6acabdc","openshift/origin-egress-router:latest"],"sizeBytes":364745713},{"names":["openshift/origin-base:latest"],"sizeBytes":363070172},{"names":["\u003cnone\u003e@\u003cnone\u003e","\u003cnone\u003e:\u003cnone\u003e"],"sizeBytes":363024702},{"names":["docker.io/openshift/origin-logging-fluentd@sha256:b55877bf5f5624c0111688db16bac54ed7b64291114323fb010d9e8e630b1c89","docker.io/openshift/origin-logging-fluentd:latest"],"sizeBytes":359219273},{"names":["docker.io/fedora@sha256:69281ddd7b2600e5f2b17f1e12d7fba25207f459204fb2d15884f8432c479136","docker.io/fedora:25"],"sizeBytes":230864375},{"names":["docker.io/openshift/origin-logging-curator@sha256:c09a1f9ef6f2ca23b4d3dca5f4a2c0f5bb76ed65351d9295d7d3344e5bce3f89","docker.io/openshift/origin-logging-curator:latest"],"sizeBytes":224977447},{"names":["rhel7.3:latest","rhel7:latest"],"sizeBytes":219121266},{"names":["openshift/origin-pod:6acabdc","openshift/origin-pod:latest"],"sizeBytes":213199843},{"names":["registry.access.redhat.com/rhel7.2@sha256:98e6ca5d226c26e31a95cd67716afe22833c943e1926a21daf1a030906a02249","registry.access.redhat.com/rhel7.2:latest"],"sizeBytes":201376319},{"names":["registry.access.redhat.com/rhel7.3@sha256:1e232401d8e0ba53b36b757b4712fbcbd1dab9c21db039c45a84871a74e89e68","registry.access.redhat.com/rhel7.3:latest"],"sizeBytes":192693772},{"names":["docker.io/centos@sha256:bba1de7c9d900a898e3cadbae040dfe8a633c06bc104a0df76ae24483e03c077"],"sizeBytes":192548999},{"names":["openshift/origin-source:latest"],"sizeBytes":192548894},{"names":["docker.io/centos@sha256:aebf12af704307dfa0079b3babdca8d7e8ff6564696882bcb5d11f1d461f9ee9","docker.io/centos:7","docker.io/centos:centos7"],"sizeBytes":192548537},{"names":["registry.access.redhat.com/rhel7.1@sha256:1bc5a4c43bbb29a5a96a61896ff696933be3502e2f5fdc4cde02d9e101731fdd","registry.access.redhat.com/rhel7.1:latest"],"sizeBytes":158229901},{"names":["openshift/hello-openshift:6acabdc","openshift/hello-openshift:latest"],"sizeBytes":5643318}]}}
There was no error output from the command.
serviceaccount "registry" created
clusterrolebinding "registry-registry-role" created
deploymentconfig "docker-registry" created
service "docker-registry" created
--> Creating router router ...
info: password for stats user admin has been set to Iepq0jCtl3
serviceaccount "router" created
clusterrolebinding "router-router-role" created
deploymentconfig "router" created
service "router" created
--> Success
Running /data/src/github.com/openshift/origin-aggregated-logging/logging.sh:162: executing 'oadm new-project logging --node-selector=''' expecting success...
SUCCESS after 1.574s: /data/src/github.com/openshift/origin-aggregated-logging/logging.sh:162: executing 'oadm new-project logging --node-selector=''' expecting success
Standard output from the command:
Created project logging
There was no error output from the command.
Running /data/src/github.com/openshift/origin-aggregated-logging/logging.sh:163: executing 'oc project logging > /dev/null' expecting success...
SUCCESS after 0.275s: /data/src/github.com/openshift/origin-aggregated-logging/logging.sh:163: executing 'oc project logging > /dev/null' expecting success
There was no output from the command.
There was no error output from the command.
apiVersion: v1
items:
- apiVersion: v1
kind: ImageStream
metadata:
labels:
build: logging-elasticsearch
component: development
logging-infra: development
provider: openshift
name: logging-elasticsearch
spec: {}
- apiVersion: v1
kind: ImageStream
metadata:
labels:
build: logging-fluentd
component: development
logging-infra: development
provider: openshift
name: logging-fluentd
spec: {}
- apiVersion: v1
kind: ImageStream
metadata:
labels:
build: logging-kibana
component: development
logging-infra: development
provider: openshift
name: logging-kibana
spec: {}
- apiVersion: v1
kind: ImageStream
metadata:
labels:
build: logging-curator
component: development
logging-infra: development
provider: openshift
name: logging-curator
spec: {}
- apiVersion: v1
kind: ImageStream
metadata:
labels:
build: logging-auth-proxy
component: development
logging-infra: development
provider: openshift
name: logging-auth-proxy
spec: {}
- apiVersion: v1
kind: ImageStream
metadata:
labels:
build: logging-deployment
component: development
logging-infra: development
provider: openshift
name: origin
spec:
dockerImageRepository: openshift/origin
tags:
- from:
kind: DockerImage
name: openshift/origin:v1.5.0-alpha.2
name: v1.5.0-alpha.2
- apiVersion: v1
kind: BuildConfig
metadata:
labels:
app: logging-elasticsearch
component: development
logging-infra: development
provider: openshift
name: logging-elasticsearch
spec:
output:
to:
kind: ImageStreamTag
name: logging-elasticsearch:latest
resources: {}
source:
contextDir: elasticsearch
git:
ref: master
uri: https://github.com/openshift/origin-aggregated-logging
type: Git
strategy:
dockerStrategy:
from:
kind: DockerImage
name: openshift/base-centos7
type: Docker
- apiVersion: v1
kind: BuildConfig
metadata:
labels:
build: logging-fluentd
component: development
logging-infra: development
provider: openshift
name: logging-fluentd
spec:
output:
to:
kind: ImageStreamTag
name: logging-fluentd:latest
resources: {}
source:
contextDir: fluentd
git:
ref: master
uri: https://github.com/openshift/origin-aggregated-logging
type: Git
strategy:
dockerStrategy:
from:
kind: DockerImage
name: openshift/base-centos7
type: Docker
- apiVersion: v1
kind: BuildConfig
metadata:
labels:
build: logging-kibana
component: development
logging-infra: development
provider: openshift
name: logging-kibana
spec:
output:
to:
kind: ImageStreamTag
name: logging-kibana:latest
resources: {}
source:
contextDir: kibana
git:
ref: master
uri: https://github.com/openshift/origin-aggregated-logging
type: Git
strategy:
dockerStrategy:
from:
kind: DockerImage
name: openshift/base-centos7
type: Docker
- apiVersion: v1
kind: BuildConfig
metadata:
labels:
build: logging-curator
component: development
logging-infra: development
provider: openshift
name: logging-curator
spec:
output:
to:
kind: ImageStreamTag
name: logging-curator:latest
resources: {}
source:
contextDir: curator
git:
ref: master
uri: https://github.com/openshift/origin-aggregated-logging
type: Git
strategy:
dockerStrategy:
from:
kind: DockerImage
name: openshift/base-centos7
type: Docker
- apiVersion: v1
kind: BuildConfig
metadata:
labels:
build: logging-auth-proxy
component: development
logging-infra: development
provider: openshift
name: logging-auth-proxy
spec:
output:
to:
kind: ImageStreamTag
name: logging-auth-proxy:latest
resources: {}
source:
contextDir: kibana-proxy
git:
ref: master
uri: https://github.com/openshift/origin-aggregated-logging
type: Git
strategy:
dockerStrategy:
from:
kind: DockerImage
name: library/node:0.10.36
type: Docker
kind: List
metadata: {}
Running hack/testing/build-images:31: executing 'oc process -o yaml -f /data/src/github.com/openshift/origin-aggregated-logging/hack/templates/dev-builds-wo-deployer.yaml -p LOGGING_FORK_URL=https://github.com/openshift/origin-aggregated-logging -p LOGGING_FORK_BRANCH=master | build_filter | oc create -f -' expecting success...
SUCCESS after 0.354s: hack/testing/build-images:31: executing 'oc process -o yaml -f /data/src/github.com/openshift/origin-aggregated-logging/hack/templates/dev-builds-wo-deployer.yaml -p LOGGING_FORK_URL=https://github.com/openshift/origin-aggregated-logging -p LOGGING_FORK_BRANCH=master | build_filter | oc create -f -' expecting success
Standard output from the command:
imagestream "logging-elasticsearch" created
imagestream "logging-fluentd" created
imagestream "logging-kibana" created
imagestream "logging-curator" created
imagestream "logging-auth-proxy" created
imagestream "origin" created
buildconfig "logging-elasticsearch" created
buildconfig "logging-fluentd" created
buildconfig "logging-kibana" created
buildconfig "logging-curator" created
buildconfig "logging-auth-proxy" created
There was no error output from the command.
Running hack/testing/build-images:9: executing 'oc get imagestreamtag origin:latest' expecting success; re-trying every 0.2s until completion or 60.000s...
SUCCESS after 2.661s: hack/testing/build-images:9: executing 'oc get imagestreamtag origin:latest' expecting success; re-trying every 0.2s until completion or 60.000s
Standard output from the command:
NAME DOCKER REF UPDATED IMAGENAME
origin:latest openshift/origin@sha256:d8af300e072a46c4298dc4b2db682a8b104964c1113d93819ce296e185a6cc4a Less than a second ago sha256:d8af300e072a46c4298dc4b2db682a8b104964c1113d93819ce296e185a6cc4a
Standard error from the command:
Error from server (NotFound): imagestreamtags.image.openshift.io "origin:latest" not found
... repeated 6 times
Uploading directory "/data/src/github.com/openshift/origin-aggregated-logging" as binary input for the build ...
build "logging-auth-proxy-1" started
Uploading directory "/data/src/github.com/openshift/origin-aggregated-logging" as binary input for the build ...
build "logging-curator-1" started
Uploading directory "/data/src/github.com/openshift/origin-aggregated-logging" as binary input for the build ...
build "logging-elasticsearch-1" started
Uploading directory "/data/src/github.com/openshift/origin-aggregated-logging" as binary input for the build ...
build "logging-fluentd-1" started
Uploading directory "/data/src/github.com/openshift/origin-aggregated-logging" as binary input for the build ...
build "logging-kibana-1" started
Running hack/testing/build-images:33: executing 'wait_for_builds_complete' expecting success...
FAILURE after 4412.962s: hack/testing/build-images:33: executing 'wait_for_builds_complete' expecting success: the command returned the wrong error code
Standard output from the command:
build "logging-kibana-3" started
build in progress for logging-kibana - delete failed build logging-kibana-1 status complete
build "logging-kibana-1" deleted
error builds are not complete
NAME TYPE FROM STATUS STARTED DURATION
logging-auth-proxy-1 Docker Binary@ce69c3a Complete About an hour ago 2m47s
logging-curator-1 Docker Binary@ce69c3a Complete About an hour ago 1m13s
logging-elasticsearch-1 Docker Binary@ce69c3a Complete About an hour ago 1m54s
logging-fluentd-1 Docker Binary@ce69c3a Complete About an hour ago 1m53s
logging-kibana-2 Docker Binary@ce69c3a Cancelled (CancelledBuild) About an hour ago
logging-kibana-3 Docker Binary@ce69c3a Complete About an hour ago 8m4s
Standard error from the command:
Uploading directory "/data/src/github.com/openshift/origin-aggregated-logging" as binary input for the build ...
Error from server (BadRequest): cannot upload file to build logging-kibana-2 with status Pending
Uploading directory "/data/src/github.com/openshift/origin-aggregated-logging" as binary input for the build ...
[ERROR] PID 4247: hack/lib/cmd.sh:241: `return "${return_code}"` exited with status 1.
[INFO] Stack Trace:
[INFO] 1: hack/lib/cmd.sh:241: `return "${return_code}"`
[INFO] 2: hack/testing/build-images:33: os::cmd::expect_success
[INFO] 3: hack/testing/init-log-stack:14: source
[INFO] 4: /data/src/github.com/openshift/origin-aggregated-logging/logging.sh:166: source
[INFO] Exiting with code 1.
/data/src/github.com/openshift/origin-aggregated-logging/hack/lib/log/system.sh: line 31: 4613 Terminated sar -A -o "${binary_logfile}" 1 86400 > /dev/null 2> "${stderr_logfile}"
[INFO] [CLEANUP] Beginning cleanup routines...
[INFO] [CLEANUP] Dumping cluster events to /tmp/origin-aggregated-logging/artifacts/events.txt
[INFO] [CLEANUP] Dumping etcd contents to /tmp/origin-aggregated-logging/artifacts/etcd
[WARNING] No compiled `etcdhelper` binary was found. Attempting to build one using:
[WARNING] $ hack/build-go.sh tools/etcdhelper
++ Building go targets for linux/amd64: tools/etcdhelper
/data/src/github.com/openshift/origin-aggregated-logging/../origin/hack/build-go.sh took 88 seconds
2017-06-08 19:08:34.451734 I | warning: ignoring ServerName for user-provided CA for backwards compatibility is deprecated
[INFO] [CLEANUP] Dumping container logs to /tmp/origin-aggregated-logging/logs/containers
[INFO] [CLEANUP] Truncating log files over 200M
[INFO] [CLEANUP] Stopping docker containers
[INFO] [CLEANUP] Removing docker containers
[INFO] [CLEANUP] Killing child processes
[INFO] [CLEANUP] Pruning etcd data directory
[ERROR] /data/src/github.com/openshift/origin-aggregated-logging/logging.sh exited with code 1 after 01h 29m 46s
Error while running ssh/sudo command:
set -e
pushd /data/src/github.com/openshift//origin-aggregated-logging/hack/testing >/dev/null
export PATH=$GOPATH/bin:$PATH
echo '***************************************************'
echo 'Running GIT_URL=https://github.com/openshift/origin-aggregated-logging GIT_BRANCH=master O_A_L_DIR=/data/src/github.com/openshift/origin-aggregated-logging OS_ROOT=/data/src/github.com/openshift/origin ENABLE_OPS_CLUSTER=true USE_LOCAL_SOURCE=true TEST_PERF=false VERBOSE=1 OS_ANSIBLE_REPO=https://github.com/openshift/openshift-ansible OS_ANSIBLE_BRANCH=master ./logging.sh...'
time GIT_URL=https://github.com/openshift/origin-aggregated-logging GIT_BRANCH=master O_A_L_DIR=/data/src/github.com/openshift/origin-aggregated-logging OS_ROOT=/data/src/github.com/openshift/origin ENABLE_OPS_CLUSTER=true USE_LOCAL_SOURCE=true TEST_PERF=false VERBOSE=1 OS_ANSIBLE_REPO=https://github.com/openshift/openshift-ansible OS_ANSIBLE_BRANCH=master ./logging.sh
echo 'Finished GIT_URL=https://github.com/openshift/origin-aggregated-logging GIT_BRANCH=master O_A_L_DIR=/data/src/github.com/openshift/origin-aggregated-logging OS_ROOT=/data/src/github.com/openshift/origin ENABLE_OPS_CLUSTER=true USE_LOCAL_SOURCE=true TEST_PERF=false VERBOSE=1 OS_ANSIBLE_REPO=https://github.com/openshift/openshift-ansible OS_ANSIBLE_BRANCH=master ./logging.sh'
echo '***************************************************'
popd >/dev/null
The SSH command responded with a non-zero exit status. Vagrant
assumes that this means the command failed. The output for this command
should be in the log above. Please read the output to determine what
went wrong.
==> openshiftdev: Downloading logs
==> openshiftdev: Downloading artifacts from '/var/log/yum.log' to '/var/lib/jenkins/jobs/test-origin-aggregated-logging/workspace/origin/artifacts/yum.log'
==> openshiftdev: Downloading artifacts from '/var/log/secure' to '/var/lib/jenkins/jobs/test-origin-aggregated-logging/workspace/origin/artifacts/secure'
==> openshiftdev: Downloading artifacts from '/var/log/audit/audit.log' to '/var/lib/jenkins/jobs/test-origin-aggregated-logging/workspace/origin/artifacts/audit.log'
==> openshiftdev: Downloading artifacts from '/tmp/origin-aggregated-logging/' to '/var/lib/jenkins/jobs/test-origin-aggregated-logging/workspace/origin/artifacts'
Build step 'Execute shell' marked build as failure
[description-setter] Could not determine description.
[PostBuildScript] - Execution post build scripts.
[workspace] $ /bin/sh -xe /tmp/hudson3051288733519376676.sh
+ INSTANCE_NAME=origin_logging-rhel7-1641
+ pushd origin
~/jobs/test-origin-aggregated-logging/workspace/origin ~/jobs/test-origin-aggregated-logging/workspace
+ rc=0
+ '[' -f .vagrant-openshift.json ']'
++ /usr/bin/vagrant ssh -c 'sudo ausearch -m avc'
+ ausearchresult='<no matches>'
+ rc=1
+ '[' '<no matches>' = '<no matches>' ']'
+ rc=0
+ /usr/bin/vagrant destroy -f
==> openshiftdev: Terminating the instance...
==> openshiftdev: Running cleanup tasks for 'shell' provisioner...
+ popd
~/jobs/test-origin-aggregated-logging/workspace
+ exit 0
[BFA] Scanning build for known causes...
[BFA] Found failure cause(s):
[BFA] Command Failure from category failure
[BFA] Done. 0s
Finished: FAILURE