Advertisement
Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- [metal-kuber@localhost dev-scripts]$ make
- ./01_install_requirements.sh
- + source common.sh
- ++++ dirname common.sh
- +++ cd .
- +++ pwd
- ++ SCRIPTDIR=/home/metal-kuber/dev-scripts
- +++ whoami
- ++ USER=metal-kuber
- ++ '[' -z '' ']'
- ++ '[' -f /home/metal-kuber/dev-scripts/config_metal-kuber.sh ']'
- ++ echo 'Using CONFIG /home/metal-kuber/dev-scripts/config_metal-kuber.sh'
- Using CONFIG /home/metal-kuber/dev-scripts/config_metal-kuber.sh
- ++ CONFIG=/home/metal-kuber/dev-scripts/config_metal-kuber.sh
- ++ source /home/metal-kuber/dev-scripts/config_metal-kuber.sh
- +++ set +x
- ++ ADDN_DNS=
- ++ EXT_IF=
- ++ PRO_IF=
- ++ MANAGE_BR_BRIDGE=y
- ++ INT_IF=
- ++ ROOT_DISK=/dev/vda
- ++ FILESYSTEM=/
- ++ WORKING_DIR=/opt/dev-scripts
- ++ NODES_FILE=/opt/dev-scripts/ironic_nodes.json
- ++ NODES_PLATFORM=baremetal
- ++ MASTER_NODES_FILE=ocp/master_nodes.json
- ++ export RHCOS_IMAGE_URL=https://releases-rhcos.svc.ci.openshift.org/storage/releases/maipo/
- ++ RHCOS_IMAGE_URL=https://releases-rhcos.svc.ci.openshift.org/storage/releases/maipo/
- +++ curl https://releases-rhcos.svc.ci.openshift.org/storage/releases/maipo//builds.json
- +++ jq -r '.builds[0]'
- % Total % Received % Xferd Average Speed Time Time Time Current
- Dload Upload Total Spent Left Speed
- 100 6067 100 6067 0 0 879 0 0:00:06 0:00:06 --:--:-- 1385
- ++ export RHCOS_LATEST=400.7.20190312.0
- ++ RHCOS_LATEST=400.7.20190312.0
- ++ export RHCOS_IMAGE_VERSION=400.7.20190312.0
- ++ RHCOS_IMAGE_VERSION=400.7.20190312.0
- +++ curl https://releases-rhcos.svc.ci.openshift.org/storage/releases/maipo//400.7.20190312.0/meta.json
- +++ jq -r .images.openstack.path
- % Total % Received % Xferd Average Speed Time Time Time Current
- Dload Upload Total Spent Left Speed
- 100 4539 100 4539 0 0 3403 0 0:00:01 0:00:01 --:--:-- 3405
- ++ export RHCOS_IMAGE_FILENAME_OPENSTACK_GZ=rhcos-maipo-400.7.20190312.0-openstack.qcow2.gz
- ++ RHCOS_IMAGE_FILENAME_OPENSTACK_GZ=rhcos-maipo-400.7.20190312.0-openstack.qcow2.gz
- +++ echo rhcos-maipo-400.7.20190312.0-openstack.qcow2.gz
- +++ sed -e 's/-openstack.*//'
- ++ export RHCOS_IMAGE_NAME=rhcos-maipo-400.7.20190312.0
- ++ RHCOS_IMAGE_NAME=rhcos-maipo-400.7.20190312.0
- ++ export RHCOS_IMAGE_FILENAME_OPENSTACK=rhcos-maipo-400.7.20190312.0-openstack.qcow2
- ++ RHCOS_IMAGE_FILENAME_OPENSTACK=rhcos-maipo-400.7.20190312.0-openstack.qcow2
- ++ export RHCOS_IMAGE_FILENAME_DUALDHCP=rhcos-maipo-400.7.20190312.0-dualdhcp.qcow2
- ++ RHCOS_IMAGE_FILENAME_DUALDHCP=rhcos-maipo-400.7.20190312.0-dualdhcp.qcow2
- ++ export RHCOS_IMAGE_FILENAME_LATEST=rhcos-maipo-latest.qcow2
- ++ RHCOS_IMAGE_FILENAME_LATEST=rhcos-maipo-latest.qcow2
- ++ export IRONIC_IMAGE=quay.io/metalkube/metalkube-ironic
- ++ IRONIC_IMAGE=quay.io/metalkube/metalkube-ironic
- ++ export IRONIC_INSPECTOR_IMAGE=quay.io/metalkube/metalkube-ironic-inspector
- ++ IRONIC_INSPECTOR_IMAGE=quay.io/metalkube/metalkube-ironic-inspector
- ++ export IRONIC_DATA_DIR=/opt/dev-scripts/ironic
- ++ IRONIC_DATA_DIR=/opt/dev-scripts/ironic
- ++ export KUBECONFIG=/home/metal-kuber/dev-scripts/ocp/auth/kubeconfig
- ++ KUBECONFIG=/home/metal-kuber/dev-scripts/ocp/auth/kubeconfig
- ++ export 'SSH=ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o ConnectTimeout=5'
- ++ SSH='ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o ConnectTimeout=5'
- ++ export LIBVIRT_DEFAULT_URI=qemu:///system
- ++ LIBVIRT_DEFAULT_URI=qemu:///system
- ++ '[' metal-kuber '!=' root -a '' == /run/user/0 ']'
- ++ sudo -n uptime
- +++ awk -F= '/^ID=/ { print $2 }' /etc/os-release
- +++ tr -d '"'
- ++ [[ ! centos =~ ^(centos|rhel)$ ]]
- +++ awk -F= '/^VERSION_ID=/ { print $2 }' /etc/os-release
- +++ tr -d '"'
- ++ [[ 7 -ne 7 ]]
- +++ df / --output=fstype
- +++ grep -v Type
- ++ FSTYPE=xfs
- ++ case ${FSTYPE} in
- +++ xfs_info /
- +++ grep -q ftype=1
- ++ [[ -n '' ]]
- ++ '[' 2758 = 0 ']'
- ++ '[' '!' -d /opt/dev-scripts ']'
- +++ dirname ./01_install_requirements.sh
- ++ LOGDIR=./logs
- ++ '[' '!' -d ./logs ']'
- +++ basename ./01_install_requirements.sh .sh
- +++ date +%F-%H%M%S
- ++ LOGFILE=./logs/01_install_requirements-2019-04-03-173807.log
- ++ echo 'Logging to ./logs/01_install_requirements-2019-04-03-173807.log'
- Logging to ./logs/01_install_requirements-2019-04-03-173807.log
- ++ exec
- +++ tee ./logs/01_install_requirements-2019-04-03-173807.log
- ++ BOOTSTRAP_SSH_READY=500
- + sudo setenforce permissive
- + sudo sed -i s/=enforcing/=permissive/g /etc/selinux/config
- + sudo yum -y update
- Loaded plugins: fastestmirror, priorities
- Loading mirror speeds from cached hostfile
- * base: centos.excellmedia.net
- * epel: repos.del.extreme-ix.org
- * extras: centos.excellmedia.net
- * updates: centos.excellmedia.net
- 7 packages excluded due to repository priority protections
- No packages marked for update
- + '[' '!' -f /etc/yum.repos.d/epel.repo ']'
- + grep -q zeromq /etc/yum.repos.d/epel.repo
- + sudo yum -y install crudini curl dnsmasq figlet golang NetworkManager nmap patch psmisc python-pip python-requests python-setuptools vim-enhanced wget
- Loaded plugins: fastestmirror, priorities
- Loading mirror speeds from cached hostfile
- * base: centos.excellmedia.net
- * epel: repos.del.extreme-ix.org
- * extras: centos.excellmedia.net
- * updates: centos.excellmedia.net
- 7 packages excluded due to repository priority protections
- Package crudini-0.9-2.el7.noarch already installed and latest version
- Package curl-7.29.0-51.el7.x86_64 already installed and latest version
- Package dnsmasq-2.76-7.el7.x86_64 already installed and latest version
- Package figlet-2.2.5-9.el7.x86_64 already installed and latest version
- Package golang-1.11.5-1.el7.x86_64 already installed and latest version
- Package 1:NetworkManager-1.12.0-10.el7_6.x86_64 already installed and latest version
- Package 2:nmap-6.40-16.el7.x86_64 already installed and latest version
- Package patch-2.7.1-10.el7_5.x86_64 already installed and latest version
- Package psmisc-22.20-15.el7.x86_64 already installed and latest version
- Package python-pip-8.1.2-1.el7.noarch is obsoleted by python2-pip-8.1.2-7.el7.noarch which is already installed
- Package python-requests-2.6.0-1.el7_1.noarch is obsoleted by python2-requests-2.21.0-2.el7.noarch which is already installed
- Package python-setuptools-0.9.8-7.el7.noarch is obsoleted by python2-setuptools-40.8.0-1.el7.noarch which is already installed
- Package 2:vim-enhanced-7.4.160-5.el7.x86_64 already installed and latest version
- Package wget-1.14-18.el7.x86_64 already installed and latest version
- Nothing to do
- + cd
- + '[' '!' -d tripleo-repos ']'
- + pushd tripleo-repos
- ~/tripleo-repos ~
- + sudo python setup.py install
- running install
- [pbr] Writing ChangeLog
- [pbr] Generating ChangeLog
- [pbr] ChangeLog complete (0.0s)
- [pbr] Generating AUTHORS
- [pbr] AUTHORS complete (0.0s)
- running build
- running build_py
- running egg_info
- writing requirements to tripleo_repos.egg-info/requires.txt
- writing tripleo_repos.egg-info/PKG-INFO
- writing top-level names to tripleo_repos.egg-info/top_level.txt
- writing dependency_links to tripleo_repos.egg-info/dependency_links.txt
- writing entry points to tripleo_repos.egg-info/entry_points.txt
- writing pbr to tripleo_repos.egg-info/pbr.json
- [pbr] Processing SOURCES.txt
- [pbr] In git context, generating filelist from git
- warning: no previously-included files matching '*.pyc' found anywhere in distribution
- writing manifest file 'tripleo_repos.egg-info/SOURCES.txt'
- running install_lib
- running install_egg_info
- removing '/usr/lib/python2.7/site-packages/tripleo_repos-0.0.1.dev75-py2.7.egg-info' (and everything under it)
- Copying tripleo_repos.egg-info to /usr/lib/python2.7/site-packages/tripleo_repos-0.0.1.dev75-py2.7.egg-info
- running install_scripts
- /usr/lib/python2.7/site-packages/pbr/packaging.py:410: EasyInstallDeprecationWarning: Use get_header
- header = easy_install.get_script_header("", executable, is_wininst)
- Installing tripleo-repos script to /usr/bin
- + popd
- ~
- + sudo tripleo-repos current-tripleo
- WARNING: --centos-mirror was deprecated in favour of --mirror
- Loaded plugins: fastestmirror, priorities
- Loading mirror speeds from cached hostfile
- * base: centos.excellmedia.net
- * epel: repos.del.extreme-ix.org
- * extras: centos.excellmedia.net
- * updates: centos.excellmedia.net
- 7 packages excluded due to repository priority protections
- Package yum-plugin-priorities-1.1.31-50.el7.noarch already installed and latest version
- Nothing to do
- Loaded plugins: fastestmirror, priorities
- Cleaning repos: base delorean delorean-master-build-deps delorean-master-testing
- : docker-ce-stable epel extras nodesource rdo-qemu-ev updates yarn
- 30 metadata files removed
- 22 sqlite files removed
- 0 metadata files removed
- Removed old repo "/etc/yum.repos.d/delorean.repo"
- Removed old repo "/etc/yum.repos.d/delorean-master-testing.repo"
- Installed repo delorean to /etc/yum.repos.d/delorean.repo
- Installed repo delorean-master-testing to /etc/yum.repos.d/delorean-master-testing.repo
- + sudo yum -y update
- Loaded plugins: fastestmirror, priorities
- Loading mirror speeds from cached hostfile
- * base: centos.excellmedia.net
- * epel: repos.del.extreme-ix.org
- * extras: centos.excellmedia.net
- * updates: centos.excellmedia.net
- 7 packages excluded due to repository priority protections
- No packages marked for update
- + sudo curl -sL https://dl.yarnpkg.com/rpm/yarn.repo -o /etc/yum.repos.d/yarn.repo
- + curl -sL https://rpm.nodesource.com/setup_10.x
- + sudo bash -
- ## Installing the NodeSource Node.js 10.x repo...
- ## Inspecting system...
- + rpm -q --whatprovides redhat-release || rpm -q --whatprovides centos-release || rpm -q --whatprovides cloudlinux-release || rpm -q --whatprovides sl-release
- + uname -m
- ## Confirming "el7-x86_64" is supported...
- + curl -sLf -o /dev/null 'https://rpm.nodesource.com/pub_10.x/el/7/x86_64/nodesource-release-el7-1.noarch.rpm'
- ## Downloading release setup RPM...
- + mktemp
- + curl -sL -o '/tmp/tmp.n1ODyMd6RD' 'https://rpm.nodesource.com/pub_10.x/el/7/x86_64/nodesource-release-el7-1.noarch.rpm'
- ## Installing release setup RPM...
- + rpm -i --nosignature --force '/tmp/tmp.n1ODyMd6RD'
- ## Cleaning up...
- + rm -f '/tmp/tmp.n1ODyMd6RD'
- ## Checking for existing installations...
- + rpm -qa 'node|npm' | grep -v nodesource
- ## Run `sudo yum install -y nodejs` to install Node.js 10.x and npm.
- ## You may also need development tools to build native addons:
- sudo yum install gcc-c++ make
- ## To install the Yarn package manager, run:
- curl -sL https://dl.yarnpkg.com/rpm/yarn.repo | sudo tee /etc/yum.repos.d/yarn.repo
- sudo yum install yarn
- + sudo yum -y install ansible bind-utils jq libguestfs-tools libvirt libvirt-devel libvirt-daemon-kvm nodejs podman python-ironicclient python-ironic-inspector-client python-lxml python-netaddr python-openstackclient python-virtualbmc qemu-kvm virt-install unzip yarn
- Loaded plugins: fastestmirror, priorities
- Loading mirror speeds from cached hostfile
- * base: centos.mirror.net.in
- * epel: repos.del.extreme-ix.org
- * extras: centos.excellmedia.net
- * updates: centos.mirror.net.in
- 7 packages excluded due to repository priority protections
- Package ansible-2.7.9-1.el7.noarch already installed and latest version
- Package 32:bind-utils-9.9.4-73.el7_6.x86_64 already installed and latest version
- Package jq-1.5-10.el7.x86_64 already installed and latest version
- Package 1:libguestfs-tools-1.38.2-12.el7_6.2.noarch already installed and latest version
- Package libvirt-4.5.0-10.el7_6.6.x86_64 already installed and latest version
- Package libvirt-devel-4.5.0-10.el7_6.6.x86_64 already installed and latest version
- Package libvirt-daemon-kvm-4.5.0-10.el7_6.6.x86_64 already installed and latest version
- Package 2:nodejs-10.15.3-1nodesource.x86_64 already installed and latest version
- Package podman-1.0.0-3.git921f98f.el7.x86_64 already installed and latest version
- Package python2-ironicclient-2.7.0-0.20190310214800.4af8a79.el7.noarch already installed and latest version
- Package python2-ironic-inspector-client-3.5.0-0.20190311120303.9bb1150.el7.noarch already installed and latest version
- Package python-lxml-3.2.1-4.el7.x86_64 already installed and latest version
- Package python-netaddr-0.7.5-9.el7.noarch is obsoleted by python2-netaddr-0.7.19-5.el7.noarch which is already installed
- Package python2-openstackclient-3.18.0-0.20190311121958.6868499.el7.noarch already installed and latest version
- Package python2-virtualbmc-1.5.0-0.20190401094255.bff0e6c.el7.noarch already installed and latest version
- Package 10:qemu-kvm-1.5.3-160.el7_6.1.x86_64 is obsoleted by 10:qemu-kvm-ev-2.12.0-18.el7_6.3.1.x86_64 which is already installed
- Package virt-install-1.5.0-1.el7.noarch already installed and latest version
- Package unzip-6.0-19.el7.x86_64 already installed and latest version
- Package yarn-1.15.2-1.noarch already installed and latest version
- Nothing to do
- + sudo pip install lolcat yq
- Requirement already satisfied (use --upgrade to upgrade): lolcat in /usr/lib/python2.7/site-packages
- Requirement already satisfied (use --upgrade to upgrade): yq in /usr/lib/python2.7/site-packages
- Requirement already satisfied (use --upgrade to upgrade): xmltodict>=0.11.0 in /usr/lib/python2.7/site-packages (from yq)
- Requirement already satisfied (use --upgrade to upgrade): setuptools in /usr/lib/python2.7/site-packages (from yq)
- Requirement already satisfied (use --upgrade to upgrade): PyYAML>=3.11 in /usr/lib64/python2.7/site-packages (from yq)
- You are using pip version 8.1.2, however version 19.0.3 is available.
- You should consider upgrading via the 'pip install --upgrade pip' command.
- + oc_version=4.0.22
- + oc_tools_dir=/home/metal-kuber/oc-4.0.22
- + oc_tools_local_file=openshift-client-4.0.22.tar.gz
- + '[' '!' -f /home/metal-kuber/oc-4.0.22/openshift-client-4.0.22.tar.gz ']'
- + '[' '!' -f /usr/local/bin/terraform ']'
- + '[' '!' -f /home/metal-kuber/.ssh/id_rsa.pub ']'
- + sudo '[' '!' -f /root/.ssh/id_rsa_virt_power ']'
- ./02_configure_host.sh
- + source common.sh
- ++++ dirname common.sh
- +++ cd .
- +++ pwd
- ++ SCRIPTDIR=/home/metal-kuber/dev-scripts
- +++ whoami
- ++ USER=metal-kuber
- ++ '[' -z '' ']'
- ++ '[' -f /home/metal-kuber/dev-scripts/config_metal-kuber.sh ']'
- ++ echo 'Using CONFIG /home/metal-kuber/dev-scripts/config_metal-kuber.sh'
- Using CONFIG /home/metal-kuber/dev-scripts/config_metal-kuber.sh
- ++ CONFIG=/home/metal-kuber/dev-scripts/config_metal-kuber.sh
- ++ source /home/metal-kuber/dev-scripts/config_metal-kuber.sh
- +++ set +x
- ++ ADDN_DNS=
- ++ EXT_IF=
- ++ PRO_IF=
- ++ MANAGE_BR_BRIDGE=y
- ++ INT_IF=
- ++ ROOT_DISK=/dev/vda
- ++ FILESYSTEM=/
- ++ WORKING_DIR=/opt/dev-scripts
- ++ NODES_FILE=/opt/dev-scripts/ironic_nodes.json
- ++ NODES_PLATFORM=baremetal
- ++ MASTER_NODES_FILE=ocp/master_nodes.json
- ++ export RHCOS_IMAGE_URL=https://releases-rhcos.svc.ci.openshift.org/storage/releases/maipo/
- ++ RHCOS_IMAGE_URL=https://releases-rhcos.svc.ci.openshift.org/storage/releases/maipo/
- +++ curl https://releases-rhcos.svc.ci.openshift.org/storage/releases/maipo//builds.json
- +++ jq -r '.builds[0]'
- % Total % Received % Xferd Average Speed Time Time Time Current
- Dload Upload Total Spent Left Speed
- 100 6067 100 6067 0 0 4378 0 0:00:01 0:00:01 --:--:-- 4377
- ++ export RHCOS_LATEST=400.7.20190312.0
- ++ RHCOS_LATEST=400.7.20190312.0
- ++ export RHCOS_IMAGE_VERSION=400.7.20190312.0
- ++ RHCOS_IMAGE_VERSION=400.7.20190312.0
- +++ curl https://releases-rhcos.svc.ci.openshift.org/storage/releases/maipo//400.7.20190312.0/meta.json
- +++ jq -r .images.openstack.path
- % Total % Received % Xferd Average Speed Time Time Time Current
- Dload Upload Total Spent Left Speed
- 100 4539 100 4539 0 0 3358 0 0:00:01 0:00:01 --:--:-- 3359
- ++ export RHCOS_IMAGE_FILENAME_OPENSTACK_GZ=rhcos-maipo-400.7.20190312.0-openstack.qcow2.gz
- ++ RHCOS_IMAGE_FILENAME_OPENSTACK_GZ=rhcos-maipo-400.7.20190312.0-openstack.qcow2.gz
- +++ echo rhcos-maipo-400.7.20190312.0-openstack.qcow2.gz
- +++ sed -e 's/-openstack.*//'
- ++ export RHCOS_IMAGE_NAME=rhcos-maipo-400.7.20190312.0
- ++ RHCOS_IMAGE_NAME=rhcos-maipo-400.7.20190312.0
- ++ export RHCOS_IMAGE_FILENAME_OPENSTACK=rhcos-maipo-400.7.20190312.0-openstack.qcow2
- ++ RHCOS_IMAGE_FILENAME_OPENSTACK=rhcos-maipo-400.7.20190312.0-openstack.qcow2
- ++ export RHCOS_IMAGE_FILENAME_DUALDHCP=rhcos-maipo-400.7.20190312.0-dualdhcp.qcow2
- ++ RHCOS_IMAGE_FILENAME_DUALDHCP=rhcos-maipo-400.7.20190312.0-dualdhcp.qcow2
- ++ export RHCOS_IMAGE_FILENAME_LATEST=rhcos-maipo-latest.qcow2
- ++ RHCOS_IMAGE_FILENAME_LATEST=rhcos-maipo-latest.qcow2
- ++ export IRONIC_IMAGE=quay.io/metalkube/metalkube-ironic
- ++ IRONIC_IMAGE=quay.io/metalkube/metalkube-ironic
- ++ export IRONIC_INSPECTOR_IMAGE=quay.io/metalkube/metalkube-ironic-inspector
- ++ IRONIC_INSPECTOR_IMAGE=quay.io/metalkube/metalkube-ironic-inspector
- ++ export IRONIC_DATA_DIR=/opt/dev-scripts/ironic
- ++ IRONIC_DATA_DIR=/opt/dev-scripts/ironic
- ++ export KUBECONFIG=/home/metal-kuber/dev-scripts/ocp/auth/kubeconfig
- ++ KUBECONFIG=/home/metal-kuber/dev-scripts/ocp/auth/kubeconfig
- ++ export 'SSH=ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o ConnectTimeout=5'
- ++ SSH='ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o ConnectTimeout=5'
- ++ export LIBVIRT_DEFAULT_URI=qemu:///system
- ++ LIBVIRT_DEFAULT_URI=qemu:///system
- ++ '[' metal-kuber '!=' root -a '' == /run/user/0 ']'
- ++ sudo -n uptime
- +++ awk -F= '/^ID=/ { print $2 }' /etc/os-release
- +++ tr -d '"'
- ++ [[ ! centos =~ ^(centos|rhel)$ ]]
- +++ awk -F= '/^VERSION_ID=/ { print $2 }' /etc/os-release
- +++ tr -d '"'
- ++ [[ 7 -ne 7 ]]
- +++ df / --output=fstype
- +++ grep -v Type
- ++ FSTYPE=xfs
- ++ case ${FSTYPE} in
- +++ xfs_info /
- +++ grep -q ftype=1
- ++ [[ -n '' ]]
- ++ '[' 2758 = 0 ']'
- ++ '[' '!' -d /opt/dev-scripts ']'
- +++ dirname ./02_configure_host.sh
- ++ LOGDIR=./logs
- ++ '[' '!' -d ./logs ']'
- +++ basename ./02_configure_host.sh .sh
- +++ date +%F-%H%M%S
- ++ LOGFILE=./logs/02_configure_host-2019-04-03-173950.log
- ++ echo 'Logging to ./logs/02_configure_host-2019-04-03-173950.log'
- Logging to ./logs/02_configure_host-2019-04-03-173950.log
- ++ exec
- +++ tee ./logs/02_configure_host-2019-04-03-173950.log
- ++ BOOTSTRAP_SSH_READY=500
- + source ocp_install_env.sh
- +++ go env
- ++ eval 'GOARCH="amd64"
- GOBIN=""
- GOCACHE="/home/metal-kuber/.cache/go-build"
- GOEXE=""
- GOFLAGS=""
- GOHOSTARCH="amd64"
- GOHOSTOS="linux"
- GOOS="linux"
- GOPATH="/home/metal-kuber/go"
- GOPROXY=""
- GORACE=""
- GOROOT="/usr/lib/golang"
- GOTMPDIR=""
- GOTOOLDIR="/usr/lib/golang/pkg/tool/linux_amd64"
- GCCGO="gccgo"
- CC="gcc"
- CXX="g++"
- CGO_ENABLED="1"
- GOMOD=""
- CGO_CFLAGS="-g -O2"
- CGO_CPPFLAGS=""
- CGO_CXXFLAGS="-g -O2"
- CGO_FFLAGS="-g -O2"
- CGO_LDFLAGS="-g -O2"
- PKG_CONFIG="pkg-config"
- GOGCCFLAGS="-fPIC -m64 -pthread -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build689166306=/tmp/go-build -gno-record-gcc-switches"'
- +++ GOARCH=amd64
- +++ GOBIN=
- +++ GOCACHE=/home/metal-kuber/.cache/go-build
- +++ GOEXE=
- +++ GOFLAGS=
- +++ GOHOSTARCH=amd64
- +++ GOHOSTOS=linux
- +++ GOOS=linux
- +++ GOPATH=/home/metal-kuber/go
- +++ GOPROXY=
- +++ GORACE=
- +++ GOROOT=/usr/lib/golang
- +++ GOTMPDIR=
- +++ GOTOOLDIR=/usr/lib/golang/pkg/tool/linux_amd64
- +++ GCCGO=gccgo
- +++ CC=gcc
- +++ CXX=g++
- +++ CGO_ENABLED=1
- +++ GOMOD=
- +++ CGO_CFLAGS='-g -O2'
- +++ CGO_CPPFLAGS=
- +++ CGO_CXXFLAGS='-g -O2'
- +++ CGO_FFLAGS='-g -O2'
- +++ CGO_LDFLAGS='-g -O2'
- +++ PKG_CONFIG=pkg-config
- +++ GOGCCFLAGS='-fPIC -m64 -pthread -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build689166306=/tmp/go-build -gno-record-gcc-switches'
- ++ export OPENSHIFT_INSTALL_DATA=/home/metal-kuber/go/src/github.com/openshift-metalkube/kni-installer/data/data
- ++ OPENSHIFT_INSTALL_DATA=/home/metal-kuber/go/src/github.com/openshift-metalkube/kni-installer/data/data
- ++ export BASE_DOMAIN=test.metalkube.org
- ++ BASE_DOMAIN=test.metalkube.org
- ++ export CLUSTER_NAME=ostest
- ++ CLUSTER_NAME=ostest
- ++ export CLUSTER_DOMAIN=ostest.test.metalkube.org
- ++ CLUSTER_DOMAIN=ostest.test.metalkube.org
- +++ cat /home/metal-kuber/.ssh/id_rsa.pub
- ++ export 'SSH_PUB_KEY=ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDZzITSzncuAsbo//8vJUplg20ClWXGtWv2Df2gMLXOKxwioFIeJzRsb8BplXp6Yag6wXxEbiFup8PjfFGAzSjh6aPlNQxQTsoy5zNYEzXMp/FRU7okLIMh8g+j5faOjwsmY5FqTDPhLMEzWN8Riiw4meEBB/dQNVSyLd0hIOhH9swsRlTCMbJ8geND+e9iyU6T+G373IKQlQ3wbx3osA2R/7afl9/B3OOsa/e7SLsX0b1Oh0vOcY0lDvFQXzkPamHfWmWttGi3LP1bkvUh7XdwRz9N/b/1sLTAOZWa78cqECPFsiNG7go9T7ETOsqP46annsX2pX0T1QyCumadJfaj metal-kuber@localhost.localdomain'
- ++ SSH_PUB_KEY='ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDZzITSzncuAsbo//8vJUplg20ClWXGtWv2Df2gMLXOKxwioFIeJzRsb8BplXp6Yag6wXxEbiFup8PjfFGAzSjh6aPlNQxQTsoy5zNYEzXMp/FRU7okLIMh8g+j5faOjwsmY5FqTDPhLMEzWN8Riiw4meEBB/dQNVSyLd0hIOhH9swsRlTCMbJ8geND+e9iyU6T+G373IKQlQ3wbx3osA2R/7afl9/B3OOsa/e7SLsX0b1Oh0vOcY0lDvFQXzkPamHfWmWttGi3LP1bkvUh7XdwRz9N/b/1sLTAOZWa78cqECPFsiNG7go9T7ETOsqP46annsX2pX0T1QyCumadJfaj metal-kuber@localhost.localdomain'
- ++ export EXTERNAL_SUBNET=192.168.111.0/24
- ++ EXTERNAL_SUBNET=192.168.111.0/24
- ++ export SSH_PRIV_KEY=/home/metal-kuber/.ssh/id_rsa
- ++ SSH_PRIV_KEY=/home/metal-kuber/.ssh/id_rsa
- ++ export OPENSHIFT_INSTALL_RELEASE_IMAGE_OVERRIDE=registry.svc.ci.openshift.org/openshift/origin-release:v4.0
- ++ OPENSHIFT_INSTALL_RELEASE_IMAGE_OVERRIDE=registry.svc.ci.openshift.org/openshift/origin-release:v4.0
- + export ANSIBLE_LIBRARY=./library
- + ANSIBLE_LIBRARY=./library
- + ANSIBLE_FORCE_COLOR=true
- + ansible-playbook -e non_root_user=metal-kuber -e working_dir=/opt/dev-scripts -e roles_path=/home/metal-kuber/dev-scripts/roles -e @tripleo-quickstart-config/metalkube-nodes.yml -e local_working_dir=/home/metal-kuber/.quickstart -e virthost=localhost.localdomain -e platform=baremetal -e manage_baremetal=y -e @config/environments/dev_privileged_libvirt.yml -i tripleo-quickstart-config/metalkube-inventory.ini -b -vvv tripleo-quickstart-config/metalkube-setup-playbook.yml
- ansible-playbook 2.7.9
- config file = /etc/ansible/ansible.cfg
- configured module search path = [u'/home/metal-kuber/dev-scripts/library']
- ansible python module location = /usr/lib/python2.7/site-packages/ansible
- executable location = /bin/ansible-playbook
- python version = 2.7.5 (default, Oct 30 2018, 23:45:53) [GCC 4.8.5 20150623 (Red Hat 4.8.5-36)]
- Using /etc/ansible/ansible.cfg as config file
- /home/metal-kuber/dev-scripts/tripleo-quickstart-config/metalkube-inventory.ini did not meet host_list requirements, check plugin documentation if this is unexpected
- /home/metal-kuber/dev-scripts/tripleo-quickstart-config/metalkube-inventory.ini did not meet script requirements, check plugin documentation if this is unexpected
- /home/metal-kuber/dev-scripts/tripleo-quickstart-config/metalkube-inventory.ini did not meet yaml requirements, check plugin documentation if this is unexpected
- Parsed /home/metal-kuber/dev-scripts/tripleo-quickstart-config/metalkube-inventory.ini inventory source with ini plugin
- PLAYBOOK: metalkube-setup-playbook.yml **************************************************************************
- 1 plays in tripleo-quickstart-config/metalkube-setup-playbook.yml
- PLAY [Setup dummy baremetal VMs via tripleo-quickstart] *********************************************************
- TASK [Gathering Facts] ******************************************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/metalkube-setup-playbook.yml:5
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293391.64-128131787305306 `" && echo ansible-tmp-1554293391.64-128131787305306="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293391.64-128131787305306 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/system/setup.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpTP0sl0 TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293391.64-128131787305306/AnsiballZ_setup.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293391.64-128131787305306/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293391.64-128131787305306/AnsiballZ_setup.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-jpvjtshwjpdcauabrlnrvmypmewsgxff; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293391.64-128131787305306/AnsiballZ_setup.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293391.64-128131787305306/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost]
- META: ran handlers
- TASK [parts/kvm : Set CPU vendor] *******************************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:1
- ok: [localhost] => {
- "ansible_facts": {
- "cpu_vendor": "intel"
- },
- "changed": false
- }
- TASK [parts/kvm : Get CPU flags] ********************************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:7
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293392.56-42717377461205 `" && echo ansible-tmp-1554293392.56-42717377461205="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293392.56-42717377461205 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/commands/command.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpOsbSv0 TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293392.56-42717377461205/AnsiballZ_command.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293392.56-42717377461205/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293392.56-42717377461205/AnsiballZ_command.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-krpcrjbumpfhwviargidiqrgymhukvej; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293392.56-42717377461205/AnsiballZ_command.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293392.56-42717377461205/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => {
- "changed": false,
- "cmd": [
- "awk",
- "-F:",
- "/^flags/ {print $2; exit}",
- "/proc/cpuinfo"
- ],
- "delta": "0:00:00.002832",
- "end": "2019-04-03 17:39:52.775004",
- "invocation": {
- "module_args": {
- "_raw_params": "awk -F: '/^flags/ {print $2; exit}' /proc/cpuinfo",
- "_uses_shell": false,
- "argv": null,
- "chdir": null,
- "creates": null,
- "executable": null,
- "removes": null,
- "stdin": null,
- "warn": true
- }
- },
- "rc": 0,
- "start": "2019-04-03 17:39:52.772172",
- "stderr": "",
- "stderr_lines": [],
- "stdout": " fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch epb cat_l3 cdp_l3 intel_pt tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm cqm mpx rdt_a avx512f avx512dq rdseed adx smap clflushopt clwb avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local dtherm ida arat pln pts hwp hwp_act_window hwp_pkg_req",
- "stdout_lines": [
- " fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch epb cat_l3 cdp_l3 intel_pt tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm cqm mpx rdt_a avx512f avx512dq rdseed adx smap clflushopt clwb avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local dtherm ida arat pln pts hwp hwp_act_window hwp_pkg_req"
- ]
- }
- TASK [parts/kvm : Check for nested support] *********************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:12
- ok: [localhost] => {
- "ansible_facts": {
- "cpu_nested_support": true
- },
- "changed": false
- }
- TASK [parts/kvm : Disable nested if not supported] **************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:18
- ok: [localhost] => {
- "ansible_facts": {
- "nested": true
- },
- "changed": false
- }
- TASK [parts/kvm : Remove previous KVM modprobe configs] *********************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:23
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.03-207574337617343 `" && echo ansible-tmp-1554293393.03-207574337617343="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.03-207574337617343 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/commands/command.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmp78hvkH TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.03-207574337617343/AnsiballZ_command.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.03-207574337617343/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.03-207574337617343/AnsiballZ_command.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-debnbmhkamxtgktmyxqpmbqaojqlvjgb; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.03-207574337617343/AnsiballZ_command.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.03-207574337617343/ > /dev/null 2>&1 && sleep 0'
- changed: [localhost] => {
- "changed": true,
- "cmd": "find /etc/modprobe.d/ -type f -print0| xargs -r -0 sed -i '/^options kvm_intel/d;/^options kvm_amd/d'",
- "delta": "0:00:00.009766",
- "end": "2019-04-03 17:39:53.143105",
- "invocation": {
- "module_args": {
- "_raw_params": "find /etc/modprobe.d/ -type f -print0| xargs -r -0 sed -i '/^options kvm_intel/d;/^options kvm_amd/d'",
- "_uses_shell": true,
- "argv": null,
- "chdir": null,
- "creates": null,
- "executable": null,
- "removes": null,
- "stdin": null,
- "warn": true
- }
- },
- "rc": 0,
- "start": "2019-04-03 17:39:53.133339",
- "stderr": "",
- "stderr_lines": [],
- "stdout": "",
- "stdout_lines": []
- }
- TASK [parts/kvm : Configure KVM module] *************************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:28
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990 `" && echo ansible-tmp-1554293393.25-185253836172990="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/files/stat.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmp8H72wX TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/AnsiballZ_stat.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/AnsiballZ_stat.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-yepocjzjtxvxwlejpjnxtprgscfwfvtq; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/AnsiballZ_stat.py'"'"' && sleep 0'
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmp6R1YKI TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/source
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/source && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/files/copy.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpClq9tY TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/AnsiballZ_copy.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/AnsiballZ_copy.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-ltphpeegbnqvuosjvoxtarmkyirqbpjt; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/AnsiballZ_copy.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/ > /dev/null 2>&1 && sleep 0'
- changed: [localhost] => {
- "changed": true,
- "checksum": "f4392f177a6f6fb1cd7ff4ed5dfb4f667f9dcf52",
- "dest": "/etc/modprobe.d/kvm.conf",
- "diff": [],
- "gid": 0,
- "group": "root",
- "invocation": {
- "module_args": {
- "_original_basename": "tmp6R1YKI",
- "attributes": null,
- "backup": false,
- "checksum": "f4392f177a6f6fb1cd7ff4ed5dfb4f667f9dcf52",
- "content": null,
- "delimiter": null,
- "dest": "/etc/modprobe.d/kvm.conf",
- "directory_mode": null,
- "follow": false,
- "force": true,
- "group": null,
- "local_follow": null,
- "mode": null,
- "owner": null,
- "regexp": null,
- "remote_src": null,
- "selevel": null,
- "serole": null,
- "setype": null,
- "seuser": null,
- "src": "/home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/source",
- "unsafe_writes": null,
- "validate": null
- }
- },
- "md5sum": "9893a93e5d7a8aeced9e597a65c927d9",
- "mode": "0644",
- "owner": "root",
- "secontext": "system_u:object_r:modules_conf_t:s0",
- "size": 27,
- "src": "/home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.25-185253836172990/source",
- "state": "file",
- "uid": 0
- }
- TASK [parts/kvm : Fetch current runtime nested setting] *********************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:35
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.71-270450154658575 `" && echo ansible-tmp-1554293393.71-270450154658575="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.71-270450154658575 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/commands/command.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpFDu0xF TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.71-270450154658575/AnsiballZ_command.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.71-270450154658575/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.71-270450154658575/AnsiballZ_command.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-fyuihrdgctpsjsydrsiagpwvqcuyswvl; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.71-270450154658575/AnsiballZ_command.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293393.71-270450154658575/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => {
- "changed": false,
- "cmd": [
- "cat",
- "/sys/module/kvm_intel/parameters/nested"
- ],
- "delta": "0:00:00.002743",
- "end": "2019-04-03 17:39:53.815065",
- "invocation": {
- "module_args": {
- "_raw_params": "cat /sys/module/kvm_intel/parameters/nested",
- "_uses_shell": false,
- "argv": null,
- "chdir": null,
- "creates": null,
- "executable": null,
- "removes": null,
- "stdin": null,
- "warn": true
- }
- },
- "rc": 0,
- "start": "2019-04-03 17:39:53.812322",
- "stderr": "",
- "stderr_lines": [],
- "stdout": "Y",
- "stdout_lines": [
- "Y"
- ]
- }
- TASK [parts/kvm : Set fact for nested to false if virtualization is not enabled] ********************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:41
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [parts/kvm : Check if nested is enabled currently] *********************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:47
- ok: [localhost] => {
- "ansible_facts": {
- "cpu_nested_enabled": true
- },
- "changed": false
- }
- TASK [parts/kvm : Unload KVM module] ****************************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:56
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [parts/kvm : Reload KVM module] ****************************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:62
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [parts/kvm : Fetch current runtime nested setting] *********************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:68
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [parts/kvm : Check again if nested is enabled currently] ***************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:73
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [parts/kvm : Fail when the desired and actual state do not match] ******************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/kvm/tasks/main.yml:78
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [parts/libvirt : Check for older iPXE rpm] *****************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/libvirt/tasks/main.yml:4
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.29-131153682053529 `" && echo ansible-tmp-1554293394.29-131153682053529="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.29-131153682053529 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/commands/command.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpvq77Cl TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.29-131153682053529/AnsiballZ_command.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.29-131153682053529/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.29-131153682053529/AnsiballZ_command.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-afquptiqkozicmondyvccplpozomdodz; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.29-131153682053529/AnsiballZ_command.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.29-131153682053529/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => {
- "changed": false,
- "cmd": "if rpm -q ipxe-roms-qemu-20130517; then\n rpm -e --nodeps ipxe-roms-qemu-20130517 && echo \"rpm removed\";\n fi;",
- "delta": "0:00:00.023339",
- "end": "2019-04-03 17:39:54.413450",
- "invocation": {
- "module_args": {
- "_raw_params": "if rpm -q ipxe-roms-qemu-20130517; then\n rpm -e --nodeps ipxe-roms-qemu-20130517 && echo \"rpm removed\";\n fi;",
- "_uses_shell": true,
- "argv": null,
- "chdir": null,
- "creates": null,
- "executable": null,
- "removes": null,
- "stdin": null,
- "warn": true
- }
- },
- "rc": 0,
- "start": "2019-04-03 17:39:54.390111",
- "stderr": "",
- "stderr_lines": [],
- "stdout": "package ipxe-roms-qemu-20130517 is not installed",
- "stdout_lines": [
- "package ipxe-roms-qemu-20130517 is not installed"
- ]
- }
- TASK [parts/libvirt : Install packages for libvirt] *************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/libvirt/tasks/main.yml:17
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.51-9558954918777 `" && echo ansible-tmp-1554293394.51-9558954918777="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.51-9558954918777 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/packaging/os/yum.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmp0kRJv6 TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.51-9558954918777/AnsiballZ_yum.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.51-9558954918777/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.51-9558954918777/AnsiballZ_yum.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-jgzfsqfcrhoevvqcdhorvsnbhocmqnwt; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.51-9558954918777/AnsiballZ_yum.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293394.51-9558954918777/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => {
- "changed": false,
- "invocation": {
- "module_args": {
- "allow_downgrade": false,
- "autoremove": false,
- "bugfix": false,
- "conf_file": null,
- "disable_excludes": null,
- "disable_gpg_check": false,
- "disable_plugin": [],
- "disablerepo": [],
- "download_only": false,
- "enable_plugin": [],
- "enablerepo": [],
- "exclude": [],
- "install_repoquery": true,
- "installroot": "/",
- "list": null,
- "name": [
- "qemu-kvm",
- "libvirt",
- "libvirt-python",
- "libguestfs-tools",
- "python-lxml",
- "polkit-pkla-compat"
- ],
- "releasever": null,
- "security": false,
- "skip_broken": false,
- "state": "present",
- "update_cache": false,
- "update_only": false,
- "use_backend": "auto",
- "validate_certs": true
- }
- },
- "msg": "",
- "rc": 0,
- "results": [
- "10:qemu-kvm-ev-2.12.0-18.el7_6.3.1.x86_64 providing qemu-kvm is already installed",
- "libvirt-4.5.0-10.el7_6.6.x86_64 providing libvirt is already installed",
- "libvirt-python-4.5.0-1.el7.x86_64 providing libvirt-python is already installed",
- "1:libguestfs-tools-1.38.2-12.el7_6.2.noarch providing libguestfs-tools is already installed",
- "python-lxml-3.2.1-4.el7.x86_64 providing python-lxml is already installed",
- "polkit-pkla-compat-0.1-4.el7.x86_64 providing polkit-pkla-compat is already installed"
- ]
- }
- TASK [parts/libvirt : Start libvirtd] ***************************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/parts/libvirt/tasks/main.yml:23
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.05-170746859456785 `" && echo ansible-tmp-1554293397.05-170746859456785="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.05-170746859456785 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/system/systemd.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpSi8d3R TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.05-170746859456785/AnsiballZ_systemd.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.05-170746859456785/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.05-170746859456785/AnsiballZ_systemd.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-gserfcofagtgewvkvbmtsidugirrhiph; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.05-170746859456785/AnsiballZ_systemd.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.05-170746859456785/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => {
- "changed": false,
- "enabled": true,
- "invocation": {
- "module_args": {
- "daemon_reload": false,
- "enabled": true,
- "force": null,
- "masked": null,
- "name": "libvirtd",
- "no_block": false,
- "scope": null,
- "state": "started",
- "user": null
- }
- },
- "name": "libvirtd",
- "state": "started",
- "status": {
- "ActiveEnterTimestamp": "Wed 2019-04-03 15:43:21 IST",
- "ActiveEnterTimestampMonotonic": "2269338481",
- "ActiveExitTimestamp": "Wed 2019-04-03 15:43:21 IST",
- "ActiveExitTimestampMonotonic": "2269289133",
- "ActiveState": "active",
- "After": "apparmor.service virtlockd.service virtlogd.socket systemd-logind.service virtlockd-admin.socket basic.target local-fs.target remote-fs.target iscsid.service systemd-journald.socket virtlogd-admin.socket virtlogd.service system.slice network.target dbus.service virtlockd.socket systemd-machined.service",
- "AllowIsolate": "no",
- "AmbientCapabilities": "0",
- "AssertResult": "yes",
- "AssertTimestamp": "Wed 2019-04-03 15:43:21 IST",
- "AssertTimestampMonotonic": "2269302592",
- "Before": "multi-user.target shutdown.target libvirt-guests.service",
- "BlockIOAccounting": "no",
- "BlockIOWeight": "18446744073709551615",
- "CPUAccounting": "no",
- "CPUQuotaPerSecUSec": "infinity",
- "CPUSchedulingPolicy": "0",
- "CPUSchedulingPriority": "0",
- "CPUSchedulingResetOnFork": "no",
- "CPUShares": "18446744073709551615",
- "CanIsolate": "no",
- "CanReload": "yes",
- "CanStart": "yes",
- "CanStop": "yes",
- "CapabilityBoundingSet": "18446744073709551615",
- "ConditionResult": "yes",
- "ConditionTimestamp": "Wed 2019-04-03 15:43:21 IST",
- "ConditionTimestampMonotonic": "2269302592",
- "Conflicts": "shutdown.target",
- "ControlGroup": "/system.slice/libvirtd.service",
- "ControlPID": "0",
- "DefaultDependencies": "yes",
- "Delegate": "no",
- "Description": "Virtualization daemon",
- "DevicePolicy": "auto",
- "Documentation": "man:libvirtd(8) https://libvirt.org",
- "EnvironmentFile": "/etc/sysconfig/libvirtd (ignore_errors=yes)",
- "ExecMainCode": "0",
- "ExecMainExitTimestampMonotonic": "0",
- "ExecMainPID": "53732",
- "ExecMainStartTimestamp": "Wed 2019-04-03 15:43:21 IST",
- "ExecMainStartTimestampMonotonic": "2269303232",
- "ExecMainStatus": "0",
- "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }",
- "ExecStart": "{ path=/usr/sbin/libvirtd ; argv[]=/usr/sbin/libvirtd $LIBVIRTD_ARGS ; ignore_errors=no ; start_time=[Wed 2019-04-03 15:43:21 IST] ; stop_time=[n/a] ; pid=53732 ; code=(null) ; status=0/0 }",
- "FailureAction": "none",
- "FileDescriptorStoreMax": "0",
- "FragmentPath": "/usr/lib/systemd/system/libvirtd.service",
- "GuessMainPID": "yes",
- "IOScheduling": "0",
- "Id": "libvirtd.service",
- "IgnoreOnIsolate": "no",
- "IgnoreOnSnapshot": "no",
- "IgnoreSIGPIPE": "yes",
- "InactiveEnterTimestamp": "Wed 2019-04-03 15:43:21 IST",
- "InactiveEnterTimestampMonotonic": "2269301188",
- "InactiveExitTimestamp": "Wed 2019-04-03 15:43:21 IST",
- "InactiveExitTimestampMonotonic": "2269303277",
- "JobTimeoutAction": "none",
- "JobTimeoutUSec": "0",
- "KillMode": "process",
- "KillSignal": "15",
- "LimitAS": "18446744073709551615",
- "LimitCORE": "18446744073709551615",
- "LimitCPU": "18446744073709551615",
- "LimitDATA": "18446744073709551615",
- "LimitFSIZE": "18446744073709551615",
- "LimitLOCKS": "18446744073709551615",
- "LimitMEMLOCK": "65536",
- "LimitMSGQUEUE": "819200",
- "LimitNICE": "0",
- "LimitNOFILE": "8192",
- "LimitNPROC": "2061866",
- "LimitRSS": "18446744073709551615",
- "LimitRTPRIO": "0",
- "LimitRTTIME": "18446744073709551615",
- "LimitSIGPENDING": "2061866",
- "LimitSTACK": "18446744073709551615",
- "LoadState": "loaded",
- "MainPID": "53732",
- "MemoryAccounting": "no",
- "MemoryCurrent": "13926400",
- "MemoryLimit": "18446744073709551615",
- "MountFlags": "0",
- "Names": "libvirtd.service",
- "NeedDaemonReload": "no",
- "Nice": "0",
- "NoNewPrivileges": "no",
- "NonBlocking": "no",
- "NotifyAccess": "main",
- "OOMScoreAdjust": "0",
- "OnFailureJobMode": "replace",
- "PermissionsStartOnly": "no",
- "PrivateDevices": "no",
- "PrivateNetwork": "no",
- "PrivateTmp": "no",
- "ProtectHome": "no",
- "ProtectSystem": "no",
- "RefuseManualStart": "no",
- "RefuseManualStop": "no",
- "RemainAfterExit": "no",
- "Requires": "virtlogd.socket basic.target virtlockd.socket",
- "Restart": "on-failure",
- "RestartUSec": "100ms",
- "Result": "success",
- "RootDirectoryStartOnly": "no",
- "RuntimeDirectoryMode": "0755",
- "SameProcessGroup": "no",
- "SecureBits": "0",
- "SendSIGHUP": "no",
- "SendSIGKILL": "yes",
- "Slice": "system.slice",
- "StandardError": "inherit",
- "StandardInput": "null",
- "StandardOutput": "journal",
- "StartLimitAction": "none",
- "StartLimitBurst": "5",
- "StartLimitInterval": "10000000",
- "StartupBlockIOWeight": "18446744073709551615",
- "StartupCPUShares": "18446744073709551615",
- "StatusErrno": "0",
- "StopWhenUnneeded": "no",
- "SubState": "running",
- "SyslogLevelPrefix": "yes",
- "SyslogPriority": "30",
- "SystemCallErrorNumber": "0",
- "TTYReset": "no",
- "TTYVHangup": "no",
- "TTYVTDisallocate": "no",
- "TasksAccounting": "no",
- "TasksCurrent": "21",
- "TasksMax": "32768",
- "TimeoutStartUSec": "1min 30s",
- "TimeoutStopUSec": "1min 30s",
- "TimerSlackNSec": "50000",
- "Transient": "no",
- "Type": "notify",
- "UMask": "0022",
- "UnitFilePreset": "enabled",
- "UnitFileState": "enabled",
- "WantedBy": "libvirt-guests.service multi-user.target",
- "Wants": "system.slice systemd-machined.service",
- "WatchdogTimestamp": "Wed 2019-04-03 15:43:21 IST",
- "WatchdogTimestampMonotonic": "2269338442",
- "WatchdogUSec": "0"
- }
- }
- TASK [Install OVS dependencies] *********************************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:6
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [environment/setup : Create OVS Bridges] *******************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:11
- skipping: [localhost] => (item={u'bridge': u'provisioning', u'forward_mode': u'bridge', u'name': u'provisioning'}) => {
- "changed": false,
- "item": {
- "bridge": "provisioning",
- "forward_mode": "bridge",
- "name": "provisioning"
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'bridge': u'baremetal', u'domain': u'ostest.test.metalkube.org', u'dhcp_hosts': [{u'ip': u'192.168.111.20', u'name': u'master-0'}, {u'ip': u'192.168.111.21', u'name': u'master-1'}, {u'ip': u'192.168.111.22', u'name': u'master-2'}], u'name': u'baremetal', u'netmask': u'255.255.255.0', u'forward_mode': u'nat', u'dhcp_range': [u'192.168.111.20', u'192.168.111.60'], u'dns': {u'hosts': [{u'ip': u'192.168.111.5', u'hostnames': [u'api']}, {u'ip': u'192.168.111.2', u'hostnames': [u'ns1']}], u'forwarders': [{u'domain': u'apps.ostest.test.metalkube.org', u'addr': u'127.0.0.1'}]}, u'address': u'192.168.111.1', u'nat_port_range': [1024, 65535]}) => {
- "changed": false,
- "item": {
- "address": "192.168.111.1",
- "bridge": "baremetal",
- "dhcp_hosts": [
- {
- "ip": "192.168.111.20",
- "name": "master-0"
- },
- {
- "ip": "192.168.111.21",
- "name": "master-1"
- },
- {
- "ip": "192.168.111.22",
- "name": "master-2"
- }
- ],
- "dhcp_range": [
- "192.168.111.20",
- "192.168.111.60"
- ],
- "dns": {
- "forwarders": [
- {
- "addr": "127.0.0.1",
- "domain": "apps.ostest.test.metalkube.org"
- }
- ],
- "hosts": [
- {
- "hostnames": [
- "api"
- ],
- "ip": "192.168.111.5"
- },
- {
- "hostnames": [
- "ns1"
- ],
- "ip": "192.168.111.2"
- }
- ]
- },
- "domain": "ostest.test.metalkube.org",
- "forward_mode": "nat",
- "name": "baremetal",
- "nat_port_range": [
- 1024,
- 65535
- ],
- "netmask": "255.255.255.0"
- },
- "skip_reason": "Conditional result was False"
- }
- TASK [environment/setup : get a list of MACs to use] ************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:20
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.74-5924442588798 `" && echo ansible-tmp-1554293397.74-5924442588798="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.74-5924442588798 `" ) && sleep 0'
- Using module file /home/metal-kuber/dev-scripts/library/generate_macs.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpQYZ9Y2 TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.74-5924442588798/AnsiballZ_generate_macs.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.74-5924442588798/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.74-5924442588798/AnsiballZ_generate_macs.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-yplolbhceqrfiurohggziliulzpcngri; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.74-5924442588798/AnsiballZ_generate_macs.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293397.74-5924442588798/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => {
- "changed": false,
- "invocation": {
- "module_args": {
- "networks": [
- {
- "bridge": "provisioning",
- "forward_mode": "bridge",
- "name": "provisioning"
- },
- {
- "address": "192.168.111.1",
- "bridge": "baremetal",
- "dhcp_hosts": [
- {
- "ip": "192.168.111.20",
- "name": "master-0"
- },
- {
- "ip": "192.168.111.21",
- "name": "master-1"
- },
- {
- "ip": "192.168.111.22",
- "name": "master-2"
- }
- ],
- "dhcp_range": [
- "192.168.111.20",
- "192.168.111.60"
- ],
- "dns": {
- "forwarders": [
- {
- "addr": "127.0.0.1",
- "domain": "apps.ostest.test.metalkube.org"
- }
- ],
- "hosts": [
- {
- "hostnames": [
- "api"
- ],
- "ip": "192.168.111.5"
- },
- {
- "hostnames": [
- "ns1"
- ],
- "ip": "192.168.111.2"
- }
- ]
- },
- "domain": "ostest.test.metalkube.org",
- "forward_mode": "nat",
- "name": "baremetal",
- "nat_port_range": [
- 1024,
- 65535
- ],
- "netmask": "255.255.255.0"
- }
- ],
- "nodes": [
- {
- "flavor": "openshift_master",
- "name": "openshift_master_0",
- "virtualbmc_port": 6230
- },
- {
- "flavor": "openshift_master",
- "name": "openshift_master_1",
- "virtualbmc_port": 6231
- },
- {
- "flavor": "openshift_master",
- "name": "openshift_master_2",
- "virtualbmc_port": 6232
- },
- {
- "flavor": "openshift_worker",
- "name": "openshift_worker_0",
- "virtualbmc_port": 6233
- },
- {
- "flavor": "openshift_worker",
- "name": "openshift_worker_1",
- "virtualbmc_port": 6234
- },
- {
- "flavor": "openshift_worker",
- "name": "openshift_worker_2",
- "virtualbmc_port": 6235
- }
- ]
- }
- },
- "openshift_master_0": {
- "baremetal": "00:c5:1c:2c:a7:8a",
- "provisioning": "00:c5:1c:2c:a7:88"
- },
- "openshift_master_1": {
- "baremetal": "00:c5:1c:2c:a7:8e",
- "provisioning": "00:c5:1c:2c:a7:8c"
- },
- "openshift_master_2": {
- "baremetal": "00:c5:1c:2c:a7:92",
- "provisioning": "00:c5:1c:2c:a7:90"
- },
- "openshift_worker_0": {
- "baremetal": "00:c5:1c:2c:a7:96",
- "provisioning": "00:c5:1c:2c:a7:94"
- },
- "openshift_worker_1": {
- "baremetal": "00:c5:1c:2c:a7:9a",
- "provisioning": "00:c5:1c:2c:a7:98"
- },
- "openshift_worker_2": {
- "baremetal": "00:c5:1c:2c:a7:9e",
- "provisioning": "00:c5:1c:2c:a7:9c"
- }
- }
- TASK [environment/setup : Create libvirt networks] **************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:30
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.04-117998675446682 `" && echo ansible-tmp-1554293398.04-117998675446682="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.04-117998675446682 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/cloud/misc/virt_net.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmp7IuBEK TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.04-117998675446682/AnsiballZ_virt_net.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.04-117998675446682/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.04-117998675446682/AnsiballZ_virt_net.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-kvqwcaltjfjvyslqmedofpprsxncmfdo; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.04-117998675446682/AnsiballZ_virt_net.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.04-117998675446682/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={u'bridge': u'provisioning', u'forward_mode': u'bridge', u'name': u'provisioning'}) => {
- "changed": false,
- "invocation": {
- "module_args": {
- "autostart": null,
- "command": "define",
- "name": "provisioning",
- "state": "present",
- "uri": "qemu:///system",
- "xml": "<network>\n <name>provisioning</name>\n <bridge name='provisioning'/>\n <forward mode='bridge'>\n </forward>\n</network>\n"
- }
- },
- "item": {
- "bridge": "provisioning",
- "forward_mode": "bridge",
- "name": "provisioning"
- }
- }
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.31-156407347959020 `" && echo ansible-tmp-1554293398.31-156407347959020="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.31-156407347959020 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/cloud/misc/virt_net.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpH9XJHJ TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.31-156407347959020/AnsiballZ_virt_net.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.31-156407347959020/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.31-156407347959020/AnsiballZ_virt_net.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-jxxkoabnkswdmyhzcmzmhtivfzbmfpty; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.31-156407347959020/AnsiballZ_virt_net.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.31-156407347959020/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={u'bridge': u'baremetal', u'domain': u'ostest.test.metalkube.org', u'dhcp_hosts': [{u'ip': u'192.168.111.20', u'name': u'master-0'}, {u'ip': u'192.168.111.21', u'name': u'master-1'}, {u'ip': u'192.168.111.22', u'name': u'master-2'}], u'name': u'baremetal', u'netmask': u'255.255.255.0', u'forward_mode': u'nat', u'dhcp_range': [u'192.168.111.20', u'192.168.111.60'], u'dns': {u'hosts': [{u'ip': u'192.168.111.5', u'hostnames': [u'api']}, {u'ip': u'192.168.111.2', u'hostnames': [u'ns1']}], u'forwarders': [{u'domain': u'apps.ostest.test.metalkube.org', u'addr': u'127.0.0.1'}]}, u'address': u'192.168.111.1', u'nat_port_range': [1024, 65535]}) => {
- "changed": false,
- "invocation": {
- "module_args": {
- "autostart": null,
- "command": "define",
- "name": "baremetal",
- "state": "present",
- "uri": "qemu:///system",
- "xml": "<network>\n <name>baremetal</name>\n <bridge name='baremetal'/>\n <forward mode='nat'>\n <nat>\n <port start='1024' end='65535' />\n </nat>\n </forward>\n <ip address='192.168.111.1' netmask='255.255.255.0'>\n <dhcp>\n <range start='192.168.111.20' end='192.168.111.60'/>\n <host mac='00:c5:1c:2c:a7:8a' name='master-0' ip='192.168.111.20'/>\n <host mac='00:c5:1c:2c:a7:8e' name='master-1' ip='192.168.111.21'/>\n <host mac='00:c5:1c:2c:a7:92' name='master-2' ip='192.168.111.22'/>\n </dhcp>\n </ip>\n <domain name='ostest.test.metalkube.org' localOnly='yes'/>\n <dns>\n <host ip='192.168.111.5'>\n <hostname>api</hostname>\n </host>\n <host ip='192.168.111.2'>\n <hostname>ns1</hostname>\n </host>\n <forwarder domain='apps.ostest.test.metalkube.org' addr='127.0.0.1' />\n </dns>\n</network>\n"
- }
- },
- "item": {
- "address": "192.168.111.1",
- "bridge": "baremetal",
- "dhcp_hosts": [
- {
- "ip": "192.168.111.20",
- "name": "master-0"
- },
- {
- "ip": "192.168.111.21",
- "name": "master-1"
- },
- {
- "ip": "192.168.111.22",
- "name": "master-2"
- }
- ],
- "dhcp_range": [
- "192.168.111.20",
- "192.168.111.60"
- ],
- "dns": {
- "forwarders": [
- {
- "addr": "127.0.0.1",
- "domain": "apps.ostest.test.metalkube.org"
- }
- ],
- "hosts": [
- {
- "hostnames": [
- "api"
- ],
- "ip": "192.168.111.5"
- },
- {
- "hostnames": [
- "ns1"
- ],
- "ip": "192.168.111.2"
- }
- ]
- },
- "domain": "ostest.test.metalkube.org",
- "forward_mode": "nat",
- "name": "baremetal",
- "nat_port_range": [
- 1024,
- 65535
- ],
- "netmask": "255.255.255.0"
- }
- }
- TASK [environment/setup : Start libvirt networks] ***************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:39
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.56-60327409647027 `" && echo ansible-tmp-1554293398.56-60327409647027="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.56-60327409647027 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/cloud/misc/virt_net.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpgTUV3N TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.56-60327409647027/AnsiballZ_virt_net.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.56-60327409647027/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.56-60327409647027/AnsiballZ_virt_net.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-jygpzzdzbmwpvadacrucqerlkqrxdmxs; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.56-60327409647027/AnsiballZ_virt_net.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.56-60327409647027/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={u'bridge': u'provisioning', u'forward_mode': u'bridge', u'name': u'provisioning'}) => {
- "changed": false,
- "invocation": {
- "module_args": {
- "autostart": null,
- "command": "start",
- "name": "provisioning",
- "state": "active",
- "uri": "qemu:///system",
- "xml": null
- }
- },
- "item": {
- "bridge": "provisioning",
- "forward_mode": "bridge",
- "name": "provisioning"
- }
- }
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.71-162536534633249 `" && echo ansible-tmp-1554293398.71-162536534633249="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.71-162536534633249 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/cloud/misc/virt_net.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpGNG5kV TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.71-162536534633249/AnsiballZ_virt_net.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.71-162536534633249/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.71-162536534633249/AnsiballZ_virt_net.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-pvidqxqwkuavwfbzusalddvttsocrnls; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.71-162536534633249/AnsiballZ_virt_net.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.71-162536534633249/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={u'bridge': u'baremetal', u'domain': u'ostest.test.metalkube.org', u'dhcp_hosts': [{u'ip': u'192.168.111.20', u'name': u'master-0'}, {u'ip': u'192.168.111.21', u'name': u'master-1'}, {u'ip': u'192.168.111.22', u'name': u'master-2'}], u'name': u'baremetal', u'netmask': u'255.255.255.0', u'forward_mode': u'nat', u'dhcp_range': [u'192.168.111.20', u'192.168.111.60'], u'dns': {u'hosts': [{u'ip': u'192.168.111.5', u'hostnames': [u'api']}, {u'ip': u'192.168.111.2', u'hostnames': [u'ns1']}], u'forwarders': [{u'domain': u'apps.ostest.test.metalkube.org', u'addr': u'127.0.0.1'}]}, u'address': u'192.168.111.1', u'nat_port_range': [1024, 65535]}) => {
- "changed": false,
- "invocation": {
- "module_args": {
- "autostart": null,
- "command": "start",
- "name": "baremetal",
- "state": "active",
- "uri": "qemu:///system",
- "xml": null
- }
- },
- "item": {
- "address": "192.168.111.1",
- "bridge": "baremetal",
- "dhcp_hosts": [
- {
- "ip": "192.168.111.20",
- "name": "master-0"
- },
- {
- "ip": "192.168.111.21",
- "name": "master-1"
- },
- {
- "ip": "192.168.111.22",
- "name": "master-2"
- }
- ],
- "dhcp_range": [
- "192.168.111.20",
- "192.168.111.60"
- ],
- "dns": {
- "forwarders": [
- {
- "addr": "127.0.0.1",
- "domain": "apps.ostest.test.metalkube.org"
- }
- ],
- "hosts": [
- {
- "hostnames": [
- "api"
- ],
- "ip": "192.168.111.5"
- },
- {
- "hostnames": [
- "ns1"
- ],
- "ip": "192.168.111.2"
- }
- ]
- },
- "domain": "ostest.test.metalkube.org",
- "forward_mode": "nat",
- "name": "baremetal",
- "nat_port_range": [
- 1024,
- 65535
- ],
- "netmask": "255.255.255.0"
- }
- }
- TASK [environment/setup : Mark libvirt networks as autostarted] ************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:47
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.94-131117576762617 `" && echo ansible-tmp-1554293398.94-131117576762617="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.94-131117576762617 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/cloud/misc/virt_net.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmp99Cbjr TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.94-131117576762617/AnsiballZ_virt_net.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.94-131117576762617/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.94-131117576762617/AnsiballZ_virt_net.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-zwwyozjaonxakrnuqnrcvnumjmioqjgl; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.94-131117576762617/AnsiballZ_virt_net.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293398.94-131117576762617/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={u'bridge': u'provisioning', u'forward_mode': u'bridge', u'name': u'provisioning'}) => {
- "changed": false,
- "invocation": {
- "module_args": {
- "autostart": true,
- "command": null,
- "name": "provisioning",
- "state": null,
- "uri": "qemu:///system",
- "xml": null
- }
- },
- "item": {
- "bridge": "provisioning",
- "forward_mode": "bridge",
- "name": "provisioning"
- }
- }
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.09-165092564613142 `" && echo ansible-tmp-1554293399.09-165092564613142="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.09-165092564613142 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/cloud/misc/virt_net.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpV6atrV TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.09-165092564613142/AnsiballZ_virt_net.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.09-165092564613142/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.09-165092564613142/AnsiballZ_virt_net.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-lvkayebtgtrsomcejduowcxdtvmmryxb; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.09-165092564613142/AnsiballZ_virt_net.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.09-165092564613142/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={u'bridge': u'baremetal', u'domain': u'ostest.test.metalkube.org', u'dhcp_hosts': [{u'ip': u'192.168.111.20', u'name': u'master-0'}, {u'ip': u'192.168.111.21', u'name': u'master-1'}, {u'ip': u'192.168.111.22', u'name': u'master-2'}], u'name': u'baremetal', u'netmask': u'255.255.255.0', u'forward_mode': u'nat', u'dhcp_range': [u'192.168.111.20', u'192.168.111.60'], u'dns': {u'hosts': [{u'ip': u'192.168.111.5', u'hostnames': [u'api']}, {u'ip': u'192.168.111.2', u'hostnames': [u'ns1']}], u'forwarders': [{u'domain': u'apps.ostest.test.metalkube.org', u'addr': u'127.0.0.1'}]}, u'address': u'192.168.111.1', u'nat_port_range': [1024, 65535]}) => {
- "changed": false,
- "invocation": {
- "module_args": {
- "autostart": true,
- "command": null,
- "name": "baremetal",
- "state": null,
- "uri": "qemu:///system",
- "xml": null
- }
- },
- "item": {
- "address": "192.168.111.1",
- "bridge": "baremetal",
- "dhcp_hosts": [
- {
- "ip": "192.168.111.20",
- "name": "master-0"
- },
- {
- "ip": "192.168.111.21",
- "name": "master-1"
- },
- {
- "ip": "192.168.111.22",
- "name": "master-2"
- }
- ],
- "dhcp_range": [
- "192.168.111.20",
- "192.168.111.60"
- ],
- "dns": {
- "forwarders": [
- {
- "addr": "127.0.0.1",
- "domain": "apps.ostest.test.metalkube.org"
- }
- ],
- "hosts": [
- {
- "hostnames": [
- "api"
- ],
- "ip": "192.168.111.5"
- },
- {
- "hostnames": [
- "ns1"
- ],
- "ip": "192.168.111.2"
- }
- ]
- },
- "domain": "ostest.test.metalkube.org",
- "forward_mode": "nat",
- "name": "baremetal",
- "nat_port_range": [
- 1024,
- 65535
- ],
- "netmask": "255.255.255.0"
- }
- }
- TASK [environment/setup : Check if "virsh net-autostart" was successful] ****************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:63
- ok: [localhost] => {
- "msg": "Some libvirt networks were not set to autostart. Please see https://bugs.launchpad.net/tripleo-quickstart/+bug/1581676"
- }
- TASK [environment/setup : Get libvirt networks xml] *************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:68
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.39-85525798875923 `" && echo ansible-tmp-1554293399.39-85525798875923="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.39-85525798875923 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/cloud/misc/virt_net.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpmcI7uN TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.39-85525798875923/AnsiballZ_virt_net.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.39-85525798875923/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.39-85525798875923/AnsiballZ_virt_net.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-yjnxqjpsgdzqtherxjnsrlqoirztuqfz; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.39-85525798875923/AnsiballZ_virt_net.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.39-85525798875923/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={u'bridge': u'provisioning', u'forward_mode': u'bridge', u'name': u'provisioning'}) => {
- "changed": false,
- "get_xml": "<network>\n <name>provisioning</name>\n <uuid>964fdb13-ba73-43a1-a739-5e7d1bca9814</uuid>\n <forward mode='bridge'/>\n <bridge name='provisioning'/>\n</network>\n",
- "invocation": {
- "module_args": {
- "autostart": null,
- "command": "get_xml",
- "name": "provisioning",
- "state": null,
- "uri": "qemu:///system",
- "xml": null
- }
- },
- "item": {
- "bridge": "provisioning",
- "forward_mode": "bridge",
- "name": "provisioning"
- }
- }
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.54-104245173320025 `" && echo ansible-tmp-1554293399.54-104245173320025="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.54-104245173320025 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/cloud/misc/virt_net.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmprum0Yg TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.54-104245173320025/AnsiballZ_virt_net.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.54-104245173320025/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.54-104245173320025/AnsiballZ_virt_net.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-oughwamjfemaocwljodmqfmowhxlynfh; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.54-104245173320025/AnsiballZ_virt_net.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.54-104245173320025/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={u'bridge': u'baremetal', u'domain': u'ostest.test.metalkube.org', u'dhcp_hosts': [{u'ip': u'192.168.111.20', u'name': u'master-0'}, {u'ip': u'192.168.111.21', u'name': u'master-1'}, {u'ip': u'192.168.111.22', u'name': u'master-2'}], u'name': u'baremetal', u'netmask': u'255.255.255.0', u'forward_mode': u'nat', u'dhcp_range': [u'192.168.111.20', u'192.168.111.60'], u'dns': {u'hosts': [{u'ip': u'192.168.111.5', u'hostnames': [u'api']}, {u'ip': u'192.168.111.2', u'hostnames': [u'ns1']}], u'forwarders': [{u'domain': u'apps.ostest.test.metalkube.org', u'addr': u'127.0.0.1'}]}, u'address': u'192.168.111.1', u'nat_port_range': [1024, 65535]}) => {
- "changed": false,
- "get_xml": "<network>\n <name>baremetal</name>\n <uuid>4844e3d7-33be-4f3a-ab42-8897e2f78fb5</uuid>\n <forward mode='nat'>\n <nat>\n <port start='1024' end='65535'/>\n </nat>\n </forward>\n <bridge name='baremetal' stp='on' delay='0'/>\n <mac address='52:54:00:6e:18:bd'/>\n <domain name='ostest.test.metalkube.org' localOnly='yes'/>\n <dns>\n <forwarder domain='apps.ostest.test.metalkube.org' addr='127.0.0.1'/>\n <host ip='192.168.111.5'>\n <hostname>api</hostname>\n </host>\n <host ip='192.168.111.2'>\n <hostname>ns1</hostname>\n </host>\n </dns>\n <ip address='192.168.111.1' netmask='255.255.255.0'>\n <dhcp>\n <range start='192.168.111.20' end='192.168.111.60'/>\n <host mac='00:d1:4e:5f:a3:22' name='master-0' ip='192.168.111.20'/>\n <host mac='00:d1:4e:5f:a3:26' name='master-1' ip='192.168.111.21'/>\n <host mac='00:d1:4e:5f:a3:2a' name='master-2' ip='192.168.111.22'/>\n </dhcp>\n </ip>\n</network>\n",
- "invocation": {
- "module_args": {
- "autostart": null,
- "command": "get_xml",
- "name": "baremetal",
- "state": null,
- "uri": "qemu:///system",
- "xml": null
- }
- },
- "item": {
- "address": "192.168.111.1",
- "bridge": "baremetal",
- "dhcp_hosts": [
- {
- "ip": "192.168.111.20",
- "name": "master-0"
- },
- {
- "ip": "192.168.111.21",
- "name": "master-1"
- },
- {
- "ip": "192.168.111.22",
- "name": "master-2"
- }
- ],
- "dhcp_range": [
- "192.168.111.20",
- "192.168.111.60"
- ],
- "dns": {
- "forwarders": [
- {
- "addr": "127.0.0.1",
- "domain": "apps.ostest.test.metalkube.org"
- }
- ],
- "hosts": [
- {
- "hostnames": [
- "api"
- ],
- "ip": "192.168.111.5"
- },
- {
- "hostnames": [
- "ns1"
- ],
- "ip": "192.168.111.2"
- }
- ]
- },
- "domain": "ostest.test.metalkube.org",
- "forward_mode": "nat",
- "name": "baremetal",
- "nat_port_range": [
- 1024,
- 65535
- ],
- "netmask": "255.255.255.0"
- }
- }
- TASK [environment/setup : copy network-xml to file] *************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:77
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.76-54289395109266 `" && echo ansible-tmp-1554293399.76-54289395109266="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.76-54289395109266 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/files/stat.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpnNWyt0 TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.76-54289395109266/AnsiballZ_stat.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.76-54289395109266/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.76-54289395109266/AnsiballZ_stat.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-nzwkbjxrdeisexxgtvrwhfvpsdbjfiep; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.76-54289395109266/AnsiballZ_stat.py'"'"' && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/files/file.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpchYBak TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.76-54289395109266/AnsiballZ_file.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.76-54289395109266/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.76-54289395109266/AnsiballZ_file.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-buvghebbcjlaqwnfffhkmuvuoqcukhyg; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.76-54289395109266/AnsiballZ_file.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293399.76-54289395109266/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={'_ansible_parsed': True, u'get_xml': u"<network>\n <name>provisioning</name>\n <uuid>964fdb13-ba73-43a1-a739-5e7d1bca9814</uuid>\n <forward mode='bridge'/>\n <bridge name='provisioning'/>\n</network>\n", '_ansible_item_result': True, '_ansible_item_label': {u'bridge': u'provisioning', u'forward_mode': u'bridge', u'name': u'provisioning'}, 'failed': False, 'changed': False, 'item': {u'bridge': u'provisioning', u'forward_mode': u'bridge', u'name': u'provisioning'}, u'invocation': {u'module_args': {u'xml': None, u'state': None, u'command': u'get_xml', u'name': u'provisioning', u'autostart': None, u'uri': u'qemu:///system'}}, '_ansible_ignore_errors': None, '_ansible_no_log': False}) => {
- "changed": false,
- "checksum": "c9397be296c9056e0b06734c7225212fae33ccb5",
- "dest": "/tmp/network-provisioning.xml",
- "diff": {
- "after": {
- "path": "/tmp/network-provisioning.xml"
- },
- "before": {
- "path": "/tmp/network-provisioning.xml"
- }
- },
- "gid": 0,
- "group": "root",
- "invocation": {
- "module_args": {
- "_diff_peek": null,
- "_original_basename": "tmpHpdWZL",
- "access_time": null,
- "access_time_format": "%Y%m%d%H%M.%S",
- "attributes": null,
- "backup": null,
- "content": null,
- "delimiter": null,
- "dest": "/tmp/network-provisioning.xml",
- "directory_mode": null,
- "follow": true,
- "force": false,
- "group": null,
- "mode": null,
- "modification_time": null,
- "modification_time_format": "%Y%m%d%H%M.%S",
- "owner": null,
- "path": "/tmp/network-provisioning.xml",
- "recurse": false,
- "regexp": null,
- "remote_src": null,
- "selevel": null,
- "serole": null,
- "setype": null,
- "seuser": null,
- "src": null,
- "state": "file",
- "unsafe_writes": null
- }
- },
- "item": {
- "changed": false,
- "failed": false,
- "get_xml": "<network>\n <name>provisioning</name>\n <uuid>964fdb13-ba73-43a1-a739-5e7d1bca9814</uuid>\n <forward mode='bridge'/>\n <bridge name='provisioning'/>\n</network>\n",
- "invocation": {
- "module_args": {
- "autostart": null,
- "command": "get_xml",
- "name": "provisioning",
- "state": null,
- "uri": "qemu:///system",
- "xml": null
- }
- },
- "item": {
- "bridge": "provisioning",
- "forward_mode": "bridge",
- "name": "provisioning"
- }
- },
- "mode": "0644",
- "owner": "root",
- "path": "/tmp/network-provisioning.xml",
- "secontext": "unconfined_u:object_r:user_tmp_t:s0",
- "size": 160,
- "state": "file",
- "uid": 0
- }
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.08-260234222065324 `" && echo ansible-tmp-1554293400.08-260234222065324="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.08-260234222065324 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/files/stat.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpV2ADwk TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.08-260234222065324/AnsiballZ_stat.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.08-260234222065324/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.08-260234222065324/AnsiballZ_stat.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-xmugtvkriewzdnlwsfqmrxlaksdmfiow; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.08-260234222065324/AnsiballZ_stat.py'"'"' && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/files/file.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpBgWTdl TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.08-260234222065324/AnsiballZ_file.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.08-260234222065324/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.08-260234222065324/AnsiballZ_file.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-kqfwpszwiituwhlmwgvrbfzwrbfkjysj; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.08-260234222065324/AnsiballZ_file.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.08-260234222065324/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={'_ansible_parsed': True, u'get_xml': u"<network>\n <name>baremetal</name>\n <uuid>4844e3d7-33be-4f3a-ab42-8897e2f78fb5</uuid>\n <forward mode='nat'>\n <nat>\n <port start='1024' end='65535'/>\n </nat>\n </forward>\n <bridge name='baremetal' stp='on' delay='0'/>\n <mac address='52:54:00:6e:18:bd'/>\n <domain name='ostest.test.metalkube.org' localOnly='yes'/>\n <dns>\n <forwarder domain='apps.ostest.test.metalkube.org' addr='127.0.0.1'/>\n <host ip='192.168.111.5'>\n <hostname>api</hostname>\n </host>\n <host ip='192.168.111.2'>\n <hostname>ns1</hostname>\n </host>\n </dns>\n <ip address='192.168.111.1' netmask='255.255.255.0'>\n <dhcp>\n <range start='192.168.111.20' end='192.168.111.60'/>\n <host mac='00:d1:4e:5f:a3:22' name='master-0' ip='192.168.111.20'/>\n <host mac='00:d1:4e:5f:a3:26' name='master-1' ip='192.168.111.21'/>\n <host mac='00:d1:4e:5f:a3:2a' name='master-2' ip='192.168.111.22'/>\n </dhcp>\n </ip>\n</network>\n", '_ansible_item_result': True, '_ansible_item_label': {u'bridge': u'baremetal', u'domain': u'ostest.test.metalkube.org', u'dhcp_hosts': [{u'ip': u'192.168.111.20', u'name': u'master-0'}, {u'ip': u'192.168.111.21', u'name': u'master-1'}, {u'ip': u'192.168.111.22', u'name': u'master-2'}], u'name': u'baremetal', u'netmask': u'255.255.255.0', u'forward_mode': u'nat', u'dhcp_range': [u'192.168.111.20', u'192.168.111.60'], u'dns': {u'hosts': [{u'ip': u'192.168.111.5', u'hostnames': [u'api']}, {u'ip': u'192.168.111.2', u'hostnames': [u'ns1']}], u'forwarders': [{u'domain': u'apps.ostest.test.metalkube.org', u'addr': u'127.0.0.1'}]}, u'address': u'192.168.111.1', u'nat_port_range': [1024, 65535]}, 'failed': False, 'changed': False, 'item': {u'bridge': u'baremetal', u'domain': u'ostest.test.metalkube.org', u'dhcp_hosts': [{u'ip': u'192.168.111.20', u'name': u'master-0'}, {u'ip': u'192.168.111.21', u'name': u'master-1'}, {u'ip': u'192.168.111.22', u'name': u'master-2'}], u'name': u'baremetal', u'netmask': u'255.255.255.0', u'forward_mode': u'nat', u'nat_port_range': [1024, 65535], u'dns': {u'hosts': [{u'ip': u'192.168.111.5', u'hostnames': [u'api']}, {u'ip': u'192.168.111.2', u'hostnames': [u'ns1']}], u'forwarders': [{u'domain': u'apps.ostest.test.metalkube.org', u'addr': u'127.0.0.1'}]}, u'address': u'192.168.111.1', u'dhcp_range': [u'192.168.111.20', u'192.168.111.60']}, u'invocation': {u'module_args': {u'xml': None, u'state': None, u'command': u'get_xml', u'name': u'baremetal', u'autostart': None, u'uri': u'qemu:///system'}}, '_ansible_ignore_errors': None, '_ansible_no_log': False}) => {
- "changed": false,
- "checksum": "9a1d944c1aaa1b01434a790e0c5b5bbdeeb623d5",
- "dest": "/tmp/network-baremetal.xml",
- "diff": {
- "after": {
- "path": "/tmp/network-baremetal.xml"
- },
- "before": {
- "path": "/tmp/network-baremetal.xml"
- }
- },
- "gid": 0,
- "group": "root",
- "invocation": {
- "module_args": {
- "_diff_peek": null,
- "_original_basename": "tmpwgCOyM",
- "access_time": null,
- "access_time_format": "%Y%m%d%H%M.%S",
- "attributes": null,
- "backup": null,
- "content": null,
- "delimiter": null,
- "dest": "/tmp/network-baremetal.xml",
- "directory_mode": null,
- "follow": true,
- "force": false,
- "group": null,
- "mode": null,
- "modification_time": null,
- "modification_time_format": "%Y%m%d%H%M.%S",
- "owner": null,
- "path": "/tmp/network-baremetal.xml",
- "recurse": false,
- "regexp": null,
- "remote_src": null,
- "selevel": null,
- "serole": null,
- "setype": null,
- "seuser": null,
- "src": null,
- "state": "file",
- "unsafe_writes": null
- }
- },
- "item": {
- "changed": false,
- "failed": false,
- "get_xml": "<network>\n <name>baremetal</name>\n <uuid>4844e3d7-33be-4f3a-ab42-8897e2f78fb5</uuid>\n <forward mode='nat'>\n <nat>\n <port start='1024' end='65535'/>\n </nat>\n </forward>\n <bridge name='baremetal' stp='on' delay='0'/>\n <mac address='52:54:00:6e:18:bd'/>\n <domain name='ostest.test.metalkube.org' localOnly='yes'/>\n <dns>\n <forwarder domain='apps.ostest.test.metalkube.org' addr='127.0.0.1'/>\n <host ip='192.168.111.5'>\n <hostname>api</hostname>\n </host>\n <host ip='192.168.111.2'>\n <hostname>ns1</hostname>\n </host>\n </dns>\n <ip address='192.168.111.1' netmask='255.255.255.0'>\n <dhcp>\n <range start='192.168.111.20' end='192.168.111.60'/>\n <host mac='00:d1:4e:5f:a3:22' name='master-0' ip='192.168.111.20'/>\n <host mac='00:d1:4e:5f:a3:26' name='master-1' ip='192.168.111.21'/>\n <host mac='00:d1:4e:5f:a3:2a' name='master-2' ip='192.168.111.22'/>\n </dhcp>\n </ip>\n</network>\n",
- "invocation": {
- "module_args": {
- "autostart": null,
- "command": "get_xml",
- "name": "baremetal",
- "state": null,
- "uri": "qemu:///system",
- "xml": null
- }
- },
- "item": {
- "address": "192.168.111.1",
- "bridge": "baremetal",
- "dhcp_hosts": [
- {
- "ip": "192.168.111.20",
- "name": "master-0"
- },
- {
- "ip": "192.168.111.21",
- "name": "master-1"
- },
- {
- "ip": "192.168.111.22",
- "name": "master-2"
- }
- ],
- "dhcp_range": [
- "192.168.111.20",
- "192.168.111.60"
- ],
- "dns": {
- "forwarders": [
- {
- "addr": "127.0.0.1",
- "domain": "apps.ostest.test.metalkube.org"
- }
- ],
- "hosts": [
- {
- "hostnames": [
- "api"
- ],
- "ip": "192.168.111.5"
- },
- {
- "hostnames": [
- "ns1"
- ],
- "ip": "192.168.111.2"
- }
- ]
- },
- "domain": "ostest.test.metalkube.org",
- "forward_mode": "nat",
- "name": "baremetal",
- "nat_port_range": [
- 1024,
- 65535
- ],
- "netmask": "255.255.255.0"
- }
- },
- "mode": "0644",
- "owner": "root",
- "path": "/tmp/network-baremetal.xml",
- "secontext": "unconfined_u:object_r:user_tmp_t:s0",
- "size": 944,
- "state": "file",
- "uid": 0
- }
- TASK [environment/setup : redefine the libvirt networks so the config is written to /etc/libvirt] ***************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:84
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.4-35359221379196 `" && echo ansible-tmp-1554293400.4-35359221379196="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.4-35359221379196 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/commands/command.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpm5PaGF TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.4-35359221379196/AnsiballZ_command.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.4-35359221379196/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.4-35359221379196/AnsiballZ_command.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-ckaruafgyuqlzlhjpxbxxbjktivjamzs; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.4-35359221379196/AnsiballZ_command.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.4-35359221379196/ > /dev/null 2>&1 && sleep 0'
- changed: [localhost] => (item={u'bridge': u'provisioning', u'forward_mode': u'bridge', u'name': u'provisioning'}) => {
- "changed": true,
- "cmd": [
- "virsh",
- "net-define",
- "/tmp/network-provisioning.xml"
- ],
- "delta": "0:00:00.012925",
- "end": "2019-04-03 17:40:00.506894",
- "invocation": {
- "module_args": {
- "_raw_params": "virsh net-define /tmp/network-provisioning.xml",
- "_uses_shell": false,
- "argv": null,
- "chdir": null,
- "creates": null,
- "executable": null,
- "removes": null,
- "stdin": null,
- "warn": true
- }
- },
- "item": {
- "bridge": "provisioning",
- "forward_mode": "bridge",
- "name": "provisioning"
- },
- "rc": 0,
- "start": "2019-04-03 17:40:00.493969",
- "stderr": "",
- "stderr_lines": [],
- "stdout": "Network provisioning defined from /tmp/network-provisioning.xml",
- "stdout_lines": [
- "Network provisioning defined from /tmp/network-provisioning.xml"
- ]
- }
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.55-281087019430520 `" && echo ansible-tmp-1554293400.55-281087019430520="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.55-281087019430520 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/commands/command.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpwZAv5O TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.55-281087019430520/AnsiballZ_command.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.55-281087019430520/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.55-281087019430520/AnsiballZ_command.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-ciuodgknkhmjhqcmaaljvacbwsklzvlb; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.55-281087019430520/AnsiballZ_command.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.55-281087019430520/ > /dev/null 2>&1 && sleep 0'
- changed: [localhost] => (item={u'bridge': u'baremetal', u'domain': u'ostest.test.metalkube.org', u'dhcp_hosts': [{u'ip': u'192.168.111.20', u'name': u'master-0'}, {u'ip': u'192.168.111.21', u'name': u'master-1'}, {u'ip': u'192.168.111.22', u'name': u'master-2'}], u'name': u'baremetal', u'netmask': u'255.255.255.0', u'forward_mode': u'nat', u'dhcp_range': [u'192.168.111.20', u'192.168.111.60'], u'dns': {u'hosts': [{u'ip': u'192.168.111.5', u'hostnames': [u'api']}, {u'ip': u'192.168.111.2', u'hostnames': [u'ns1']}], u'forwarders': [{u'domain': u'apps.ostest.test.metalkube.org', u'addr': u'127.0.0.1'}]}, u'address': u'192.168.111.1', u'nat_port_range': [1024, 65535]}) => {
- "changed": true,
- "cmd": [
- "virsh",
- "net-define",
- "/tmp/network-baremetal.xml"
- ],
- "delta": "0:00:00.019641",
- "end": "2019-04-03 17:40:00.665016",
- "invocation": {
- "module_args": {
- "_raw_params": "virsh net-define /tmp/network-baremetal.xml",
- "_uses_shell": false,
- "argv": null,
- "chdir": null,
- "creates": null,
- "executable": null,
- "removes": null,
- "stdin": null,
- "warn": true
- }
- },
- "item": {
- "address": "192.168.111.1",
- "bridge": "baremetal",
- "dhcp_hosts": [
- {
- "ip": "192.168.111.20",
- "name": "master-0"
- },
- {
- "ip": "192.168.111.21",
- "name": "master-1"
- },
- {
- "ip": "192.168.111.22",
- "name": "master-2"
- }
- ],
- "dhcp_range": [
- "192.168.111.20",
- "192.168.111.60"
- ],
- "dns": {
- "forwarders": [
- {
- "addr": "127.0.0.1",
- "domain": "apps.ostest.test.metalkube.org"
- }
- ],
- "hosts": [
- {
- "hostnames": [
- "api"
- ],
- "ip": "192.168.111.5"
- },
- {
- "hostnames": [
- "ns1"
- ],
- "ip": "192.168.111.2"
- }
- ]
- },
- "domain": "ostest.test.metalkube.org",
- "forward_mode": "nat",
- "name": "baremetal",
- "nat_port_range": [
- 1024,
- 65535
- ],
- "netmask": "255.255.255.0"
- },
- "rc": 0,
- "start": "2019-04-03 17:40:00.645375",
- "stderr": "",
- "stderr_lines": [],
- "stdout": "Network baremetal defined from /tmp/network-baremetal.xml",
- "stdout_lines": [
- "Network baremetal defined from /tmp/network-baremetal.xml"
- ]
- }
- TASK [environment/setup : Mark libvirt networks as autostarted] *************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:90
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.77-272019492065226 `" && echo ansible-tmp-1554293400.77-272019492065226="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.77-272019492065226 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/cloud/misc/virt_net.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpWuv6AG TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.77-272019492065226/AnsiballZ_virt_net.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.77-272019492065226/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.77-272019492065226/AnsiballZ_virt_net.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-zymogcpcwsjoggfckiaejsctiahdcbob; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.77-272019492065226/AnsiballZ_virt_net.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.77-272019492065226/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={u'bridge': u'provisioning', u'forward_mode': u'bridge', u'name': u'provisioning'}) => {
- "changed": false,
- "invocation": {
- "module_args": {
- "autostart": true,
- "command": null,
- "name": "provisioning",
- "state": null,
- "uri": "qemu:///system",
- "xml": null
- }
- },
- "item": {
- "bridge": "provisioning",
- "forward_mode": "bridge",
- "name": "provisioning"
- }
- }
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.94-3561008481955 `" && echo ansible-tmp-1554293400.94-3561008481955="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.94-3561008481955 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/cloud/misc/virt_net.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpE3Sai3 TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.94-3561008481955/AnsiballZ_virt_net.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.94-3561008481955/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.94-3561008481955/AnsiballZ_virt_net.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-fjykhxyejlvhacadpbtocdyevfokhgds; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.94-3561008481955/AnsiballZ_virt_net.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293400.94-3561008481955/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={u'bridge': u'baremetal', u'domain': u'ostest.test.metalkube.org', u'dhcp_hosts': [{u'ip': u'192.168.111.20', u'name': u'master-0'}, {u'ip': u'192.168.111.21', u'name': u'master-1'}, {u'ip': u'192.168.111.22', u'name': u'master-2'}], u'name': u'baremetal', u'netmask': u'255.255.255.0', u'forward_mode': u'nat', u'dhcp_range': [u'192.168.111.20', u'192.168.111.60'], u'dns': {u'hosts': [{u'ip': u'192.168.111.5', u'hostnames': [u'api']}, {u'ip': u'192.168.111.2', u'hostnames': [u'ns1']}], u'forwarders': [{u'domain': u'apps.ostest.test.metalkube.org', u'addr': u'127.0.0.1'}]}, u'address': u'192.168.111.1', u'nat_port_range': [1024, 65535]}) => {
- "changed": false,
- "invocation": {
- "module_args": {
- "autostart": true,
- "command": null,
- "name": "baremetal",
- "state": null,
- "uri": "qemu:///system",
- "xml": null
- }
- },
- "item": {
- "address": "192.168.111.1",
- "bridge": "baremetal",
- "dhcp_hosts": [
- {
- "ip": "192.168.111.20",
- "name": "master-0"
- },
- {
- "ip": "192.168.111.21",
- "name": "master-1"
- },
- {
- "ip": "192.168.111.22",
- "name": "master-2"
- }
- ],
- "dhcp_range": [
- "192.168.111.20",
- "192.168.111.60"
- ],
- "dns": {
- "forwarders": [
- {
- "addr": "127.0.0.1",
- "domain": "apps.ostest.test.metalkube.org"
- }
- ],
- "hosts": [
- {
- "hostnames": [
- "api"
- ],
- "ip": "192.168.111.5"
- },
- {
- "hostnames": [
- "ns1"
- ],
- "ip": "192.168.111.2"
- }
- ]
- },
- "domain": "ostest.test.metalkube.org",
- "forward_mode": "nat",
- "name": "baremetal",
- "nat_port_range": [
- 1024,
- 65535
- ],
- "netmask": "255.255.255.0"
- }
- }
- TASK [environment/setup : Whitelist bridges for unprivileged access] ********************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:103
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.17-130176243958622 `" && echo ansible-tmp-1554293401.17-130176243958622="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.17-130176243958622 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/files/lineinfile.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmprXSuhB TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.17-130176243958622/AnsiballZ_lineinfile.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.17-130176243958622/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.17-130176243958622/AnsiballZ_lineinfile.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-jtsngcfxyqhbknqqneppzxaytygxtpvw; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.17-130176243958622/AnsiballZ_lineinfile.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.17-130176243958622/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={u'bridge': u'provisioning', u'forward_mode': u'bridge', u'name': u'provisioning'}) => {
- "backup": "",
- "changed": false,
- "diff": [
- {
- "after": "",
- "after_header": "/etc/qemu-kvm/bridge.conf (content)",
- "before": "",
- "before_header": "/etc/qemu-kvm/bridge.conf (content)"
- },
- {
- "after_header": "/etc/qemu-kvm/bridge.conf (file attributes)",
- "before_header": "/etc/qemu-kvm/bridge.conf (file attributes)"
- }
- ],
- "invocation": {
- "module_args": {
- "attributes": null,
- "backrefs": false,
- "backup": false,
- "content": null,
- "create": false,
- "delimiter": null,
- "dest": "/etc/qemu-kvm/bridge.conf",
- "directory_mode": null,
- "firstmatch": false,
- "follow": false,
- "force": null,
- "group": null,
- "insertafter": null,
- "insertbefore": null,
- "line": "allow provisioning",
- "mode": null,
- "owner": null,
- "path": "/etc/qemu-kvm/bridge.conf",
- "regexp": null,
- "remote_src": null,
- "selevel": null,
- "serole": null,
- "setype": null,
- "seuser": null,
- "src": null,
- "state": "present",
- "unsafe_writes": null,
- "validate": null
- }
- },
- "item": {
- "bridge": "provisioning",
- "forward_mode": "bridge",
- "name": "provisioning"
- },
- "msg": ""
- }
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.39-184519719075621 `" && echo ansible-tmp-1554293401.39-184519719075621="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.39-184519719075621 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/files/lineinfile.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmp3S0zpp TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.39-184519719075621/AnsiballZ_lineinfile.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.39-184519719075621/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.39-184519719075621/AnsiballZ_lineinfile.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-hwuucfwjjwimexrjehtzfaetqqmfmrdq; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.39-184519719075621/AnsiballZ_lineinfile.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.39-184519719075621/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => (item={u'bridge': u'baremetal', u'domain': u'ostest.test.metalkube.org', u'dhcp_hosts': [{u'ip': u'192.168.111.20', u'name': u'master-0'}, {u'ip': u'192.168.111.21', u'name': u'master-1'}, {u'ip': u'192.168.111.22', u'name': u'master-2'}], u'name': u'baremetal', u'netmask': u'255.255.255.0', u'forward_mode': u'nat', u'dhcp_range': [u'192.168.111.20', u'192.168.111.60'], u'dns': {u'hosts': [{u'ip': u'192.168.111.5', u'hostnames': [u'api']}, {u'ip': u'192.168.111.2', u'hostnames': [u'ns1']}], u'forwarders': [{u'domain': u'apps.ostest.test.metalkube.org', u'addr': u'127.0.0.1'}]}, u'address': u'192.168.111.1', u'nat_port_range': [1024, 65535]}) => {
- "backup": "",
- "changed": false,
- "diff": [
- {
- "after": "",
- "after_header": "/etc/qemu-kvm/bridge.conf (content)",
- "before": "",
- "before_header": "/etc/qemu-kvm/bridge.conf (content)"
- },
- {
- "after_header": "/etc/qemu-kvm/bridge.conf (file attributes)",
- "before_header": "/etc/qemu-kvm/bridge.conf (file attributes)"
- }
- ],
- "invocation": {
- "module_args": {
- "attributes": null,
- "backrefs": false,
- "backup": false,
- "content": null,
- "create": false,
- "delimiter": null,
- "dest": "/etc/qemu-kvm/bridge.conf",
- "directory_mode": null,
- "firstmatch": false,
- "follow": false,
- "force": null,
- "group": null,
- "insertafter": null,
- "insertbefore": null,
- "line": "allow baremetal",
- "mode": null,
- "owner": null,
- "path": "/etc/qemu-kvm/bridge.conf",
- "regexp": null,
- "remote_src": null,
- "selevel": null,
- "serole": null,
- "setype": null,
- "seuser": null,
- "src": null,
- "state": "present",
- "unsafe_writes": null,
- "validate": null
- }
- },
- "item": {
- "address": "192.168.111.1",
- "bridge": "baremetal",
- "dhcp_hosts": [
- {
- "ip": "192.168.111.20",
- "name": "master-0"
- },
- {
- "ip": "192.168.111.21",
- "name": "master-1"
- },
- {
- "ip": "192.168.111.22",
- "name": "master-2"
- }
- ],
- "dhcp_range": [
- "192.168.111.20",
- "192.168.111.60"
- ],
- "dns": {
- "forwarders": [
- {
- "addr": "127.0.0.1",
- "domain": "apps.ostest.test.metalkube.org"
- }
- ],
- "hosts": [
- {
- "hostnames": [
- "api"
- ],
- "ip": "192.168.111.5"
- },
- {
- "hostnames": [
- "ns1"
- ],
- "ip": "192.168.111.2"
- }
- ]
- },
- "domain": "ostest.test.metalkube.org",
- "forward_mode": "nat",
- "name": "baremetal",
- "nat_port_range": [
- 1024,
- 65535
- ],
- "netmask": "255.255.255.0"
- },
- "msg": ""
- }
- TASK [environment/setup : Ensure remote working dir exists] *****************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/environment/setup/tasks/main.yml:112
- <localhost> ESTABLISH LOCAL CONNECTION FOR USER: metal-kuber
- <localhost> EXEC /bin/sh -c 'echo ~metal-kuber && sleep 0'
- <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.58-118663203433722 `" && echo ansible-tmp-1554293401.58-118663203433722="` echo /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.58-118663203433722 `" ) && sleep 0'
- Using module file /usr/lib/python2.7/site-packages/ansible/modules/files/file.py
- <localhost> PUT /home/metal-kuber/.ansible/tmp/ansible-local-35529vbCWty/tmpTL_ryu TO /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.58-118663203433722/AnsiballZ_file.py
- <localhost> EXEC /bin/sh -c 'chmod u+x /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.58-118663203433722/ /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.58-118663203433722/AnsiballZ_file.py && sleep 0'
- <localhost> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-wjnreupnhbshnkdkqytbdtuxchlfcmoy; /usr/bin/python /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.58-118663203433722/AnsiballZ_file.py'"'"' && sleep 0'
- <localhost> EXEC /bin/sh -c 'rm -f -r /home/metal-kuber/.ansible/tmp/ansible-tmp-1554293401.58-118663203433722/ > /dev/null 2>&1 && sleep 0'
- ok: [localhost] => {
- "changed": false,
- "diff": {
- "after": {
- "path": "/opt/dev-scripts"
- },
- "before": {
- "path": "/opt/dev-scripts"
- }
- },
- "gid": 1000,
- "group": "metal-kuber",
- "invocation": {
- "module_args": {
- "_diff_peek": null,
- "_original_basename": null,
- "access_time": null,
- "access_time_format": "%Y%m%d%H%M.%S",
- "attributes": null,
- "backup": null,
- "content": null,
- "delimiter": null,
- "directory_mode": null,
- "follow": true,
- "force": false,
- "group": null,
- "mode": null,
- "modification_time": null,
- "modification_time_format": "%Y%m%d%H%M.%S",
- "owner": null,
- "path": "/opt/dev-scripts",
- "recurse": false,
- "regexp": null,
- "remote_src": null,
- "selevel": null,
- "serole": null,
- "setype": null,
- "seuser": null,
- "src": null,
- "state": "directory",
- "unsafe_writes": null
- }
- },
- "mode": "0755",
- "owner": "metal-kuber",
- "path": "/opt/dev-scripts",
- "secontext": "unconfined_u:object_r:usr_t:s0",
- "size": 80,
- "state": "directory",
- "uid": 1000
- }
- TASK [libvirt/setup/common : ensure libvirt volume path exists] *************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/common/tasks/main.yml:5
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/common : Check volume pool] *****************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/common/tasks/main.yml:12
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/common : create the volume pool xml file] ***************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/common/tasks/main.yml:21
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/common : Define volume pool] ****************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/common/tasks/main.yml:27
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/common : Start volume pool] *****************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/common/tasks/main.yml:33
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/common : ensure tripleo-quickstart volume pool is defined] **********************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/common/tasks/main.yml:42
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/common : Mark volume pool for autostart] ****************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/common/tasks/main.yml:50
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/overcloud : Include vars for libvirt-libvirt-nodepool] **************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/overcloud/tasks/main.yml:1
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/overcloud : ensure python-netaddr] **********************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/overcloud/tasks/main.yml:12
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/overcloud : Check if overcloud volumes exist] ***********************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/overcloud/tasks/main.yml:19
- skipping: [localhost] => (item={u'flavor': u'openshift_master', u'virtualbmc_port': 6230, u'name': u'openshift_master_0'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_master",
- "name": "openshift_master_0",
- "virtualbmc_port": 6230
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_master', u'virtualbmc_port': 6231, u'name': u'openshift_master_1'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_master",
- "name": "openshift_master_1",
- "virtualbmc_port": 6231
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_master', u'virtualbmc_port': 6232, u'name': u'openshift_master_2'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_master",
- "name": "openshift_master_2",
- "virtualbmc_port": 6232
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_worker', u'virtualbmc_port': 6233, u'name': u'openshift_worker_0'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_worker",
- "name": "openshift_worker_0",
- "virtualbmc_port": 6233
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_worker', u'virtualbmc_port': 6234, u'name': u'openshift_worker_1'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_worker",
- "name": "openshift_worker_1",
- "virtualbmc_port": 6234
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_worker', u'virtualbmc_port': 6235, u'name': u'openshift_worker_2'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_worker",
- "name": "openshift_worker_2",
- "virtualbmc_port": 6235
- },
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/overcloud : Create overcloud vm storage] ****************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/overcloud/tasks/main.yml:26
- skipping: [localhost] => (item={'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', '_ansible_item_result': True, 'item': {u'flavor': u'openshift_master', u'virtualbmc_port': 6230, u'name': u'openshift_master_0'}, 'changed': False, '_ansible_ignore_errors': True, '_ansible_item_label': {u'flavor': u'openshift_master', u'virtualbmc_port': 6230, u'name': u'openshift_master_0'}}) => {
- "changed": false,
- "item": {
- "changed": false,
- "item": {
- "flavor": "openshift_master",
- "name": "openshift_master_0",
- "virtualbmc_port": 6230
- },
- "skip_reason": "Conditional result was False",
- "skipped": true
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', '_ansible_item_result': True, 'item': {u'flavor': u'openshift_master', u'virtualbmc_port': 6231, u'name': u'openshift_master_1'}, 'changed': False, '_ansible_ignore_errors': True, '_ansible_item_label': {u'flavor': u'openshift_master', u'virtualbmc_port': 6231, u'name': u'openshift_master_1'}}) => {
- "changed": false,
- "item": {
- "changed": false,
- "item": {
- "flavor": "openshift_master",
- "name": "openshift_master_1",
- "virtualbmc_port": 6231
- },
- "skip_reason": "Conditional result was False",
- "skipped": true
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', '_ansible_item_result': True, 'item': {u'flavor': u'openshift_master', u'virtualbmc_port': 6232, u'name': u'openshift_master_2'}, 'changed': False, '_ansible_ignore_errors': True, '_ansible_item_label': {u'flavor': u'openshift_master', u'virtualbmc_port': 6232, u'name': u'openshift_master_2'}}) => {
- "changed": false,
- "item": {
- "changed": false,
- "item": {
- "flavor": "openshift_master",
- "name": "openshift_master_2",
- "virtualbmc_port": 6232
- },
- "skip_reason": "Conditional result was False",
- "skipped": true
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', '_ansible_item_result': True, 'item': {u'flavor': u'openshift_worker', u'virtualbmc_port': 6233, u'name': u'openshift_worker_0'}, 'changed': False, '_ansible_ignore_errors': True, '_ansible_item_label': {u'flavor': u'openshift_worker', u'virtualbmc_port': 6233, u'name': u'openshift_worker_0'}}) => {
- "changed": false,
- "item": {
- "changed": false,
- "item": {
- "flavor": "openshift_worker",
- "name": "openshift_worker_0",
- "virtualbmc_port": 6233
- },
- "skip_reason": "Conditional result was False",
- "skipped": true
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', '_ansible_item_result': True, 'item': {u'flavor': u'openshift_worker', u'virtualbmc_port': 6234, u'name': u'openshift_worker_1'}, 'changed': False, '_ansible_ignore_errors': True, '_ansible_item_label': {u'flavor': u'openshift_worker', u'virtualbmc_port': 6234, u'name': u'openshift_worker_1'}}) => {
- "changed": false,
- "item": {
- "changed": false,
- "item": {
- "flavor": "openshift_worker",
- "name": "openshift_worker_1",
- "virtualbmc_port": 6234
- },
- "skip_reason": "Conditional result was False",
- "skipped": true
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', '_ansible_item_result': True, 'item': {u'flavor': u'openshift_worker', u'virtualbmc_port': 6235, u'name': u'openshift_worker_2'}, 'changed': False, '_ansible_ignore_errors': True, '_ansible_item_label': {u'flavor': u'openshift_worker', u'virtualbmc_port': 6235, u'name': u'openshift_worker_2'}}) => {
- "changed": false,
- "item": {
- "changed": false,
- "item": {
- "flavor": "openshift_worker",
- "name": "openshift_worker_2",
- "virtualbmc_port": 6235
- },
- "skip_reason": "Conditional result was False",
- "skipped": true
- },
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/overcloud : Define overcloud vms] ***********************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/overcloud/tasks/main.yml:38
- skipping: [localhost] => (item={u'flavor': u'openshift_master', u'virtualbmc_port': 6230, u'name': u'openshift_master_0'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_master",
- "name": "openshift_master_0",
- "virtualbmc_port": 6230
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_master', u'virtualbmc_port': 6231, u'name': u'openshift_master_1'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_master",
- "name": "openshift_master_1",
- "virtualbmc_port": 6231
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_master', u'virtualbmc_port': 6232, u'name': u'openshift_master_2'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_master",
- "name": "openshift_master_2",
- "virtualbmc_port": 6232
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_worker', u'virtualbmc_port': 6233, u'name': u'openshift_worker_0'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_worker",
- "name": "openshift_worker_0",
- "virtualbmc_port": 6233
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_worker', u'virtualbmc_port': 6234, u'name': u'openshift_worker_1'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_worker",
- "name": "openshift_worker_1",
- "virtualbmc_port": 6234
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_worker', u'virtualbmc_port': 6235, u'name': u'openshift_worker_2'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_worker",
- "name": "openshift_worker_2",
- "virtualbmc_port": 6235
- },
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/overcloud : Define overcloud vms] ***********************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/overcloud/tasks/main.yml:47
- skipping: [localhost] => (item={u'flavor': u'openshift_master', u'virtualbmc_port': 6230, u'name': u'openshift_master_0'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_master",
- "name": "openshift_master_0",
- "virtualbmc_port": 6230
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_master', u'virtualbmc_port': 6231, u'name': u'openshift_master_1'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_master",
- "name": "openshift_master_1",
- "virtualbmc_port": 6231
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_master', u'virtualbmc_port': 6232, u'name': u'openshift_master_2'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_master",
- "name": "openshift_master_2",
- "virtualbmc_port": 6232
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_worker', u'virtualbmc_port': 6233, u'name': u'openshift_worker_0'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_worker",
- "name": "openshift_worker_0",
- "virtualbmc_port": 6233
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_worker', u'virtualbmc_port': 6234, u'name': u'openshift_worker_1'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_worker",
- "name": "openshift_worker_1",
- "virtualbmc_port": 6234
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={u'flavor': u'openshift_worker', u'virtualbmc_port': 6235, u'name': u'openshift_worker_2'}) => {
- "changed": false,
- "item": {
- "flavor": "openshift_worker",
- "name": "openshift_worker_2",
- "virtualbmc_port": 6235
- },
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/overcloud : include_tasks] ******************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/overcloud/tasks/main.yml:56
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/overcloud : Create additional blockdevice for objectstorage nodes] **************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/overcloud/tasks/main.yml:61
- skipping: [localhost] => (item=[{u'flavor': u'openshift_master', u'virtualbmc_port': 6230, u'name': u'openshift_master_0'}, u'vdb']) => {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_master",
- "name": "openshift_master_0",
- "virtualbmc_port": 6230
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item=[{u'flavor': u'openshift_master', u'virtualbmc_port': 6231, u'name': u'openshift_master_1'}, u'vdb']) => {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_master",
- "name": "openshift_master_1",
- "virtualbmc_port": 6231
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item=[{u'flavor': u'openshift_master', u'virtualbmc_port': 6232, u'name': u'openshift_master_2'}, u'vdb']) => {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_master",
- "name": "openshift_master_2",
- "virtualbmc_port": 6232
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item=[{u'flavor': u'openshift_worker', u'virtualbmc_port': 6233, u'name': u'openshift_worker_0'}, u'vdb']) => {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_worker",
- "name": "openshift_worker_0",
- "virtualbmc_port": 6233
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item=[{u'flavor': u'openshift_worker', u'virtualbmc_port': 6234, u'name': u'openshift_worker_1'}, u'vdb']) => {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_worker",
- "name": "openshift_worker_1",
- "virtualbmc_port": 6234
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item=[{u'flavor': u'openshift_worker', u'virtualbmc_port': 6235, u'name': u'openshift_worker_2'}, u'vdb']) => {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_worker",
- "name": "openshift_worker_2",
- "virtualbmc_port": 6235
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/overcloud : Check if additional blockdevices are attached] **********************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/overcloud/tasks/main.yml:69
- skipping: [localhost] => (item=[{u'flavor': u'openshift_master', u'virtualbmc_port': 6230, u'name': u'openshift_master_0'}, u'vdb']) => {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_master",
- "name": "openshift_master_0",
- "virtualbmc_port": 6230
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item=[{u'flavor': u'openshift_master', u'virtualbmc_port': 6231, u'name': u'openshift_master_1'}, u'vdb']) => {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_master",
- "name": "openshift_master_1",
- "virtualbmc_port": 6231
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item=[{u'flavor': u'openshift_master', u'virtualbmc_port': 6232, u'name': u'openshift_master_2'}, u'vdb']) => {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_master",
- "name": "openshift_master_2",
- "virtualbmc_port": 6232
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item=[{u'flavor': u'openshift_worker', u'virtualbmc_port': 6233, u'name': u'openshift_worker_0'}, u'vdb']) => {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_worker",
- "name": "openshift_worker_0",
- "virtualbmc_port": 6233
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item=[{u'flavor': u'openshift_worker', u'virtualbmc_port': 6234, u'name': u'openshift_worker_1'}, u'vdb']) => {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_worker",
- "name": "openshift_worker_1",
- "virtualbmc_port": 6234
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item=[{u'flavor': u'openshift_worker', u'virtualbmc_port': 6235, u'name': u'openshift_worker_2'}, u'vdb']) => {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_worker",
- "name": "openshift_worker_2",
- "virtualbmc_port": 6235
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/overcloud : Attach additional blockdevices to overcloud objectstorage VMs] ******************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/overcloud/tasks/main.yml:80
- skipping: [localhost] => (item={'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', '_ansible_item_result': True, 'item': [{u'flavor': u'openshift_master', u'virtualbmc_port': 6230, u'name': u'openshift_master_0'}, u'vdb'], 'changed': False, '_ansible_ignore_errors': True, '_ansible_item_label': [{u'flavor': u'openshift_master', u'virtualbmc_port': 6230, u'name': u'openshift_master_0'}, u'vdb']}) => {
- "changed": false,
- "item": {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_master",
- "name": "openshift_master_0",
- "virtualbmc_port": 6230
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False",
- "skipped": true
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', '_ansible_item_result': True, 'item': [{u'flavor': u'openshift_master', u'virtualbmc_port': 6231, u'name': u'openshift_master_1'}, u'vdb'], 'changed': False, '_ansible_ignore_errors': True, '_ansible_item_label': [{u'flavor': u'openshift_master', u'virtualbmc_port': 6231, u'name': u'openshift_master_1'}, u'vdb']}) => {
- "changed": false,
- "item": {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_master",
- "name": "openshift_master_1",
- "virtualbmc_port": 6231
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False",
- "skipped": true
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', '_ansible_item_result': True, 'item': [{u'flavor': u'openshift_master', u'virtualbmc_port': 6232, u'name': u'openshift_master_2'}, u'vdb'], 'changed': False, '_ansible_ignore_errors': True, '_ansible_item_label': [{u'flavor': u'openshift_master', u'virtualbmc_port': 6232, u'name': u'openshift_master_2'}, u'vdb']}) => {
- "changed": false,
- "item": {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_master",
- "name": "openshift_master_2",
- "virtualbmc_port": 6232
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False",
- "skipped": true
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', '_ansible_item_result': True, 'item': [{u'flavor': u'openshift_worker', u'virtualbmc_port': 6233, u'name': u'openshift_worker_0'}, u'vdb'], 'changed': False, '_ansible_ignore_errors': True, '_ansible_item_label': [{u'flavor': u'openshift_worker', u'virtualbmc_port': 6233, u'name': u'openshift_worker_0'}, u'vdb']}) => {
- "changed": false,
- "item": {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_worker",
- "name": "openshift_worker_0",
- "virtualbmc_port": 6233
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False",
- "skipped": true
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', '_ansible_item_result': True, 'item': [{u'flavor': u'openshift_worker', u'virtualbmc_port': 6234, u'name': u'openshift_worker_1'}, u'vdb'], 'changed': False, '_ansible_ignore_errors': True, '_ansible_item_label': [{u'flavor': u'openshift_worker', u'virtualbmc_port': 6234, u'name': u'openshift_worker_1'}, u'vdb']}) => {
- "changed": false,
- "item": {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_worker",
- "name": "openshift_worker_1",
- "virtualbmc_port": 6234
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False",
- "skipped": true
- },
- "skip_reason": "Conditional result was False"
- }
- skipping: [localhost] => (item={'skipped': True, '_ansible_no_log': False, 'skip_reason': u'Conditional result was False', '_ansible_item_result': True, 'item': [{u'flavor': u'openshift_worker', u'virtualbmc_port': 6235, u'name': u'openshift_worker_2'}, u'vdb'], 'changed': False, '_ansible_ignore_errors': True, '_ansible_item_label': [{u'flavor': u'openshift_worker', u'virtualbmc_port': 6235, u'name': u'openshift_worker_2'}, u'vdb']}) => {
- "changed": false,
- "item": {
- "changed": false,
- "item": [
- {
- "flavor": "openshift_worker",
- "name": "openshift_worker_2",
- "virtualbmc_port": 6235
- },
- "vdb"
- ],
- "skip_reason": "Conditional result was False",
- "skipped": true
- },
- "skip_reason": "Conditional result was False"
- }
- TASK [libvirt/setup/overcloud : Write ironic node json files] ***************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/libvirt/setup/overcloud/tasks/main.yml:90
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- TASK [virtbmc : Include VBMC setup if enabled] ******************************************************************
- task path: /home/metal-kuber/dev-scripts/tripleo-quickstart-config/roles/virtbmc/tasks/main.yml:3
- skipping: [localhost] => {
- "changed": false,
- "skip_reason": "Conditional result was False"
- }
- META: ran handlers
- META: ran handlers
- PLAY RECAP ******************************************************************************************************
- localhost : ok=23 changed=3 unreachable=0 failed=0
- + grep -q libvirt
- + id metal-kuber
- + virsh pool-uuid default
- + '[' '!' -e /etc/sysconfig/network-scripts/ifcfg-provisioning ']'
- + sudo ifdown provisioning
- + sudo ifup provisioning
- + '[' '' ']'
- + '[' '!' -e /etc/sysconfig/network-scripts/ifcfg-baremetal ']'
- + sudo ifdown baremetal
- + sudo ifup baremetal
- + '[' '' ']'
- + '[' y == y ']'
- + sudo virsh net-destroy baremetal
- Network baremetal destroyed
- + sudo virsh net-start baremetal
- Network baremetal started
- + for port in 80 5050 6385
- + sudo iptables -C INPUT -i provisioning -p tcp -m tcp --dport 80 -j ACCEPT
- + for port in 80 5050 6385
- + sudo iptables -C INPUT -i provisioning -p tcp -m tcp --dport 5050 -j ACCEPT
- + for port in 80 5050 6385
- + sudo iptables -C INPUT -i provisioning -p tcp -m tcp --dport 6385 -j ACCEPT
- + sudo iptables -C INPUT -i baremetal -p udp -m udp --dport 6230:6235 -j ACCEPT
- + for port in 67 69
- + sudo iptables -C INPUT -i provisioning -p udp --dport 67 -j ACCEPT
- + for port in 67 69
- + sudo iptables -C INPUT -i provisioning -p udp --dport 69 -j ACCEPT
- + '[' '' ']'
- + sudo iptables -C INPUT -p tcp --dport 8080 -j ACCEPT
- + sudo iptables -C INPUT -p tcp --dport 3000 -j ACCEPT
- + '[' y == y ']'
- + sudo mkdir -p /etc/NetworkManager/conf.d/
- + sudo crudini --set /etc/NetworkManager/conf.d/dnsmasq.conf main dns dnsmasq
- + '[' '' ']'
- + systemctl is-active --quiet NetworkManager
- + sudo systemctl reload NetworkManager
- ./03_ocp_repo_sync.sh
- + source common.sh
- ++++ dirname common.sh
- +++ cd .
- +++ pwd
- ++ SCRIPTDIR=/home/metal-kuber/dev-scripts
- +++ whoami
- ++ USER=metal-kuber
- ++ '[' -z '' ']'
- ++ '[' -f /home/metal-kuber/dev-scripts/config_metal-kuber.sh ']'
- ++ echo 'Using CONFIG /home/metal-kuber/dev-scripts/config_metal-kuber.sh'
- Using CONFIG /home/metal-kuber/dev-scripts/config_metal-kuber.sh
- ++ CONFIG=/home/metal-kuber/dev-scripts/config_metal-kuber.sh
- ++ source /home/metal-kuber/dev-scripts/config_metal-kuber.sh
- +++ set +x
- ++ ADDN_DNS=
- ++ EXT_IF=
- ++ PRO_IF=
- ++ MANAGE_BR_BRIDGE=y
- ++ INT_IF=
- ++ ROOT_DISK=/dev/vda
- ++ FILESYSTEM=/
- ++ WORKING_DIR=/opt/dev-scripts
- ++ NODES_FILE=/opt/dev-scripts/ironic_nodes.json
- ++ NODES_PLATFORM=baremetal
- ++ MASTER_NODES_FILE=ocp/master_nodes.json
- ++ export RHCOS_IMAGE_URL=https://releases-rhcos.svc.ci.openshift.org/storage/releases/maipo/
- ++ RHCOS_IMAGE_URL=https://releases-rhcos.svc.ci.openshift.org/storage/releases/maipo/
- +++ jq -r '.builds[0]'
- +++ curl https://releases-rhcos.svc.ci.openshift.org/storage/releases/maipo//builds.json
- % Total % Received % Xferd Average Speed Time Time Time Current
- Dload Upload Total Spent Left Speed
- 100 6067 100 6067 0 0 4575 0 0:00:01 0:00:01 --:--:-- 4578
- ++ export RHCOS_LATEST=400.7.20190312.0
- ++ RHCOS_LATEST=400.7.20190312.0
- ++ export RHCOS_IMAGE_VERSION=400.7.20190312.0
- ++ RHCOS_IMAGE_VERSION=400.7.20190312.0
- +++ curl https://releases-rhcos.svc.ci.openshift.org/storage/releases/maipo//400.7.20190312.0/meta.json
- +++ jq -r .images.openstack.path
- % Total % Received % Xferd Average Speed Time Time Time Current
- Dload Upload Total Spent Left Speed
- 100 4539 100 4539 0 0 3429 0 0:00:01 0:00:01 --:--:-- 3430
- ++ export RHCOS_IMAGE_FILENAME_OPENSTACK_GZ=rhcos-maipo-400.7.20190312.0-openstack.qcow2.gz
- ++ RHCOS_IMAGE_FILENAME_OPENSTACK_GZ=rhcos-maipo-400.7.20190312.0-openstack.qcow2.gz
- +++ echo rhcos-maipo-400.7.20190312.0-openstack.qcow2.gz
- +++ sed -e 's/-openstack.*//'
- ++ export RHCOS_IMAGE_NAME=rhcos-maipo-400.7.20190312.0
- ++ RHCOS_IMAGE_NAME=rhcos-maipo-400.7.20190312.0
- ++ export RHCOS_IMAGE_FILENAME_OPENSTACK=rhcos-maipo-400.7.20190312.0-openstack.qcow2
- ++ RHCOS_IMAGE_FILENAME_OPENSTACK=rhcos-maipo-400.7.20190312.0-openstack.qcow2
- ++ export RHCOS_IMAGE_FILENAME_DUALDHCP=rhcos-maipo-400.7.20190312.0-dualdhcp.qcow2
- ++ RHCOS_IMAGE_FILENAME_DUALDHCP=rhcos-maipo-400.7.20190312.0-dualdhcp.qcow2
- ++ export RHCOS_IMAGE_FILENAME_LATEST=rhcos-maipo-latest.qcow2
- ++ RHCOS_IMAGE_FILENAME_LATEST=rhcos-maipo-latest.qcow2
- ++ export IRONIC_IMAGE=quay.io/metalkube/metalkube-ironic
- ++ IRONIC_IMAGE=quay.io/metalkube/metalkube-ironic
- ++ export IRONIC_INSPECTOR_IMAGE=quay.io/metalkube/metalkube-ironic-inspector
- ++ IRONIC_INSPECTOR_IMAGE=quay.io/metalkube/metalkube-ironic-inspector
- ++ export IRONIC_DATA_DIR=/opt/dev-scripts/ironic
- ++ IRONIC_DATA_DIR=/opt/dev-scripts/ironic
- ++ export KUBECONFIG=/home/metal-kuber/dev-scripts/ocp/auth/kubeconfig
- ++ KUBECONFIG=/home/metal-kuber/dev-scripts/ocp/auth/kubeconfig
- ++ export 'SSH=ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o ConnectTimeout=5'
- ++ SSH='ssh -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o ConnectTimeout=5'
- ++ export LIBVIRT_DEFAULT_URI=qemu:///system
- ++ LIBVIRT_DEFAULT_URI=qemu:///system
- ++ '[' metal-kuber '!=' root -a '' == /run/user/0 ']'
- ++ sudo -n uptime
- +++ awk -F= '/^ID=/ { print $2 }' /etc/os-release
- +++ tr -d '"'
- ++ [[ ! centos =~ ^(centos|rhel)$ ]]
- +++ awk -F= '/^VERSION_ID=/ { print $2 }' /etc/os-release
- +++ tr -d '"'
- ++ [[ 7 -ne 7 ]]
- +++ df / --output=fstype
- +++ grep -v Type
- ++ FSTYPE=xfs
- ++ case ${FSTYPE} in
- +++ xfs_info /
- +++ grep -q ftype=1
- ++ [[ -n '' ]]
- ++ '[' 2758 = 0 ']'
- ++ '[' '!' -d /opt/dev-scripts ']'
- +++ dirname ./03_ocp_repo_sync.sh
- ++ LOGDIR=./logs
- ++ '[' '!' -d ./logs ']'
- +++ basename ./03_ocp_repo_sync.sh .sh
- +++ date +%F-%H%M%S
- ++ LOGFILE=./logs/03_ocp_repo_sync-2019-04-03-174011.log
- ++ echo 'Logging to ./logs/03_ocp_repo_sync-2019-04-03-174011.log'
- Logging to ./logs/03_ocp_repo_sync-2019-04-03-174011.log
- ++ exec
- +++ tee ./logs/03_ocp_repo_sync-2019-04-03-174011.log
- ++ BOOTSTRAP_SSH_READY=500
- ++ go env
- + eval 'GOARCH="amd64"
- GOBIN=""
- GOCACHE="/home/metal-kuber/.cache/go-build"
- GOEXE=""
- GOFLAGS=""
- GOHOSTARCH="amd64"
- GOHOSTOS="linux"
- GOOS="linux"
- GOPATH="/home/metal-kuber/go"
- GOPROXY=""
- GORACE=""
- GOROOT="/usr/lib/golang"
- GOTMPDIR=""
- GOTOOLDIR="/usr/lib/golang/pkg/tool/linux_amd64"
- GCCGO="gccgo"
- CC="gcc"
- CXX="g++"
- CGO_ENABLED="1"
- GOMOD=""
- CGO_CFLAGS="-g -O2"
- CGO_CPPFLAGS=""
- CGO_CXXFLAGS="-g -O2"
- CGO_FFLAGS="-g -O2"
- CGO_LDFLAGS="-g -O2"
- PKG_CONFIG="pkg-config"
- GOGCCFLAGS="-fPIC -m64 -pthread -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build017946053=/tmp/go-build -gno-record-gcc-switches"'
- ++ GOARCH=amd64
- ++ GOBIN=
- ++ GOCACHE=/home/metal-kuber/.cache/go-build
- ++ GOEXE=
- ++ GOFLAGS=
- ++ GOHOSTARCH=amd64
- ++ GOHOSTOS=linux
- ++ GOOS=linux
- ++ GOPATH=/home/metal-kuber/go
- ++ GOPROXY=
- ++ GORACE=
- ++ GOROOT=/usr/lib/golang
- ++ GOTMPDIR=
- ++ GOTOOLDIR=/usr/lib/golang/pkg/tool/linux_amd64
- ++ GCCGO=gccgo
- ++ CC=gcc
- ++ CXX=g++
- ++ CGO_ENABLED=1
- ++ GOMOD=
- ++ CGO_CFLAGS='-g -O2'
- ++ CGO_CPPFLAGS=
- ++ CGO_CXXFLAGS='-g -O2'
- ++ CGO_FFLAGS='-g -O2'
- ++ CGO_LDFLAGS='-g -O2'
- ++ PKG_CONFIG=pkg-config
- ++ GOGCCFLAGS='-fPIC -m64 -pthread -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build017946053=/tmp/go-build -gno-record-gcc-switches'
- + echo /home/metal-kuber/go
- + lolcat
- /home/metal-kuber/go
- + sync_go_repo_and_patch github.com/openshift-metalkube/kni-installer https://github.com/openshift-metalkube/kni-installer.git
- + DEST=/home/metal-kuber/go/src/github.com/openshift-metalkube/kni-installer
- + figlet 'Syncing github.com/openshift-metalkube/kni-installer'
- + lolcat
- ____ _
- / ___| _ _ _ __ ___(_)_ __ __ _
- \___ \| | | | '_ \ / __| | '_ \ / _` |
- ___) | |_| | | | | (__| | | | | (_| |
- |____/ \__, |_| |_|\___|_|_| |_|\__, |
- |___/ |___/
- _ _ _ _ __
- __ _(_) |_| |__ _ _| |__ ___ ___ _ __ ___ / /__ _ __ ___ _ __
- / _` | | __| '_ \| | | | '_ \ / __/ _ \| '_ ` _ \ / / _ \| '_ \ / _ \ '_ \
- | (_| | | |_| | | | |_| | |_) | (_| (_) | | | | | |/ / (_) | |_) | __/ | | |
- \__, |_|\__|_| |_|\__,_|_.__(_)___\___/|_| |_| |_/_/ \___/| .__/ \___|_| |_|
- |___/ |_|
- _ _ __ _ _ _ _ _ __
- ___| |__ (_)/ _| |_ _ __ ___ ___| |_ __ _| | | ___ _| |__ ___ / /
- / __| '_ \| | |_| __|____| '_ ` _ \ / _ \ __/ _` | | |/ / | | | '_ \ / _ \ / /
- \__ \ | | | | _| ||_____| | | | | | __/ || (_| | | <| |_| | |_) | __// /
- |___/_| |_|_|_| \__| |_| |_| |_|\___|\__\__,_|_|_|\_\\__,_|_.__/ \___/_/
- _ _ _ _ _ _
- | | ___ __ (_) (_)_ __ ___| |_ __ _| | | ___ _ __
- | |/ / '_ \| |_____| | '_ \/ __| __/ _` | | |/ _ \ '__|
- | <| | | | |_____| | | | \__ \ || (_| | | | __/ |
- |_|\_\_| |_|_| |_|_| |_|___/\__\__,_|_|_|\___|_|
- + '[' '!' -d /home/metal-kuber/go/src/github.com/openshift-metalkube/kni-installer ']'
- + pushd /home/metal-kuber/go/src/github.com/openshift-metalkube/kni-installer
- ~/go/src/github.com/openshift-metalkube/kni-installer ~/dev-scripts
- + git am --abort
- Resolve operation not in progress, we are not resuming.
- + true
- + git checkout master
- Switched to branch 'master'
- Your branch is ahead of 'origin/master' by 2 commits.
- (use "git push" to publish your local commits)
- + git pull --rebase origin master
- From https://github.com/openshift-metalkube/kni-installer
- * branch master -> FETCH_HEAD
- Current branch master is up to date.
- + git branch -D we_dont_need_no_stinkin_patches
- Deleted branch we_dont_need_no_stinkin_patches (was a5a5bd3).
- + git checkout -b we_dont_need_no_stinkin_patches
- Switched to a new branch 'we_dont_need_no_stinkin_patches'
- + shift
- + shift
- + popd
- ~/dev-scripts
- + sync_go_repo_and_patch github.com/openshift-metalkube/terraform-provider-ironic https://github.com/openshift-metalkube/terraform-provider-ironic.git
- + DEST=/home/metal-kuber/go/src/github.com/openshift-metalkube/terraform-provider-ironic
- + figlet 'Syncing github.com/openshift-metalkube/terraform-provider-ironic'
- + lolcat
- ____ _
- / ___| _ _ _ __ ___(_)_ __ __ _
- \___ \| | | | '_ \ / __| | '_ \ / _` |
- ___) | |_| | | | | (__| | | | | (_| |
- |____/ \__, |_| |_|\___|_|_| |_|\__, |
- |___/ |___/
- _ _ _ _ __
- __ _(_) |_| |__ _ _| |__ ___ ___ _ __ ___ / /__ _ __ ___ _ __
- / _` | | __| '_ \| | | | '_ \ / __/ _ \| '_ ` _ \ / / _ \| '_ \ / _ \ '_ \
- | (_| | | |_| | | | |_| | |_) | (_| (_) | | | | | |/ / (_) | |_) | __/ | | |
- \__, |_|\__|_| |_|\__,_|_.__(_)___\___/|_| |_| |_/_/ \___/| .__/ \___|_| |_|
- |___/ |_|
- _ _ __ _ _ _ _ _ __
- ___| |__ (_)/ _| |_ _ __ ___ ___| |_ __ _| | | ___ _| |__ ___ / /
- / __| '_ \| | |_| __|____| '_ ` _ \ / _ \ __/ _` | | |/ / | | | '_ \ / _ \ / /
- \__ \ | | | | _| ||_____| | | | | | __/ || (_| | | <| |_| | |_) | __// /
- |___/_| |_|_|_| \__| |_| |_| |_|\___|\__\__,_|_|_|\_\\__,_|_.__/ \___/_/
- _ __ _
- | |_ ___ _ __ _ __ __ _ / _| ___ _ __ _ __ ___ _ __ _ __ _____ _(_)
- | __/ _ \ '__| '__/ _` | |_ / _ \| '__| '_ ` _ \ _____| '_ \| '__/ _ \ \ / / |
- | || __/ | | | | (_| | _| (_) | | | | | | | |_____| |_) | | | (_) \ V /| |
- \__\___|_| |_| \__,_|_| \___/|_| |_| |_| |_| | .__/|_| \___/ \_/ |_|
- |_|
- _ _ _
- __| | ___ _ __ (_)_ __ ___ _ __ (_) ___
- / _` |/ _ \ '__|____| | '__/ _ \| '_ \| |/ __|
- | (_| | __/ | |_____| | | | (_) | | | | | (__
- \__,_|\___|_| |_|_| \___/|_| |_|_|\___|
- + '[' '!' -d /home/metal-kuber/go/src/github.com/openshift-metalkube/terraform-provider-ironic ']'
- + pushd /home/metal-kuber/go/src/github.com/openshift-metalkube/terraform-provider-ironic
- ~/go/src/github.com/openshift-metalkube/terraform-provider-ironic ~/dev-scripts
- + git am --abort
- Resolve operation not in progress, we are not resuming.
- + true
- + git checkout master
- Switched to branch 'master'
- + git pull --rebase origin master
- From https://github.com/openshift-metalkube/terraform-provider-ironic
- * branch master -> FETCH_HEAD
- Current branch master is up to date.
- + git branch -D we_dont_need_no_stinkin_patches
- Deleted branch we_dont_need_no_stinkin_patches (was 7cf70fc).
- + git checkout -b we_dont_need_no_stinkin_patches
- Switched to a new branch 'we_dont_need_no_stinkin_patches'
- + shift
- + shift
- + popd
- ~/dev-scripts
- + sync_go_repo_and_patch github.com/openshift-metalkube/facet https://github.com/openshift-metalkube/facet.git
- + DEST=/home/metal-kuber/go/src/github.com/openshift-metalkube/facet
- + lolcat
- + figlet 'Syncing github.com/openshift-metalkube/facet'
- ____ _
- / ___| _ _ _ __ ___(_)_ __ __ _
- \___ \| | | | '_ \ / __| | '_ \ / _` |
- ___) | |_| | | | | (__| | | | | (_| |
- |____/ \__, |_| |_|\___|_|_| |_|\__, |
- |___/ |___/
- _ _ _ _ __
- __ _(_) |_| |__ _ _| |__ ___ ___ _ __ ___ / /__ _ __ ___ _ __
- / _` | | __| '_ \| | | | '_ \ / __/ _ \| '_ ` _ \ / / _ \| '_ \ / _ \ '_ \
- | (_| | | |_| | | | |_| | |_) | (_| (_) | | | | | |/ / (_) | |_) | __/ | | |
- \__, |_|\__|_| |_|\__,_|_.__(_)___\___/|_| |_| |_/_/ \___/| .__/ \___|_| |_|
- |___/ |_|
- _ _ __ _ _ _ _ _ __
- ___| |__ (_)/ _| |_ _ __ ___ ___| |_ __ _| | | ___ _| |__ ___ / /
- / __| '_ \| | |_| __|____| '_ ` _ \ / _ \ __/ _` | | |/ / | | | '_ \ / _ \ / /
- \__ \ | | | | _| ||_____| | | | | | __/ || (_| | | <| |_| | |_) | __// /
- |___/_| |_|_|_| \__| |_| |_| |_|\___|\__\__,_|_|_|\_\\__,_|_.__/ \___/_/
- __ _
- / _| __ _ ___ ___| |_
- | |_ / _` |/ __/ _ \ __|
- | _| (_| | (_| __/ |_
- |_| \__,_|\___\___|\__|
- + '[' '!' -d /home/metal-kuber/go/src/github.com/openshift-metalkube/facet ']'
- + pushd /home/metal-kuber/go/src/github.com/openshift-metalkube/facet
- ~/go/src/github.com/openshift-metalkube/facet ~/dev-scripts
- + git am --abort
- Resolve operation not in progress, we are not resuming.
- + true
- + git checkout master
- Switched to branch 'master'
- + git pull --rebase origin master
- From https://github.com/openshift-metalkube/facet
- * branch master -> FETCH_HEAD
- Current branch master is up to date.
- + git branch -D we_dont_need_no_stinkin_patches
- Deleted branch we_dont_need_no_stinkin_patches (was b940bfa).
- + git checkout -b we_dont_need_no_stinkin_patches
- Switched to a new branch 'we_dont_need_no_stinkin_patches'
- + shift
- + shift
- + popd
- ~/dev-scripts
- + go get -v github.com/rakyll/statik
- + pushd /home/metal-kuber/go/src/github.com/openshift-metalkube/facet
- ~/go/src/github.com/openshift-metalkube/facet ~/dev-scripts
- + yarn install
- yarn install v1.15.2
- [1/4] Resolving packages...
- success Already up-to-date.
- Done in 0.53s.
- + ./build.sh
- + yarn build
- yarn run v1.15.2
- $ react-scripts build
- Creating an optimized production build...
- Compiled successfully.
- File sizes after gzip:
- 152.1 KB build/static/js/1.5b6f39a9.chunk.js
- 93.51 KB build/static/css/main.91413470.chunk.css
- 4.37 KB build/static/js/main.93496c7d.chunk.js
- 763 B build/static/js/runtime~main.229c360f.js
- The project was built assuming it is hosted at the server root.
- You can control this with the homepage field in your package.json.
- For example, add this to build it for GitHub Pages:
- "homepage" : "http://myname.github.io/myapp",
- The build folder is ready to be deployed.
- You may serve it with a static server:
- yarn global add serve
- serve -s build
- Find out more about deployment here:
- http://bit.ly/CRA-deploy
- Done in 20.74s.
- + /home/metal-kuber/go/bin/statik -f -src build
- + go build main.go
- + mkdir -p bin
- + mv main bin/facet
- + git checkout statik/statik.go
- + popd
- ~/dev-scripts
- + sh
- + curl https://raw.githubusercontent.com/golang/dep/master/install.sh
- % Total % Received % Xferd Average Speed Time Time Time Current
- Dload Upload Total Spent Left Speed
- 100 5230 100 5230 0 0 866 0 0:00:06 0:00:06 --:--:-- 1484
- ARCH = amd64
- OS = linux
- Will install into /home/metal-kuber/go/bin
- Fetching https://github.com/golang/dep/releases/latest..
- Release Tag = v0.5.1
- Fetching https://github.com/golang/dep/releases/tag/v0.5.1..
- Fetching https://github.com/golang/dep/releases/download/v0.5.1/dep-linux-amd64..
- Setting executable permissions.
- Moving executable to /home/metal-kuber/go/bin/dep
- + export PATH=/home/metal-kuber/go/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/home/metal-kuber/.local/bin:/home/metal-kuber/bin
- + PATH=/home/metal-kuber/go/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/home/metal-kuber/.local/bin:/home/metal-kuber/bin
- + sync_go_repo_and_patch github.com/operator-framework/operator-sdk https://github.com/operator-framework/operator-sdk.git
- + DEST=/home/metal-kuber/go/src/github.com/operator-framework/operator-sdk
- + figlet 'Syncing github.com/operator-framework/operator-sdk'
- + lolcat
- ____ _
- / ___| _ _ _ __ ___(_)_ __ __ _
- \___ \| | | | '_ \ / __| | '_ \ / _` |
- ___) | |_| | | | | (__| | | | | (_| |
- |____/ \__, |_| |_|\___|_|_| |_|\__, |
- |___/ |___/
- _ _ _ _ __
- __ _(_) |_| |__ _ _| |__ ___ ___ _ __ ___ / /__ _ __ ___ _ __
- / _` | | __| '_ \| | | | '_ \ / __/ _ \| '_ ` _ \ / / _ \| '_ \ / _ \ '__|
- | (_| | | |_| | | | |_| | |_) | (_| (_) | | | | | |/ / (_) | |_) | __/ |
- \__, |_|\__|_| |_|\__,_|_.__(_)___\___/|_| |_| |_/_/ \___/| .__/ \___|_|
- |___/ |_|
- _ __ _
- __ _| |_ ___ _ __ / _|_ __ __ _ _ __ ___ _____ _____ _ __| | __
- / _` | __/ _ \| '__|____| |_| '__/ _` | '_ ` _ \ / _ \ \ /\ / / _ \| '__| |/ /
- | (_| | || (_) | | |_____| _| | | (_| | | | | | | __/\ V V / (_) | | | <
- \__,_|\__\___/|_| |_| |_| \__,_|_| |_| |_|\___| \_/\_/ \___/|_| |_|\_\
- __ _ _ _
- / /__ _ __ ___ _ __ __ _| |_ ___ _ __ ___ __| | | __
- / / _ \| '_ \ / _ \ '__/ _` | __/ _ \| '__|____/ __|/ _` | |/ /
- / / (_) | |_) | __/ | | (_| | || (_) | | |_____\__ \ (_| | <
- /_/ \___/| .__/ \___|_| \__,_|\__\___/|_| |___/\__,_|_|\_\
- |_|
- + '[' '!' -d /home/metal-kuber/go/src/github.com/operator-framework/operator-sdk ']'
- + pushd /home/metal-kuber/go/src/github.com/operator-framework/operator-sdk
- ~/go/src/github.com/operator-framework/operator-sdk ~/dev-scripts
- + git am --abort
- Resolve operation not in progress, we are not resuming.
- + true
- + git checkout master
- Already on 'master'
- + git pull --rebase origin master
- From https://github.com/operator-framework/operator-sdk
- * branch master -> FETCH_HEAD
- Current branch master is up to date.
- + git branch -D we_dont_need_no_stinkin_patches
- Deleted branch we_dont_need_no_stinkin_patches (was a81f08f).
- + git checkout -b we_dont_need_no_stinkin_patches
- Switched to a new branch 'we_dont_need_no_stinkin_patches'
- + shift
- + shift
- + popd
- ~/dev-scripts
- + pushd /home/metal-kuber/go/src/github.com/operator-framework/operator-sdk
- ~/go/src/github.com/operator-framework/operator-sdk ~/dev-scripts
- + git checkout master
- Switched to branch 'master'
- + make dep
- make[1]: Entering directory `/home/metal-kuber/go/src/github.com/operator-framework/operator-sdk'
- The following issues were found in Gopkg.toml:
- ✗ unable to deduce repository and source type for "k8s.io/apiextensions-apiserver": unable to read metadata: unable to fetch raw metadata: failed HTTP request to URL "http://k8s.io/apiextensions-apiserver?go-get=1": Get http://k8s.io/apiextensions-apiserver?go-get=1: dial tcp: lookup k8s.io on 127.0.0.1:53: read udp 127.0.0.1:53380->127.0.0.1:53: i/o timeout
- ✗ unable to deduce repository and source type for "k8s.io/kube-openapi": unable to read metadata: unable to fetch raw metadata: failed HTTP request to URL "http://k8s.io/kube-openapi?go-get=1": Get http://k8s.io/kube-openapi?go-get=1: dial tcp: lookup k8s.io on 127.0.0.1:53: read udp 127.0.0.1:53380->127.0.0.1:53: i/o timeout
- ✗ unable to deduce repository and source type for "k8s.io/client-go": unable to read metadata: unable to fetch raw metadata: failed HTTP request to URL "http://k8s.io/client-go?go-get=1": Get http://k8s.io/client-go?go-get=1: dial tcp: lookup k8s.io on 127.0.0.1:53: read udp 127.0.0.1:53380->127.0.0.1:53: i/o timeout
- ✗ unable to deduce repository and source type for "k8s.io/apiserver": unable to read metadata: unable to fetch raw metadata: failed HTTP request to URL "http://k8s.io/apiserver?go-get=1": Get http://k8s.io/apiserver?go-get=1: dial tcp: lookup k8s.io on 127.0.0.1:53: read udp 127.0.0.1:53380->127.0.0.1:53: i/o timeout
- ✗ unable to deduce repository and source type for "k8s.io/apimachinery": unable to read metadata: unable to fetch raw metadata: failed HTTP request to URL "http://k8s.io/apimachinery?go-get=1": Get http://k8s.io/apimachinery?go-get=1: dial tcp: lookup k8s.io on 127.0.0.1:53: read udp 127.0.0.1:53380->127.0.0.1:53: i/o timeout
- ✗ unable to deduce repository and source type for "k8s.io/cli-runtime": unable to read metadata: unable to fetch raw metadata: failed HTTP request to URL "http://k8s.io/cli-runtime?go-get=1": Get http://k8s.io/cli-runtime?go-get=1: dial tcp: lookup k8s.io on 127.0.0.1:53: read udp 127.0.0.1:53380->127.0.0.1:53: i/o timeout
- ✗ unable to deduce repository and source type for "k8s.io/kube-aggregator": unable to read metadata: unable to fetch raw metadata: failed HTTP request to URL "http://k8s.io/kube-aggregator?go-get=1": Get http://k8s.io/kube-aggregator?go-get=1: dial tcp: lookup k8s.io on 127.0.0.1:53: read udp 127.0.0.1:53380->127.0.0.1:53: i/o timeout
- ✗ unable to deduce repository and source type for "k8s.io/kubernetes": unable to read metadata: unable to fetch raw metadata: failed HTTP request to URL "http://k8s.io/kubernetes?go-get=1": Get http://k8s.io/kubernetes?go-get=1: dial tcp: lookup k8s.io on 127.0.0.1:53: read udp 127.0.0.1:53380->127.0.0.1:53: i/o timeout
- ✗ unable to deduce repository and source type for "k8s.io/api": unable to read metadata: unable to fetch raw metadata: failed HTTP request to URL "http://k8s.io/api?go-get=1": Get http://k8s.io/api?go-get=1: dial tcp: lookup k8s.io on 127.0.0.1:53: read udp 127.0.0.1:53380->127.0.0.1:53: i/o timeout
- ✗ unable to deduce repository and source type for "k8s.io/helm": unable to read metadata: unable to fetch raw metadata: failed HTTP request to URL "http://k8s.io/helm?go-get=1": Get http://k8s.io/helm?go-get=1: dial tcp: lookup k8s.io on 127.0.0.1:53: read udp 127.0.0.1:53380->127.0.0.1:53: i/o timeout
- ✗ unable to deduce repository and source type for "sigs.k8s.io/controller-runtime": unable to read metadata: unable to fetch raw metadata: failed HTTP request to URL "http://sigs.k8s.io/controller-runtime?go-get=1": Get http://sigs.k8s.io/controller-runtime?go-get=1: dial tcp: lookup sigs.k8s.io on 127.0.0.1:53: read udp 127.0.0.1:56082->127.0.0.1:53: i/o timeout
- ProjectRoot name validation failed
- make[1]: *** [dep] Error 1
- make[1]: Leaving directory `/home/metal-kuber/go/src/github.com/operator-framework/operator-sdk'
- make: *** [repo_sync] Error 2
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement