X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=mcp%2Fscripts%2Flib.sh;h=3f740589d4977df7c28f02eedebe4d3528f3aaa4;hb=1201b3cc9d52e30d1c19e725d943b2826e3a384c;hp=b3abd21c75bc9051fc727ce5e6150fd75a7cfcb3;hpb=911afead7d1f3ff1cf1b3c44f282e80140d45785;p=fuel.git diff --git a/mcp/scripts/lib.sh b/mcp/scripts/lib.sh index b3abd21c7..3f740589d 100644 --- a/mcp/scripts/lib.sh +++ b/mcp/scripts/lib.sh @@ -1,24 +1,37 @@ +#!/bin/bash # # Library of shell functions # generate_ssh_key() { - [ -f "$SSH_KEY" ] || ssh-keygen -f ${SSH_KEY} -N '' - install -o jenkins -m 0600 ${SSH_KEY} /tmp/ + local user=${USER} + if [ -n "${SUDO_USER}" ] && [ "${SUDO_USER}" != 'root' ]; then + user=${SUDO_USER} + fi + + if [ -f "/tmp/${SSH_KEY}" ]; then + cp "/tmp/${SSH_KEY}" . + ssh-keygen -f "${SSH_KEY}" -y > "${SSH_KEY}.pub" + fi + + [ -f "${SSH_KEY}" ] || ssh-keygen -f "${SSH_KEY}" -N '' + install -o "${user}" -m 0600 "${SSH_KEY}" /tmp/ } get_base_image() { local base_image=$1 mkdir -p images - wget -P /tmp -nc $base_image + wget -P /tmp -N "${base_image}" } cleanup_vms() { # clean up existing nodes for node in $(virsh list --name | grep -P '\w{3}\d{2}'); do - virsh destroy $node - virsh undefine $node + virsh destroy "${node}" + done + for node in $(virsh list --name --all | grep -P '\w{3}\d{2}'); do + virsh undefine --nvram "${node}" done } @@ -27,52 +40,88 @@ prepare_vms() { local base_image=$2 cleanup_vms - get_base_image $base_image + get_base_image "${base_image}" envsubst < user-data.template > user-data.sh for node in "${vnodes[@]}"; do # create/prepare images - ./create-config-drive.sh -k ${SSH_KEY}.pub -u user-data.sh -h ${node} images/mcp_${node}.iso - cp /tmp/${base_image/*\/} images/mcp_${node}.qcow2 - qemu-img resize images/mcp_${node}.qcow2 100G + ./create-config-drive.sh -k "${SSH_KEY}.pub" -u user-data.sh \ + -h "${node}" "images/mcp_${node}.iso" + cp "/tmp/${base_image/*\/}" "images/mcp_${node}.qcow2" + qemu-img resize "images/mcp_${node}.qcow2" 100G done } create_networks() { - # create required networks - for net in pxe mgmt internal public; do - if virsh net-info $net >/dev/null 2>&1; then - virsh net-destroy ${net} - virsh net-undefine ${net} + local -n vnode_networks=$1 + # create required networks, including constant "mcpcontrol" + # FIXME(alav): since we renamed "pxe" to "mcpcontrol", we need to make sure + # we delete the old "pxe" virtual network, or it would cause IP conflicts. + for net in "pxe" "mcpcontrol" "${vnode_networks[@]}"; do + if virsh net-info "${net}" >/dev/null 2>&1; then + virsh net-destroy "${net}" || true + virsh net-undefine "${net}" + fi + # in case of custom network, host should already have the bridge in place + if [ -f "net_${net}.xml" ]; then + virsh net-define "net_${net}.xml" + virsh net-autostart "${net}" + virsh net-start "${net}" fi - virsh net-define net_${net}.xml - virsh net-autostart ${net} - virsh net-start ${net} done } create_vms() { local -n vnodes=$1 local -n vnodes_ram=$2 + local -n vnodes_vcpus=$3 + local -n vnode_networks=$4 + + # AArch64: prepare arch specific arguments + local virt_extra_args="" + if [ "$(uname -i)" = "aarch64" ]; then + # No Cirrus VGA on AArch64, use vga std + virt_extra_args="$virt_extra_args --video=vga" + fi # create vms with specified options for node in "${vnodes[@]}"; do - virt-install --name ${node} --ram ${vnodes_ram[$node]} --vcpus=2 --cpu host --accelerate \ - --network network:pxe,model=virtio \ - --network network:mgmt,model=virtio \ - --network network:internal,model=virtio \ - --network network:public,model=virtio \ - --disk path=$(pwd)/images/mcp_${node}.qcow2,format=qcow2,bus=virtio,cache=none,io=native \ + # prepare network args + net_args=" --network network=mcpcontrol,model=virtio" + if [ "${node}" = "mas01" ]; then + # MaaS node's 3rd interface gets connected to PXE/Admin Bridge + vnode_networks[2]="${vnode_networks[0]}" + fi + for net in "${vnode_networks[@]:1}"; do + net_type="network" + # in case of custom network, host should already have the bridge in place + if [ ! -f "net_${net}.xml" ]; then + net_type="bridge" + fi + net_args="${net_args} --network ${net_type}=${net},model=virtio" + done + + # shellcheck disable=SC2086 + virt-install --name "${node}" \ + --ram "${vnodes_ram[$node]}" --vcpus "${vnodes_vcpus[$node]}" \ + --cpu host-passthrough --accelerate ${net_args} \ + --disk path="$(pwd)/images/mcp_${node}.qcow2",format=qcow2,bus=virtio,cache=none,io=native \ --os-type linux --os-variant none \ --boot hd --vnc --console pty --autostart --noreboot \ - --disk path=$(pwd)/images/mcp_${node}.iso,device=cdrom + --disk path="$(pwd)/images/mcp_${node}.iso",device=cdrom \ + --noautoconsole \ + ${virt_extra_args} done } -update_pxe_network() { - # set static ip address for salt master node - virsh net-update pxe add ip-dhcp-host \ - "" --live +update_mcpcontrol_network() { + # set static ip address for salt master node, MaaS node + local cmac=$(virsh domiflist cfg01 2>&1| awk '/mcpcontrol/ {print $5; exit}') + local amac=$(virsh domiflist mas01 2>&1| awk '/mcpcontrol/ {print $5; exit}') + virsh net-update "mcpcontrol" add ip-dhcp-host \ + "" --live + [ -z "${amac}" ] || virsh net-update "mcpcontrol" add ip-dhcp-host \ + "" --live } start_vms() { @@ -80,7 +129,7 @@ start_vms() { # start vms for node in "${vnodes[@]}"; do - virsh start ${node} + virsh start "${node}" sleep $[RANDOM%5+1] done } @@ -94,8 +143,9 @@ check_connection() { echo '[INFO] Attempting to get into Salt master ...' # wait until ssh on Salt master is available - while (($attempt <= $total_attempts)); do - ssh ${SSH_OPTS} ubuntu@${SALT_MASTER} uptime + while ((attempt <= total_attempts)); do + # shellcheck disable=SC2086 + ssh ${SSH_OPTS} "ubuntu@${SALT_MASTER}" uptime case $? in 0) echo "${attempt}> Success"; break ;; *) echo "${attempt}/${total_attempts}> ssh server ain't ready yet, waiting for ${sleep_time} seconds ..." ;;