#!/bin/bash
+##############################################################################
+# Copyright (c) 2015 Tim Rozet (Red Hat), Dan Radez (Red Hat) and others.
+#
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
# Deploy script to install provisioning server for OPNFV Apex
# author: Dan Radez (dradez@redhat.com)
set -e
##VARIABLES
-if [ "$TERM" != "unknown" ]; then
- reset=$(tput sgr0)
- blue=$(tput setaf 4)
- red=$(tput setaf 1)
- green=$(tput setaf 2)
-else
- reset=""
- blue=""
- red=""
- green=""
-fi
+reset=$(tput sgr0 || echo "")
+blue=$(tput setaf 4 || echo "")
+red=$(tput setaf 1 || echo "")
+green=$(tput setaf 2 || echo "")
vm_index=4
-ha_enabled="TRUE"
+interactive="FALSE"
ping_site="8.8.8.8"
ntp_server="pool.ntp.org"
net_isolation_enabled="TRUE"
post_config="TRUE"
+debug="FALSE"
+nics_cfg=''
declare -i CNT
declare UNDERCLOUD
SSH_OPTIONS=(-o StrictHostKeyChecking=no -o GlobalKnownHostsFile=/dev/null -o UserKnownHostsFile=/dev/null -o LogLevel=error)
DEPLOY_OPTIONS=""
-RESOURCES=/var/opt/opnfv/stack
+RESOURCES=/var/opt/opnfv/images
CONFIG=/var/opt/opnfv
-INSTACKENV=$CONFIG/instackenv.json
OPNFV_NETWORK_TYPES="admin_network private_network public_network storage_network"
# Netmap used to map networks to OVS bridge names
-NET_MAP['admin_network']="brbm"
-NET_MAP['private_network']="brbm1"
-NET_MAP['public_network']="brbm2"
-NET_MAP['storage_network']="brbm3"
-
-##LIBRARIES
-source $CONFIG/lib/common-functions.sh
+NET_MAP['admin_network']="br-admin"
+NET_MAP['private_network']="br-private"
+NET_MAP['public_network']="br-public"
+NET_MAP['storage_network']="br-storage"
##FUNCTIONS
##translates yaml into variables
elif [ "${network}" == 'admin_network' ]; then
echo -e "${red}ERROR: You must enable admin_network and configure it explicitly or use auto-detection${reset}"
exit 1
- elif [ "${network}" == 'public_network' ]; then
+ elif [[ "${network}" == 'public_network' && "$net_isolation_enabled" == "TRUE" ]]; then
echo -e "${red}ERROR: You must enable public_network and configure it explicitly or use auto-detection${reset}"
exit 1
else
done
echo -e "${blue}INFO: Detecting Network Specific settings for: ${enabled_network}${reset}"
# detect network specific settings
- if [ -n $(eval echo \${${network}_optional_settings}) ]; then
- eval "network_specific_settings=\${${enabled_network}_optional_settings}"
- for setting in ${network_specific_settings}; do
- eval "setting_value=\${${enabled_network}_${setting}}"
- if [ -z "${setting_value}" ]; then
- if [ -n "$nic_value" ]; then
- setting_value=$(eval find_${setting} ${nic_value})
+ for setting in $(eval echo \${${enabled_network}_optional_settings}); do
+ eval "setting_value=\${${enabled_network}_${setting}}"
+ if [ -z "${setting_value}" ]; then
+ if [ -n "$nic_value" ]; then
+ setting_value=$(eval find_${setting} ${nic_value})
+ else
+ setting_value=''
+ echo -e "${blue}INFO: Skipping Auto-detection, NIC not specified for ${enabled_network}. Attempting Auto-generation...${reset}"
+ fi
+ if [ -n "$setting_value" ]; then
+ eval "${enabled_network}_${setting}=${setting_value}"
+ echo -e "${blue}INFO: Auto-detection: ${enabled_network}_${setting}: ${setting_value}${reset}"
+ else
+ eval "cidr=\${${enabled_network}_cidr}"
+ if [ -n "$cidr" ]; then
+ setting_value=$(eval generate_${setting} ${cidr})
else
setting_value=''
- echo -e "${blue}INFO: Skipping Auto-detection, NIC not specified for ${enabled_network}. Attempting Auto-generation...${reset}"
+ echo -e "${red}ERROR: Auto-generation failed: required parameter CIDR missing for network ${enabled_network}${reset}"
fi
if [ -n "$setting_value" ]; then
eval "${enabled_network}_${setting}=${setting_value}"
- echo -e "${blue}INFO: Auto-detection: ${enabled_network}_${setting}: ${setting_value}${reset}"
+ echo -e "${blue}INFO: Auto-generated: ${enabled_network}_${setting}: ${setting_value}${reset}"
else
- eval "cidr=\${${enabled_network}_cidr}"
- if [ -n "$cidr" ]; then
- setting_value=$(eval generate_${setting} ${cidr})
- else
- setting_value=''
- echo -e "${red}ERROR: Auto-generation failed: required parameter CIDR missing for network ${enabled_network}${reset}"
- fi
- if [ -n "$setting_value" ]; then
- eval "${enabled_network}_${setting}=${setting_value}"
- echo -e "${blue}INFO: Auto-generated: ${enabled_network}_${setting}: ${setting_value}${reset}"
- else
- echo -e "${red}ERROR: Auto-generation failed: ${setting} not found${reset}"
- exit 1
- fi
+ echo -e "${red}ERROR: Auto-generation failed: ${setting} not found${reset}"
+ exit 1
fi
- else
- echo -e "${blue}INFO: ${enabled_network}_${setting}: ${setting_value}${reset}"
fi
- done
- fi
+ else
+ echo -e "${blue}INFO: ${enabled_network}_${setting}: ${setting_value}${reset}"
+ fi
+ done
done
}
##parses deploy settings yaml into globals and options array
for entry in $inventory; do
if echo $entry | grep -Eo "^nodes_node[0-9]+_" > /dev/null; then
this_node=$(echo $entry | grep -Eo "^nodes_node[0-9]+_")
- if [[ $inventory_list != *"$this_node"* ]]; then
+ if [[ "$inventory_list" != *"$this_node"* ]]; then
inventory_list+="$this_node "
fi
fi
node_total=$node_count
- if [[ "$node_total" -lt 5 && ha_enabled == "TRUE" ]]; then
+ if [[ "$node_total" -lt 5 && ( ha_enabled == "TRUE" || "$ha_enabled" == "true" ) ]]; then
echo -e "${red}ERROR: You must provide at least 5 nodes for HA baremetal deployment${reset}"
exit 1
elif [[ "$node_total" -lt 2 ]]; then
exit 1
fi
- eval $(parse_yaml $INVENTORY_FILE)
+ eval $(parse_yaml $INVENTORY_FILE) || {
+ echo "${red}Failed to parse inventory.yaml. Aborting.${reset}"
+ exit 1
+ }
- instack_env_output="
+ instackenv_output="
{
\"nodes\" : [
node_output="
{
\"pm_password\": \"$(eval echo \${${node}ipmi_pass})\",
- \"pm_type\": \"pxe_ipmitool\",
+ \"pm_type\": \"$(eval echo \${${node}pm_type})\",
\"mac\": [
\"$(eval echo \${${node}mac_address})\"
],
\"pm_addr\": \"$(eval echo \${${node}ipmi_ip})\",
\"capabilities\": \"$(eval echo \${${node}capabilities})\"
"
- instack_env_output+=${node_output}
+ instackenv_output+=${node_output}
if [ $node_count -lt $node_total ]; then
- instack_env_output+=" },"
+ instackenv_output+=" },"
else
- instack_env_output+=" }"
+ instackenv_output+=" }"
fi
done
- instack_env_output+='
+ instackenv_output+='
]
}
'
#Copy instackenv.json to undercloud for baremetal
- echo -e "{blue}Parsed instackenv JSON:\n${instack_env_output}${reset}"
+ echo -e "{blue}Parsed instackenv JSON:\n${instackenv_output}${reset}"
ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
cat > instackenv.json << EOF
-$instack_env_output
+$instackenv_output
EOF
EOI
fi
# ensure networks are configured
- systemctl start openvswitch
+ systemctl status libvirtd || systemctl start libvirtd
+ systemctl status openvswitch || systemctl start openvswitch
# If flat we only use admin network
if [[ "$net_isolation_enabled" == "FALSE" ]]; then
virsh_enabled_networks="admin_network"
- # For baremetal we only need to create/attach instack to admin and public
+ enabled_network_list="admin_network"
+ # For baremetal we only need to create/attach Undercloud to admin and public
elif [ "$virtual" == "FALSE" ]; then
virsh_enabled_networks="admin_network public_network"
else
virsh_enabled_networks=$enabled_network_list
fi
+ # ensure default network is configured correctly
+ libvirt_dir="/usr/share/libvirt/networks"
+ virsh net-list --all | grep default || virsh net-define ${libvirt_dir}/default.xml
+ virsh net-list --all | grep -E "default\s+active" > /dev/null || virsh net-start default
+ virsh net-list --all | grep -E "default\s+active\s+yes" > /dev/null || virsh net-autostart --network default
+
for network in ${OPNFV_NETWORK_TYPES}; do
+ echo "${blue}INFO: Creating Virsh Network: $network & OVS Bridge: ${NET_MAP[$network]}${reset}"
ovs-vsctl list-br | grep ${NET_MAP[$network]} > /dev/null || ovs-vsctl add-br ${NET_MAP[$network]}
- virsh net-list --all | grep ${NET_MAP[$network]} > /dev/null || virsh net-create $CONFIG/${NET_MAP[$network]}-net.xml
- virsh net-list | grep -E "${NET_MAP[$network]}\s+active" > /dev/null || virsh net-start ${NET_MAP[$network]}
+ virsh net-list --all | grep $network > /dev/null || (cat > ${libvirt_dir}/apex-virsh-net.xml && virsh net-define ${libvirt_dir}/apex-virsh-net.xml) << EOF
+<network>
+ <name>$network</name>
+ <forward mode='bridge'/>
+ <bridge name='${NET_MAP[$network]}'/>
+ <virtualport type='openvswitch'/>
+</network>
+EOF
+ if ! (virsh net-list --all | grep $network > /dev/null); then
+ echo "${red}ERROR: unable to create network: ${network}${reset}"
+ exit 1;
+ fi
+ rm -f ${libvirt_dir}/apex-virsh-net.xml &> /dev/null;
+ virsh net-list | grep -E "$network\s+active" > /dev/null || virsh net-start $network
+ virsh net-list | grep -E "$network\s+active\s+yes" > /dev/null || virsh net-autostart --network $network
done
echo -e "${blue}INFO: Bridges set: ${reset}"
if [[ -z "$virtual" || "$virtual" == "FALSE" ]]; then
# bridge interfaces to correct OVS instances for baremetal deployment
for network in ${enabled_network_list}; do
+ if [[ "$network" != "admin_network" && "$network" != "public_network" ]]; then
+ continue
+ fi
this_interface=$(eval echo \${${network}_bridged_interface})
# check if this a bridged interface for this network
if [[ ! -z "$this_interface" || "$this_interface" != "none" ]]; then
fi
# ensure storage pool exists and is started
- virsh pool-list --all | grep default > /dev/null || virsh pool-create $CONFIG/default-pool.xml
- virsh pool-list | grep -Eo "default\s+active" > /dev/null || virsh pool-start default
-
- if virsh net-list | grep default > /dev/null; then
- num_ints_same_subnet=$(ip addr show | grep "inet 192.168.122" | wc -l)
- if [ "$num_ints_same_subnet" -gt 1 ]; then
- virsh net-destroy default
- ##go edit /etc/libvirt/qemu/networks/default.xml
- sed -i 's/192.168.122/192.168.123/g' /etc/libvirt/qemu/networks/default.xml
- sed -i 's/192.168.122/192.168.123/g' instackenv-virt.json
- sleep 5
- virsh net-start default
- virsh net-autostart default
- fi
- fi
+ virsh pool-list --all | grep default > /dev/null || virsh pool-define-as --name default dir --target /var/lib/libvirt/images
+ virsh pool-list | grep -Eo "default\s+active" > /dev/null || (virsh pool-autostart default; virsh pool-start default)
if ! egrep '^flags.*(vmx|svm)' /proc/cpuinfo > /dev/null; then
echo "${red}virtualization extensions not found, kvm kernel module insertion may fail.\n \
##verify vm exists, an has a dhcp lease assigned to it
##params: none
-function setup_instack_vm {
- if ! virsh list --all | grep instack > /dev/null; then
- #virsh vol-create default instack.qcow2.xml
- virsh define $CONFIG/instack.xml
-
- #Upload instack image
- #virsh vol-create default --file instack.qcow2.xml
- virsh vol-create-as default instack.qcow2 30G --format qcow2
+function setup_undercloud_vm {
+ if ! virsh list --all | grep undercloud > /dev/null; then
+ undercloud_nets="default admin_network"
+ if [[ $enabled_network_list =~ "public_network" ]]; then
+ undercloud_nets+=" public_network"
+ fi
+ define_vm undercloud hd 30 "$undercloud_nets"
### this doesn't work for some reason I was getting hangup events so using cp instead
- #virsh vol-upload --pool default --vol instack.qcow2 --file $CONFIG/stack/instack.qcow2
+ #virsh vol-upload --pool default --vol undercloud.qcow2 --file $CONFIG/stack/undercloud.qcow2
#2015-12-05 12:57:20.569+0000: 8755: info : libvirt version: 1.2.8, package: 16.el7_1.5 (CentOS BuildSystem <http://bugs.centos.org>, 2015-11-03-13:56:46, worker1.bsys.centos.org)
#2015-12-05 12:57:20.569+0000: 8755: warning : virKeepAliveTimerInternal:143 : No response from client 0x7ff1e231e630 after 6 keepalive messages in 35 seconds
#2015-12-05 12:57:20.569+0000: 8756: warning : virKeepAliveTimerInternal:143 : No response from client 0x7ff1e231e630 after 6 keepalive messages in 35 seconds
- #error: cannot close volume instack.qcow2
+ #error: cannot close volume undercloud.qcow2
#error: internal error: received hangup / error event on socket
#error: Reconnected to the hypervisor
- instack_dst=/var/lib/libvirt/images/instack.qcow2
- cp -f $RESOURCES/instack.qcow2 $instack_dst
-
- # resize instack machine
- echo "Checking if instack needs to be resized..."
- instack_size=$(LIBGUESTFS_BACKEND=direct virt-filesystems --long -h --all -a $instack_dst |grep device | grep -Eo "[0-9\.]+G" | sed -n 's/\([0-9][0-9]*\).*/\1/p')
- if [ "$instack_size" -lt 30 ]; then
- qemu-img resize /var/lib/libvirt/images/instack.qcow2 +25G
- LIBGUESTFS_BACKEND=direct virt-resize --expand /dev/sda1 $RESOURCES/instack.qcow2 $instack_dst
- LIBGUESTFS_BACKEND=direct virt-customize -a $instack_dst --run-command 'xfs_growfs -d /dev/sda1 || true'
- new_size=$(LIBGUESTFS_BACKEND=direct virt-filesystems --long -h --all -a $instack_dst |grep filesystem | grep -Eo "[0-9\.]+G" | sed -n 's/\([0-9][0-9]*\).*/\1/p')
+ local undercloud_dst=/var/lib/libvirt/images/undercloud.qcow2
+ cp -f $RESOURCES/undercloud.qcow2 $undercloud_dst
+
+ # resize Undercloud machine
+ echo "Checking if Undercloud needs to be resized..."
+ undercloud_size=$(LIBGUESTFS_BACKEND=direct virt-filesystems --long -h --all -a $undercloud_dst |grep device | grep -Eo "[0-9\.]+G" | sed -n 's/\([0-9][0-9]*\).*/\1/p')
+ if [ "$undercloud_size" -lt 30 ]; then
+ qemu-img resize /var/lib/libvirt/images/undercloud.qcow2 +25G
+ LIBGUESTFS_BACKEND=direct virt-resize --expand /dev/sda1 $RESOURCES/undercloud.qcow2 $undercloud_dst
+ LIBGUESTFS_BACKEND=direct virt-customize -a $undercloud_dst --run-command 'xfs_growfs -d /dev/sda1 || true'
+ new_size=$(LIBGUESTFS_BACKEND=direct virt-filesystems --long -h --all -a $undercloud_dst |grep filesystem | grep -Eo "[0-9\.]+G" | sed -n 's/\([0-9][0-9]*\).*/\1/p')
if [ "$new_size" -lt 30 ]; then
- echo "Error resizing instack machine, disk size is ${new_size}"
+ echo "Error resizing Undercloud machine, disk size is ${new_size}"
exit 1
else
- echo "instack successfully resized"
+ echo "Undercloud successfully resized"
fi
else
- echo "skipped instack resize, upstream is large enough"
+ echo "Skipped Undercloud resize, upstream is large enough"
fi
else
- echo "Found Instack VM, using existing VM"
+ echo "Found Undercloud VM, using existing VM"
fi
# if the VM is not running update the authkeys and start it
- if ! virsh list | grep instack > /dev/null; then
- echo "Injecting ssh key to instack VM"
- virt-customize -c qemu:///system -d instack --run-command "mkdir -p /root/.ssh/" \
+ if ! virsh list | grep undercloud > /dev/null; then
+ echo "Injecting ssh key to Undercloud VM"
+ LIBGUESTFS_BACKEND=direct virt-customize -a $undercloud_dst --run-command "mkdir -p /root/.ssh/" \
--upload ~/.ssh/id_rsa.pub:/root/.ssh/authorized_keys \
--run-command "chmod 600 /root/.ssh/authorized_keys && restorecon /root/.ssh/authorized_keys" \
--run-command "cp /root/.ssh/authorized_keys /home/stack/.ssh/" \
--run-command "chown stack:stack /home/stack/.ssh/authorized_keys && chmod 600 /home/stack/.ssh/authorized_keys"
- virsh start instack
+ virsh start undercloud
fi
- sleep 3 # let DHCP happen
+ sleep 10 # let undercloud get started up
+ # get the undercloud VM IP
CNT=10
- echo -n "${blue}Waiting for instack's dhcp address${reset}"
- while ! grep instack /var/lib/libvirt/dnsmasq/default.leases > /dev/null && [ $CNT -gt 0 ]; do
+ echo -n "${blue}Waiting for Undercloud's dhcp address${reset}"
+ undercloud_mac=$(virsh domiflist undercloud | grep default | awk '{ print $5 }')
+ while ! $(arp -e | grep ${undercloud_mac} > /dev/null) && [ $CNT -gt 0 ]; do
echo -n "."
- sleep 3
- CNT=CNT-1
+ sleep 10
+ CNT=$((CNT-1))
done
+ UNDERCLOUD=$(arp -e | grep ${undercloud_mac} | awk {'print $1'})
- # get the instack VM IP
- UNDERCLOUD=$(grep instack /var/lib/libvirt/dnsmasq/default.leases | awk '{print $3}' | head -n 1)
if [ -z "$UNDERCLOUD" ]; then
- #if not found then dnsmasq may be using leasefile-ro
- instack_mac=$(virsh domiflist instack | grep default | \
- grep -Eo "[0-9a-f\]+:[0-9a-f\]+:[0-9a-f\]+:[0-9a-f\]+:[0-9a-f\]+:[0-9a-f\]+")
- UNDERCLOUD=$(/usr/sbin/arp -e | grep ${instack_mac} | awk {'print $1'})
-
- if [ -z "$UNDERCLOUD" ]; then
- echo "\n\nNever got IP for Instack. Can Not Continue."
- exit 1
- else
- echo -e "${blue}\rInstack VM has IP $UNDERCLOUD${reset}"
- fi
+ echo "\n\nCan't get IP for Undercloud. Can Not Continue."
+ exit 1
else
- echo -e "${blue}\rInstack VM has IP $UNDERCLOUD${reset}"
+ echo -e "${blue}\rUndercloud VM has IP $UNDERCLOUD${reset}"
fi
CNT=10
- echo -en "${blue}\rValidating instack VM connectivity${reset}"
+ echo -en "${blue}\rValidating Undercloud VM connectivity${reset}"
while ! ping -c 1 $UNDERCLOUD > /dev/null && [ $CNT -gt 0 ]; do
echo -n "."
sleep 3
- CNT=$CNT-1
+ CNT=$((CNT-1))
done
if [ "$CNT" -eq 0 ]; then
- echo "Failed to contact Instack. Can Not Continue"
+ echo "Failed to contact Undercloud. Can Not Continue"
exit 1
fi
CNT=10
while ! ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "echo ''" 2>&1> /dev/null && [ $CNT -gt 0 ]; do
echo -n "."
sleep 3
- CNT=$CNT-1
+ CNT=$((CNT-1))
done
if [ "$CNT" -eq 0 ]; then
- echo "Failed to connect to Instack. Can Not Continue"
+ echo "Failed to connect to Undercloud. Can Not Continue"
exit 1
fi
# extra space to overwrite the previous connectivity output
echo -e "${blue}\r ${reset}"
+ sleep 1
+ ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "if ! ip a s eth2 | grep ${public_network_provisioner_ip} > /dev/null; then ip a a ${public_network_provisioner_ip}/${public_network_cidr##*/} dev eth2; ip link set up dev eth2; fi"
- #add the instack public interface if net isolation is enabled (more than just admin network)
- if [[ "$net_isolation_enabled" == "TRUE" ]]; then
- virsh attach-interface --domain instack --type network --source ${NET_MAP['public_network']} --model rtl8139 --config --live
- sleep 1
- ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "if ! ip a s eth2 | grep ${public_network_provisioner_ip} > /dev/null; then ip a a ${public_network_provisioner_ip}/${public_network_cidr##*/} dev eth2; ip link set up dev eth2; fi"
- fi
# ssh key fix for stack user
ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "restorecon -r /home/stack"
}
##Create virtual nodes in virsh
##params: none
function setup_virtual_baremetal {
+ #start by generating the opening json for instackenv.json
+ cat > $CONFIG/instackenv-virt.json << EOF
+{
+ "nodes": [
+EOF
+
+ # next create the virtual machines and add their definitions to the file
for i in $(seq 0 $vm_index); do
- if ! virsh list --all | grep baremetalbrbm_brbm1_brbm2_brbm3_${i} > /dev/null; then
- if [ ! -e $CONFIG/baremetalbrbm_brbm1_brbm2_brbm3_${i}.xml ]; then
- define_virtual_node baremetalbrbm_brbm1_brbm2_brbm3_${i}
- fi
- # Fix for ramdisk using wrong pxeboot interface
- # TODO: revisit this and see if there's a more proper fix
- sed -i "/^\s*<source network='brbm2'\/>/{
- N
- s/^\(.*\)virtio\(.*\)$/\1rtl8139\2/
- }" $CONFIG/baremetalbrbm_brbm1_brbm2_brbm3_${i}.xml
- virsh define $CONFIG/baremetalbrbm_brbm1_brbm2_brbm3_${i}.xml
+ if ! virsh list --all | grep baremetal${i} > /dev/null; then
+ define_vm baremetal${i} network 41 'admin_network'
+ for n in private_network public_network storage_network; do
+ if [[ $enabled_network_list =~ $n ]]; then
+ echo -n "$n "
+ virsh attach-interface --domain baremetal${i} --type network --source $n --model rtl8139 --config
+ fi
+ done
else
echo "Found Baremetal ${i} VM, using existing VM"
fi
- virsh vol-list default | grep baremetalbrbm_brbm1_brbm2_brbm3_${i} 2>&1> /dev/null || virsh vol-create-as default baremetalbrbm_brbm1_brbm2_brbm3_${i}.qcow2 40G --format qcow2
+ #virsh vol-list default | grep baremetal${i} 2>&1> /dev/null || virsh vol-create-as default baremetal${i}.qcow2 41G --format qcow2
+ mac=$(virsh domiflist baremetal${i} | grep admin_network | awk '{ print $5 }')
+
+ cat >> $CONFIG/instackenv-virt.json << EOF
+ {
+ "pm_addr": "192.168.122.1",
+ "pm_user": "root",
+ "pm_password": "INSERT_STACK_USER_PRIV_KEY",
+ "pm_type": "pxe_ssh",
+ "mac": [
+ "$mac"
+ ],
+ "cpu": "2",
+ "memory": "8192",
+ "disk": "41",
+ "arch": "x86_64"
+ },
+EOF
done
+ #truncate the last line to remove the comma behind the bracket
+ tail -n 1 $CONFIG/instackenv-virt.json | wc -c | xargs -I {} truncate $CONFIG/instackenv-virt.json -s -{}
+
+ #finally reclose the bracket and close the instackenv.json file
+ cat >> $CONFIG/instackenv-virt.json << EOF
+ }
+ ],
+ "arch": "x86_64",
+ "host-ip": "192.168.122.1",
+ "power_manager": "nova.virt.baremetal.virtual_power_driver.VirtualPowerManager",
+ "seed-ip": "",
+ "ssh-key": "INSERT_STACK_USER_PRIV_KEY",
+ "ssh-user": "root"
+}
+EOF
+}
+
+##Create virtual nodes in virsh
+##params: name - String: libvirt name for VM
+## bootdev - String: boot device for the VM
+## disksize - Number: size of the disk in GB
+## ovs_bridges: - List: list of ovs bridges
+function define_vm () {
+ # Create the libvirt storage volume
+ if virsh vol-list default | grep ${1}.qcow2 2>&1> /dev/null; then
+ volume_path=$(virsh vol-path --pool default ${1}.qcow2 || echo "/var/lib/libvirt/images/${1}.qcow2")
+ echo "Volume ${1} exists. Deleting Existing Volume $volume_path"
+ virsh vol-dumpxml ${1}.qcow2 --pool default > /dev/null || echo '' #ok for this to fail
+ touch $volume_path
+ virsh vol-delete ${1}.qcow2 --pool default
+ fi
+ virsh vol-create-as default ${1}.qcow2 ${3}G --format qcow2
+ volume_path=$(virsh vol-path --pool default ${1}.qcow2)
+ if [ ! -f $volume_path ]; then
+ echo "$volume_path Not created successfully... Aborting"
+ exit 1
+ fi
+
+ # create the VM
+ /usr/libexec/openstack-tripleo/configure-vm --name $1 \
+ --bootdev $2 \
+ --image "$volume_path" \
+ --diskbus sata \
+ --arch x86_64 \
+ --cpus 2 \
+ --memory 8388608 \
+ --libvirt-nic-driver virtio \
+ --baremetal-interface $4
}
##Set network-environment settings
##params: network-environment file to edit
function configure_network_environment {
- local tht_dir nic_ext
+ local tht_dir
tht_dir=/usr/share/openstack-tripleo-heat-templates/network
- nic_ext=''
sed -i '/ControlPlaneSubnetCidr/c\\ ControlPlaneSubnetCidr: "'${admin_network_cidr##*/}'"' $1
sed -i '/ControlPlaneDefaultRoute/c\\ ControlPlaneDefaultRoute: '${admin_network_provisioner_ip}'' $1
sed -i 's#^.*Compute::Ports::TenantPort:.*$# OS::TripleO::Compute::Ports::TenantPort: '${tht_dir}'/ports/tenant.yaml#' $1
sed -i "/TenantAllocationPools/c\\ TenantAllocationPools: [{'start': '${private_network_usable_ip_range%%,*}', 'end': '${private_network_usable_ip_range##*,}'}]" $1
sed -i '/TenantNetCidr/c\\ TenantNetCidr: '${private_network_cidr}'' $1
- nic_ext+=_private
+ nics_cfg+=_private
else
sed -i 's#^.*Network::Tenant.*$# OS::TripleO::Network::Tenant: '${tht_dir}'/noop.yaml#' $1
sed -i 's#^.*Controller::Ports::TenantPort:.*$# OS::TripleO::Controller::Ports::TenantPort: '${tht_dir}'/ports/noop.yaml#' $1
# check for storage network
if [[ ! -z "$storage_network_enabled" && "$storage_network_enabled" == "true" ]]; then
- sed -i 's#^.*Network::Storage.*$# OS::TripleO::Network::Storage: '${tht_dir}'/storage.yaml#' $1
+ sed -i 's#^.*Network::Storage:.*$# OS::TripleO::Network::Storage: '${tht_dir}'/storage.yaml#' $1
+ sed -i 's#^.*Network::Ports::StorageVipPort:.*$# OS::TripleO::Network::Ports::StorageVipPort: '${tht_dir}'/ports/storage.yaml#' $1
sed -i 's#^.*Controller::Ports::StoragePort:.*$# OS::TripleO::Controller::Ports::StoragePort: '${tht_dir}'/ports/storage.yaml#' $1
sed -i 's#^.*Compute::Ports::StoragePort:.*$# OS::TripleO::Compute::Ports::StoragePort: '${tht_dir}'/ports/storage.yaml#' $1
sed -i "/StorageAllocationPools/c\\ StorageAllocationPools: [{'start': '${storage_network_usable_ip_range%%,*}', 'end': '${storage_network_usable_ip_range##*,}'}]" $1
sed -i '/StorageNetCidr/c\\ StorageNetCidr: '${storage_network_cidr}'' $1
- nic_ext+=_storage
+ nics_cfg+=_storage
else
- sed -i 's#^.*Network::Storage.*$# OS::TripleO::Network::Storage: '${tht_dir}'/noop.yaml#' $1
+ sed -i 's#^.*Network::Storage:.*$# OS::TripleO::Network::Storage: '${tht_dir}'/noop.yaml#' $1
+ sed -i 's#^.*Network::Ports::StorageVipPort:.*$# OS::TripleO::Network::Ports::StorageVipPort: '${tht_dir}'/ports/noop.yaml#' $1
sed -i 's#^.*Controller::Ports::StoragePort:.*$# OS::TripleO::Controller::Ports::StoragePort: '${tht_dir}'/ports/noop.yaml#' $1
sed -i 's#^.*Compute::Ports::StoragePort:.*$# OS::TripleO::Compute::Ports::StoragePort: '${tht_dir}'/ports/noop.yaml#' $1
fi
- # set nics appropriately
- sed -i 's#^.*Compute::Net::SoftwareConfig:.*$# OS::TripleO::Compute::Net::SoftwareConfig: nics/compute'${nic_ext}'.yaml#' $1
- sed -i 's#^.*Controller::Net::SoftwareConfig:.*$# OS::TripleO::Controller::Net::SoftwareConfig: nics/controller'${nic_ext}'.yaml#' $1
+ # check for ODL L3
+ if [ "${deploy_options_array['sdn_l3']}" == 'true' ]; then
+ nics_cfg+=_br-ex_no-public-ip
+ fi
+
}
-##Copy over the glance images and instack json file
+##Copy over the glance images and instackenv json file
##params: none
function configure_undercloud {
echo
- echo "Copying configuration file and disk images to instack"
- scp ${SSH_OPTIONS[@]} $RESOURCES/overcloud-full.qcow2 "stack@$UNDERCLOUD":
+ echo "Copying configuration files to Undercloud"
if [[ "$net_isolation_enabled" == "TRUE" ]]; then
configure_network_environment $CONFIG/network-environment.yaml
echo -e "${blue}Network Environment set for Deployment: ${reset}"
cat $CONFIG/network-environment.yaml
scp ${SSH_OPTIONS[@]} $CONFIG/network-environment.yaml "stack@$UNDERCLOUD":
+ ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" << EOI
+mkdir nics/
+cat > nics/controller.yaml << EOF
+$(nics_cfg=$nics_cfg sh $CONFIG/nics-controller.yaml.template)
+EOF
+cat > nics/compute.yaml << EOF
+$(nics_cfg=$nics_cfg sh $CONFIG/nics-compute.yaml.template)
+EOF
+EOI
fi
- scp ${SSH_OPTIONS[@]} -r $CONFIG/nics/ "stack@$UNDERCLOUD":
- # ensure stack user on instack machine has an ssh key
+ # ensure stack user on Undercloud machine has an ssh key
ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "if [ ! -e ~/.ssh/id_rsa.pub ]; then ssh-keygen -t rsa -N '' -f ~/.ssh/id_rsa; fi"
if [ "$virtual" == "TRUE" ]; then
- # copy the instack vm's stack user's pub key to
- # root's auth keys so that instack can control
+ # copy the Undercloud VM's stack user's pub key to
+ # root's auth keys so that Undercloud can control
# vm power on the hypervisor
ssh ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "cat /home/stack/.ssh/id_rsa.pub" >> /root/.ssh/authorized_keys
- # fix MACs to match new setup
- for i in $(seq 0 $vm_index); do
- pyscript="import json
-data = json.load(open('$CONFIG/instackenv-virt.json'))
-print data['nodes'][$i]['mac'][0]"
-
- old_mac=$(python -c "$pyscript")
- new_mac=$(virsh dumpxml baremetalbrbm_brbm1_brbm2_brbm3_$i | grep "mac address" | cut -d = -f2 | grep -Eo "[0-9a-f:]+")
- # this doesn't work with multiple vnics on the vms
- #if [ "$old_mac" != "$new_mac" ]; then
- # echo "${blue}Modifying MAC for node from $old_mac to ${new_mac}${reset}"
- # sed -i 's/'"$old_mac"'/'"$new_mac"'/' $CONFIG/instackenv-virt.json
- #fi
- done
-
DEPLOY_OPTIONS+=" --libvirt-type qemu"
INSTACKENV=$CONFIG/instackenv-virt.json
- # upload instackenv file to Instack for virtual deployment
+ # upload instackenv file to Undercloud for virtual deployment
scp ${SSH_OPTIONS[@]} $INSTACKENV "stack@$UNDERCLOUD":instackenv.json
fi
# configure undercloud on Undercloud VM
echo "Running undercloud configuration."
- echo "Logging undercloud configuration to instack:/home/stack/apex-undercloud-install.log"
+ echo "Logging undercloud configuration to undercloud:/home/stack/apex-undercloud-install.log"
ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" << EOI
if [[ "$net_isolation_enabled" == "TRUE" ]]; then
sed -i 's/#local_ip/local_ip/' undercloud.conf
sudo sed -i '/CephMonKey:/c\\ CephMonKey: \\x27'"\$(ceph-authtool --gen-print-key)"'\\x27' /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml
sudo sed -i '/CephAdminKey:/c\\ CephAdminKey: \\x27'"\$(ceph-authtool --gen-print-key)"'\\x27' /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml
-openstack undercloud install &> apex-undercloud-install.log
+# we assume that packages will not need to be updated with undercloud install
+# and that it will be used only to configure the undercloud
+# packages updates would need to be handled manually with yum update
+sudo cp -f /usr/share/diskimage-builder/elements/yum/bin/install-packages /usr/share/diskimage-builder/elements/yum/bin/install-packages.bak
+cat << 'EOF' | sudo tee /usr/share/diskimage-builder/elements/yum/bin/install-packages > /dev/null
+#!/bin/sh
+exit 0
+EOF
+
+openstack undercloud install &> apex-undercloud-install.log || {
+ # cat the undercloud install log incase it fails
+ echo "ERROR: openstack undercloud install has failed. Dumping Log:"
+ cat apex-undercloud-install.log
+ exit 1
+}
+
sleep 30
sudo systemctl restart openstack-glance-api
sudo systemctl restart openstack-nova-conductor
##preping it for deployment and launch the deploy
##params: none
function undercloud_prep_overcloud_deploy {
-
- if [[ ${#deploy_options_array[@]} -eq 0 || ${deploy_options_array['sdn_controller']} == 'opendaylight' ]]; then
- DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight.yaml"
- elif [ ${deploy_options_array['sdn_controller']} == 'opendaylight-external' ]; then
+ if [[ "${#deploy_options_array[@]}" -eq 0 || "${deploy_options_array['sdn_controller']}" == 'opendaylight' ]]; then
+ if [ "${deploy_options_array['sdn_l3']}" == 'true' ]; then
+ DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight_l3.yaml"
+ elif [ "${deploy_options_array['sfc']}" == 'true' ]; then
+ DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight_sfc.yaml"
+ elif [ "${deploy_options_array['vpn']}" == 'true' ]; then
+ DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight_sdnvpn.yaml"
+ else
+ DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight.yaml"
+ fi
+ SDN_IMAGE=opendaylight
+ if [ "${deploy_options_array['sfc']}" == 'true' ]; then
+ SDN_IMAGE+=-sfc
+ if [ ! -f $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 ]; then
+ echo "${red} $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 is required to execute an SFC deployment."
+ echo "Please install the opnfv-apex-opendaylight-sfc package to provide this overcloud image for deployment.${reset}"
+ exit 1
+ fi
+ fi
+ elif [ "${deploy_options_array['sdn_controller']}" == 'opendaylight-external' ]; then
DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight-external.yaml"
- elif [ ${deploy_options_array['sdn_controller']} == 'onos' ]; then
+ SDN_IMAGE=opendaylight
+ elif [ "${deploy_options_array['sdn_controller']}" == 'onos' ]; then
DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/onos.yaml"
- elif [ ${deploy_options_array['sdn_controller']} == 'opencontrail' ]; then
+ SDN_IMAGE=onos
+ elif [ "${deploy_options_array['sdn_controller']}" == 'opencontrail' ]; then
echo -e "${red}ERROR: OpenContrail is currently unsupported...exiting${reset}"
exit 1
+ elif [[ -z "${deploy_options_array['sdn_controller']}" || "${deploy_options_array['sdn_controller']}" == 'false' ]]; then
+ echo -e "${blue}INFO: SDN Controller disabled...will deploy nosdn scenario${reset}"
+ SDN_IMAGE=opendaylight
+ else
+ echo "${red}Invalid sdn_controller: ${deploy_options_array['sdn_controller']}${reset}"
+ echo "${red}Valid choices are opendaylight, opendaylight-external, onos, opencontrail, false, or null${reset}"
+ exit 1
fi
+ # Make sure the correct overcloud image is available
+ if [ ! -f $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 ]; then
+ echo "${red} $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 is required to execute your deployment."
+ echo "Both ONOS and OpenDaylight are currently deployed from this image."
+ echo "Please install the opnfv-apex package to provide this overcloud image for deployment.${reset}"
+ exit 1
+ fi
+
+ echo "Copying overcloud image to Undercloud"
+ ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "rm -f overcloud-full.qcow2"
+ scp ${SSH_OPTIONS[@]} $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 "stack@$UNDERCLOUD":overcloud-full.qcow2
+
# make sure ceph is installed
DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml"
+ # scale compute nodes according to inventory
+ total_nodes=$(ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "cat /home/stack/instackenv.json | grep -c memory")
+
# check if HA is enabled
- if [[ "$ha_enabled" == "TRUE" ]]; then
- DEPLOY_OPTIONS+=" --control-scale 3 --compute-scale 2"
+ if [[ "$ha_enabled" == "TRUE" || "$ha_enabled" == "true" ]]; then
+ DEPLOY_OPTIONS+=" --control-scale 3"
+ compute_nodes=$((total_nodes - 3))
DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/puppet-pacemaker.yaml"
+ else
+ compute_nodes=$((total_nodes - 1))
+ fi
+
+ if [ "$compute_nodes" -le 0 ]; then
+ echo -e "${red}ERROR: Invalid number of compute nodes: ${compute_nodes}. Check your inventory file.${reset}"
+ exit 1
+ else
+ echo -e "${blue}INFO: Number of compute nodes set for deployment: ${compute_nodes}${reset}"
+ DEPLOY_OPTIONS+=" --compute-scale ${compute_nodes}"
fi
if [[ "$net_isolation_enabled" == "TRUE" ]]; then
DEPLOY_OPTIONS+=" -e network-environment.yaml"
fi
- if [[ "$ha_enabled" == "TRUE" ]] || [[ "$net_isolation_enabled" == "TRUE" ]]; then
+ if [[ "$ha_enabled" == "TRUE" || "$ha_enabled" == "true" ]] || [[ "$net_isolation_enabled" == "TRUE" ]]; then
DEPLOY_OPTIONS+=" --ntp-server $ntp_server"
fi
DEPLOY_OPTIONS+=" --control-flavor control --compute-flavor compute"
fi
+ echo -e "${blue}INFO: Deploy options set:\n${DEPLOY_OPTIONS}${reset}"
+
ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+if [ "$debug" == 'TRUE' ]; then
+ LIBGUESTFS_BACKEND=direct virt-customize -a overcloud-full.qcow2 --root-password password:opnfvapex
+fi
+
source stackrc
set -o errexit
echo "Uploading overcloud glance images"
echo "Configuring undercloud and discovering nodes"
openstack baremetal import --json instackenv.json
openstack baremetal configure boot
-openstack baremetal introspection bulk start
+#if [[ -z "$virtual" ]]; then
+# openstack baremetal introspection bulk start
+#fi
echo "Configuring flavors"
for flavor in baremetal control compute; do
echo -e "${blue}INFO: Updating flavor: \${flavor}${reset}"
sleep 60 #wait for Hypervisor stats to check-in to nova
# save deploy command so it can be used for debugging
cat > deploy_command << EOF
-openstack overcloud deploy --templates $DEPLOY_OPTIONS
+openstack overcloud deploy --templates $DEPLOY_OPTIONS --timeout 90
EOF
-openstack overcloud deploy --templates $DEPLOY_OPTIONS
EOI
+ if [ "$interactive" == "TRUE" ]; then
+ if ! prompt_user "Overcloud Deployment"; then
+ echo -e "${blue}INFO: User requests exit${reset}"
+ exit 0
+ fi
+ fi
+
+ ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+source stackrc
+set -o errexit
+openstack overcloud deploy --templates $DEPLOY_OPTIONS --timeout 90
+EOI
+
+ if [ "$debug" == 'TRUE' ]; then
+ ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+source overcloudrc
+echo "Keystone Endpoint List:"
+keystone endpoint-list
+echo "Keystone Service List"
+keystone service-list
+cinder quota-show \$(openstack project list | grep admin | awk {'print \$2'})
+EOI
+ fi
}
##Post configuration after install
source overcloudrc
set -o errexit
echo "Configuring Neutron external network"
-neutron net-create external --router:external=True
-neutron subnet-create --name external-net --disable-dhcp external --gateway ${public_network_gateway} --allocation-pool start=${public_network_floating_ip_range%%,*},end=${public_network_floating_ip_range##*,} ${public_network_cidr}
+neutron net-create external --router:external=True --tenant-id \$(keystone tenant-get service | grep id | awk '{ print \$4 }')
+neutron subnet-create --name external-net --tenant-id \$(keystone tenant-get service | grep id | awk '{ print \$4 }') --disable-dhcp external --gateway ${public_network_gateway} --allocation-pool start=${public_network_floating_ip_range%%,*},end=${public_network_floating_ip_range##*,} ${public_network_cidr}
EOI
echo -e "${blue}INFO: Checking if OVS bridges have IP addresses...${reset}"
fi
fi
done
+
+ # for virtual, we NAT public network through Undercloud
+ if [ "$virtual" == "TRUE" ]; then
+ if ! configure_undercloud_nat ${public_network_cidr}; then
+ echo -e "${red}ERROR: Unable to NAT undercloud with external net: ${public_network_cidr}${reset}"
+ exit 1
+ else
+ echo -e "${blue}INFO: Undercloud VM has been setup to NAT Overcloud public network${reset}"
+ fi
+ fi
+
+ # for sfc deployments we need the vxlan workaround
+ if [ "${deploy_options_array['sfc']}" == 'true' ]; then
+ ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+source stackrc
+set -o errexit
+for node in \$(nova list | grep -Eo "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+"); do
+ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
+sudo ifconfig br-int up
+sudo ip route add 123.123.123.0/24 dev br-int
+EOF
+done
+EOI
+ fi
+
+ # Collect deployment logs
+ ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+mkdir -p ~/deploy_logs
+rm -rf deploy_logs/*
+source stackrc
+set -o errexit
+for node in \$(nova list | grep -Eo "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+"); do
+ ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
+ sudo cp /var/log/messages /home/heat-admin/messages.log
+ sudo chown heat-admin /home/heat-admin/messages.log
+EOF
+scp ${SSH_OPTIONS[@]} heat-admin@\$node:/home/heat-admin/messages.log ~/deploy_logs/\$node.messages.log
+if [ "$debug" == "TRUE" ]; then
+ nova list --ip \$node
+ echo "---------------------------"
+ echo "-----/var/log/messages-----"
+ echo "---------------------------"
+ cat ~/deploy_logs/\$node.messages.log
+ echo "---------------------------"
+ echo "----------END LOG----------"
+ echo "---------------------------"
+fi
+ ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
+ sudo rm -f /home/heat-admin/messages.log
+EOF
+done
+
+# Print out the dashboard URL
+source stackrc
+echo "Overcloud dashboard available at http://\$(heat output-show overcloud PublicVip | sed 's/"//g')/dashboard"
+EOI
+
}
display_usage() {
echo -e " --no-ha : disable High Availability deployment scheme, this assumes a single controller and single compute node \n"
echo -e " --flat : disable Network Isolation and use a single flat network for the underlay network.\n"
echo -e " --no-post-config : disable Post Install configuration."
+ echo -e " --debug : enable debug output."
+ echo -e " --interactive : enable interactive deployment mode which requires user to confirm steps of deployment."
}
##translates the command line parameters into variables
;;
--no-ha )
ha_enabled="FALSE"
+ vm_index=1
echo "HA Deployment Disabled"
shift 1
;;
echo "Post install configuration disabled"
shift 1
;;
+ --debug )
+ debug="TRUE"
+ echo "Enable debug output"
+ shift 1
+ ;;
+ --interactive )
+ interactive="TRUE"
+ echo "Interactive mode enabled"
+ shift 1
+ ;;
*)
display_usage
exit 1
done
if [[ ! -z "$NETSETS" && "$net_isolation_enabled" == "FALSE" ]]; then
- echo -e "${red}INFO: Single flat network requested. Ignoring any network settings!${reset}"
- elif [[ -z "$NETSETS" && "$net_isolation_enabled" == "TRUE" ]]; then
- echo -e "${red}ERROR: You must provide a network_settings file with -n or use --flat to force a single flat network{reset}"
+ echo -e "${red}INFO: Single flat network requested. Only admin_network settings will be used!${reset}"
+ elif [[ -z "$NETSETS" ]]; then
+ echo -e "${red}ERROR: You must provide a network_settings file with -n.${reset}"
exit 1
fi
exit 1
fi
- if [[ ! -z "$DEPLOY_SETTINGS_FILE" && ! -f "$DEPLOY_SETTINGS_FILE" ]]; then
- echo -e "${red}ERROR: ${DEPLOY_SETTINGS_FILE} does not exist! Exiting...${reset}"
+ if [[ -z "$DEPLOY_SETTINGS_FILE" || ! -f "$DEPLOY_SETTINGS_FILE" ]]; then
+ echo -e "${red}ERROR: Deploy Settings: ${DEPLOY_SETTINGS_FILE} does not exist! Exiting...${reset}"
exit 1
fi
if [[ ! -z "$NETSETS" && ! -f "$NETSETS" ]]; then
- echo -e "${red}ERROR: ${NETSETS} does not exist! Exiting...${reset}"
+ echo -e "${red}ERROR: Network Settings: ${NETSETS} does not exist! Exiting...${reset}"
exit 1
fi
if [[ ! -z "$INVENTORY_FILE" && ! -f "$INVENTORY_FILE" ]]; then
- echo -e "{$red}ERROR: ${DEPLOY_SETTINGS_FILE} does not exist! Exiting...${reset}"
+ echo -e "{$red}ERROR: Inventory File: ${INVENTORY_FILE} does not exist! Exiting...${reset}"
exit 1
fi
echo -e "${red}ERROR: You must specify an inventory file for baremetal deployments! Exiting...${reset}"
exit 1
fi
+
+ if [[ "$net_isolation_enabled" == "FALSE" && "$post_config" == "TRUE" ]]; then
+ echo -e "${blue}INFO: Post Install Configuration will be skipped. It is not supported with --flat${reset}"
+ post_config="FALSE"
+ fi
+
+ ##LIBRARIES
+ # Do this after cli parse so that $CONFIG is set properly
+ source $CONFIG/lib/common-functions.sh
+ source $CONFIG/lib/installer/onos/onos_gw_mac_update.sh
+
}
##END FUNCTIONS
main() {
parse_cmdline "$@"
- if [[ "$net_isolation_enabled" == "TRUE" ]]; then
- echo -e "${blue}INFO: Parsing network settings file...${reset}"
- parse_network_settings
- fi
+ echo -e "${blue}INFO: Parsing network settings file...${reset}"
+ parse_network_settings
if ! configure_deps; then
echo -e "${red}Dependency Validation Failed, Exiting.${reset}"
exit 1
if [ -n "$DEPLOY_SETTINGS_FILE" ]; then
parse_deploy_settings
fi
- setup_instack_vm
+ setup_undercloud_vm
if [ "$virtual" == "TRUE" ]; then
setup_virtual_baremetal
elif [ -n "$INVENTORY_FILE" ]; then
echo -e "${blue}INFO: Post Install Configuration Complete${reset}"
fi
fi
+ if [[ "${deploy_options_array['sdn_controller']}" == 'onos' ]]; then
+ if ! onos_update_gw_mac ${public_network_cidr} ${public_network_gateway}; then
+ echo -e "${red}ERROR:ONOS Post Install Configuration Failed, Exiting.${reset}"
+ exit 1
+ else
+ echo -e "${blue}INFO: ONOS Post Install Configuration Complete${reset}"
+ fi
+ fi
}
main "$@"