Merge "forcing rpm-check if files added to the commit"
authorMichael Chapman <michapma@redhat.com>
Fri, 8 Jul 2016 02:20:55 +0000 (02:20 +0000)
committerGerrit Code Review <gerrit@172.30.200.206>
Fri, 8 Jul 2016 02:20:55 +0000 (02:20 +0000)
build/opnfv-apex-common.spec
ci/deploy.sh
lib/common-functions.sh
lib/configure-deps-functions.sh [new file with mode: 0755]
lib/overcloud-deploy-functions.sh [new file with mode: 0755]
lib/post-install-functions.sh [new file with mode: 0755]
lib/undercloud-functions.sh [new file with mode: 0755]
lib/virtual-setup-functions.sh [new file with mode: 0755]

index ce77d6d..8808cdb 100644 (file)
@@ -52,7 +52,12 @@ install config/network/network_settings_v6.yaml %{buildroot}%{_sysconfdir}/opnfv
 
 mkdir -p %{buildroot}%{_var}/opt/opnfv/lib/python/apex
 install lib/common-functions.sh %{buildroot}%{_var}/opt/opnfv/lib/
+install lib/configure-deps-functions.sh %{buildroot}%{_var}/opt/opnfv/lib/
 install lib/parse-functions.sh %{buildroot}%{_var}/opt/opnfv/lib/
+install lib/virtual-setup-functions.sh %{buildroot}%{_var}/opt/opnfv/lib/
+install lib/undercloud-functions.sh %{buildroot}%{_var}/opt/opnfv/lib/
+install lib/overcloud-deploy-functions.sh %{buildroot}%{_var}/opt/opnfv/lib/
+install lib/post-install-functions.sh %{buildroot}%{_var}/opt/opnfv/lib/
 install lib/utility-functions.sh %{buildroot}%{_var}/opt/opnfv/lib/
 install lib/python/apex_python_utils.py %{buildroot}%{_var}/opt/opnfv/lib/python/
 mkdir -p %{buildroot}%{python3_sitelib}/apex/
@@ -85,7 +90,12 @@ install config/inventory/pod_example_settings.yaml %{buildroot}%{_docdir}/opnfv/
 %attr(755,root,root) %{_bindir}/opnfv-clean
 %attr(755,root,root) %{_bindir}/opnfv-util
 %{_var}/opt/opnfv/lib/common-functions.sh
+%{_var}/opt/opnfv/lib/configure-deps-functions.sh
 %{_var}/opt/opnfv/lib/parse-functions.sh
+%{_var}/opt/opnfv/lib/virtual-setup-functions.sh
+%{_var}/opt/opnfv/lib/undercloud-functions.sh
+%{_var}/opt/opnfv/lib/overcloud-deploy-functions.sh
+%{_var}/opt/opnfv/lib/post-install-functions.sh
 %{_var}/opt/opnfv/lib/utility-functions.sh
 %{_var}/opt/opnfv/lib/python/
 %{python3_sitelib}/apex/
@@ -116,7 +126,7 @@ install config/inventory/pod_example_settings.yaml %{buildroot}%{_docdir}/opnfv/
 
 %changelog
 * Tue Jul 5 2016 Dan Radez <dradez@redhat.com> - 3.0-10
-- Adding parse-functions.sh
+- Adding functions.sh files
 * Thu Jun 15 2016 Tim Rozet <trozet@redhat.com> - 3.0-9
 - Add fdio scenarios.
 * Tue Jun 14 2016 Feng Pan <fpan@redhat.com> - 3.0-8
index bb4e12e..57c6376 100755 (executable)
@@ -58,7 +58,12 @@ ip_address_family=4
 # Libraries
 lib_files=(
 $LIB/common-functions.sh
+$LIB/configure-deps-functions.sh
 $LIB/parse-functions.sh
+$LIB/virtual-setup-functions.sh
+$LIB/undercloud-functions.sh
+$LIB/overcloud-deploy-functions.sh
+$LIB/post-install-functions.sh
 $LIB/utility-functions.sh
 $LIB/installer/onos/onos_gw_mac_update.sh
 )
@@ -69,997 +74,6 @@ for lib_file in ${lib_files[@]}; do
   fi
 done
 
-##FUNCTIONS
-##checks if prefix exists in string
-##params: string, prefix
-##usage: contains_prefix "deploy_setting_launcher=1" "deploy_setting"
-contains_prefix() {
-  local mystr=$1
-  local prefix=$2
-  if echo $mystr | grep -E "^$prefix.*$" > /dev/null; then
-    return 0
-  else
-    return 1
-  fi
-}
-
-##verify internet connectivity
-#params: none
-function verify_internet {
-  if ping -c 2 $ping_site > /dev/null; then
-    if ping -c 2 www.google.com > /dev/null; then
-      echo "${blue}Internet connectivity detected${reset}"
-      return 0
-    else
-      echo "${red}Internet connectivity detected, but DNS lookup failed${reset}"
-      return 1
-    fi
-  else
-    echo "${red}No internet connectivity detected${reset}"
-    return 1
-  fi
-}
-
-##download dependencies if missing and configure host
-#params: none
-function configure_deps {
-  if ! verify_internet; then
-    echo "${red}Will not download dependencies${reset}"
-    internet=false
-  fi
-
-  # verify ip forwarding
-  if sysctl net.ipv4.ip_forward | grep 0; then
-    sudo sysctl -w net.ipv4.ip_forward=1
-    sudo sh -c "echo 'net.ipv4.ip_forward = 1' >> /etc/sysctl.conf"
-  fi
-
-  # ensure no dhcp server is running on jumphost
-  if ! sudo systemctl status dhcpd | grep dead; then
-    echo "${red}WARN: DHCP Server detected on jumphost, disabling...${reset}"
-    sudo systemctl stop dhcpd
-    sudo systemctl disable dhcpd
-  fi
-
-  # ensure networks are configured
-  systemctl status libvirtd || systemctl start libvirtd
-  systemctl status openvswitch || systemctl start openvswitch
-
-  # If flat we only use admin network
-  if [[ "$net_isolation_enabled" == "FALSE" ]]; then
-    virsh_enabled_networks="admin_network"
-    enabled_network_list="admin_network"
-  # For baremetal we only need to create/attach Undercloud to admin and public
-  elif [ "$virtual" == "FALSE" ]; then
-    virsh_enabled_networks="admin_network public_network"
-  else
-    virsh_enabled_networks=$enabled_network_list
-  fi
-
-  # ensure default network is configured correctly
-  libvirt_dir="/usr/share/libvirt/networks"
-  virsh net-list --all | grep default || virsh net-define ${libvirt_dir}/default.xml
-  virsh net-list --all | grep -E "default\s+active" > /dev/null || virsh net-start default
-  virsh net-list --all | grep -E "default\s+active\s+yes" > /dev/null || virsh net-autostart --network default
-
-  if [[ -z "$virtual" || "$virtual" == "FALSE" ]]; then
-    for network in ${enabled_network_list}; do
-      echo "${blue}INFO: Creating Virsh Network: $network & OVS Bridge: ${NET_MAP[$network]}${reset}"
-      ovs-vsctl list-br | grep "^${NET_MAP[$network]}$" > /dev/null || ovs-vsctl add-br ${NET_MAP[$network]}
-      virsh net-list --all | grep $network > /dev/null || (cat > ${libvirt_dir}/apex-virsh-net.xml && virsh net-define ${libvirt_dir}/apex-virsh-net.xml) << EOF
-<network>
-  <name>$network</name>
-  <forward mode='bridge'/>
-  <bridge name='${NET_MAP[$network]}'/>
-  <virtualport type='openvswitch'/>
-</network>
-EOF
-      if ! (virsh net-list --all | grep $network > /dev/null); then
-          echo "${red}ERROR: unable to create network: ${network}${reset}"
-          exit 1;
-      fi
-      rm -f ${libvirt_dir}/apex-virsh-net.xml &> /dev/null;
-      virsh net-list | grep -E "$network\s+active" > /dev/null || virsh net-start $network
-      virsh net-list | grep -E "$network\s+active\s+yes" > /dev/null || virsh net-autostart --network $network
-    done
-
-    echo -e "${blue}INFO: Bridges set: ${reset}"
-    ovs-vsctl list-br
-
-    # bridge interfaces to correct OVS instances for baremetal deployment
-    for network in ${enabled_network_list}; do
-      if [[ "$network" != "admin_network" && "$network" != "public_network" ]]; then
-        continue
-      fi
-      this_interface=$(eval echo \${${network}_bridged_interface})
-      # check if this a bridged interface for this network
-      if [[ ! -z "$this_interface" || "$this_interface" != "none" ]]; then
-        if ! attach_interface_to_ovs ${NET_MAP[$network]} ${this_interface} ${network}; then
-          echo -e "${red}ERROR: Unable to bridge interface ${this_interface} to bridge ${NET_MAP[$network]} for enabled network: ${network}${reset}"
-          exit 1
-        else
-          echo -e "${blue}INFO: Interface ${this_interface} bridged to bridge ${NET_MAP[$network]} for enabled network: ${network}${reset}"
-        fi
-      else
-        echo "${red}ERROR: Unable to determine interface to bridge to for enabled network: ${network}${reset}"
-        exit 1
-      fi
-    done
-  else
-    for network in ${OPNFV_NETWORK_TYPES}; do
-      echo "${blue}INFO: Creating Virsh Network: $network${reset}"
-      virsh net-list --all | grep $network > /dev/null || (cat > ${libvirt_dir}/apex-virsh-net.xml && virsh net-define ${libvirt_dir}/apex-virsh-net.xml) << EOF
-<network ipv6='yes'>
-<name>$network</name>
-<bridge name='${NET_MAP[$network]}'/>
-</network>
-EOF
-      if ! (virsh net-list --all | grep $network > /dev/null); then
-          echo "${red}ERROR: unable to create network: ${network}${reset}"
-          exit 1;
-      fi
-      rm -f ${libvirt_dir}/apex-virsh-net.xml &> /dev/null;
-      virsh net-list | grep -E "$network\s+active" > /dev/null || virsh net-start $network
-      virsh net-list | grep -E "$network\s+active\s+yes" > /dev/null || virsh net-autostart --network $network
-    done
-
-    echo -e "${blue}INFO: Bridges set: ${reset}"
-    brctl show
-  fi
-
-  echo -e "${blue}INFO: virsh networks set: ${reset}"
-  virsh net-list
-
-  # ensure storage pool exists and is started
-  virsh pool-list --all | grep default > /dev/null || virsh pool-define-as --name default dir --target /var/lib/libvirt/images
-  virsh pool-list | grep -Eo "default\s+active" > /dev/null || (virsh pool-autostart default; virsh pool-start default)
-
-  if ! egrep '^flags.*(vmx|svm)' /proc/cpuinfo > /dev/null; then
-    echo "${red}virtualization extensions not found, kvm kernel module insertion may fail.\n  \
-Are you sure you have enabled vmx in your bios or hypervisor?${reset}"
-  fi
-
-  if ! lsmod | grep kvm > /dev/null; then modprobe kvm; fi
-  if ! lsmod | grep kvm_intel > /dev/null; then modprobe kvm_intel; fi
-
-  if ! lsmod | grep kvm > /dev/null; then
-    echo "${red}kvm kernel modules not loaded!${reset}"
-    return 1
-  fi
-
-  ##sshkeygen for root
-  if [ ! -e ~/.ssh/id_rsa.pub ]; then
-    ssh-keygen -t rsa -N "" -f ~/.ssh/id_rsa
-  fi
-
-  echo "${blue}All dependencies installed and running${reset}"
-}
-
-##verify vm exists, an has a dhcp lease assigned to it
-##params: none
-function setup_undercloud_vm {
-  if ! virsh list --all | grep undercloud > /dev/null; then
-      undercloud_nets="default admin_network"
-      if [[ $enabled_network_list =~ "public_network" ]]; then
-        undercloud_nets+=" public_network"
-      fi
-      define_vm undercloud hd 30 "$undercloud_nets" 4 12288
-
-      ### this doesn't work for some reason I was getting hangup events so using cp instead
-      #virsh vol-upload --pool default --vol undercloud.qcow2 --file $CONFIG/stack/undercloud.qcow2
-      #2015-12-05 12:57:20.569+0000: 8755: info : libvirt version: 1.2.8, package: 16.el7_1.5 (CentOS BuildSystem <http://bugs.centos.org>, 2015-11-03-13:56:46, worker1.bsys.centos.org)
-      #2015-12-05 12:57:20.569+0000: 8755: warning : virKeepAliveTimerInternal:143 : No response from client 0x7ff1e231e630 after 6 keepalive messages in 35 seconds
-      #2015-12-05 12:57:20.569+0000: 8756: warning : virKeepAliveTimerInternal:143 : No response from client 0x7ff1e231e630 after 6 keepalive messages in 35 seconds
-      #error: cannot close volume undercloud.qcow2
-      #error: internal error: received hangup / error event on socket
-      #error: Reconnected to the hypervisor
-
-      local undercloud_dst=/var/lib/libvirt/images/undercloud.qcow2
-      cp -f $RESOURCES/undercloud.qcow2 $undercloud_dst
-
-      # resize Undercloud machine
-      echo "Checking if Undercloud needs to be resized..."
-      undercloud_size=$(LIBGUESTFS_BACKEND=direct virt-filesystems --long -h --all -a $undercloud_dst |grep device | grep -Eo "[0-9\.]+G" | sed -n 's/\([0-9][0-9]*\).*/\1/p')
-      if [ "$undercloud_size" -lt 30 ]; then
-        qemu-img resize /var/lib/libvirt/images/undercloud.qcow2 +25G
-        LIBGUESTFS_BACKEND=direct virt-resize --expand /dev/sda1 $RESOURCES/undercloud.qcow2 $undercloud_dst
-        LIBGUESTFS_BACKEND=direct virt-customize -a $undercloud_dst --run-command 'xfs_growfs -d /dev/sda1 || true'
-        new_size=$(LIBGUESTFS_BACKEND=direct virt-filesystems --long -h --all -a $undercloud_dst |grep filesystem | grep -Eo "[0-9\.]+G" | sed -n 's/\([0-9][0-9]*\).*/\1/p')
-        if [ "$new_size" -lt 30 ]; then
-          echo "Error resizing Undercloud machine, disk size is ${new_size}"
-          exit 1
-        else
-          echo "Undercloud successfully resized"
-        fi
-      else
-        echo "Skipped Undercloud resize, upstream is large enough"
-      fi
-
-  else
-      echo "Found Undercloud VM, using existing VM"
-  fi
-
-  # if the VM is not running update the authkeys and start it
-  if ! virsh list | grep undercloud > /dev/null; then
-    echo "Injecting ssh key to Undercloud VM"
-    LIBGUESTFS_BACKEND=direct virt-customize -a $undercloud_dst --run-command "mkdir -p /root/.ssh/" \
-        --upload ~/.ssh/id_rsa.pub:/root/.ssh/authorized_keys \
-        --run-command "chmod 600 /root/.ssh/authorized_keys && restorecon /root/.ssh/authorized_keys" \
-        --run-command "cp /root/.ssh/authorized_keys /home/stack/.ssh/" \
-        --run-command "chown stack:stack /home/stack/.ssh/authorized_keys && chmod 600 /home/stack/.ssh/authorized_keys"
-    virsh start undercloud
-  fi
-
-  sleep 10 # let undercloud get started up
-
-  # get the undercloud VM IP
-  CNT=10
-  echo -n "${blue}Waiting for Undercloud's dhcp address${reset}"
-  undercloud_mac=$(virsh domiflist undercloud | grep default | awk '{ print $5 }')
-  while ! $(arp -e | grep ${undercloud_mac} > /dev/null) && [ $CNT -gt 0 ]; do
-      echo -n "."
-      sleep 10
-      CNT=$((CNT-1))
-  done
-  UNDERCLOUD=$(arp -e | grep ${undercloud_mac} | awk {'print $1'})
-
-  if [ -z "$UNDERCLOUD" ]; then
-    echo "\n\nCan't get IP for Undercloud. Can Not Continue."
-    exit 1
-  else
-     echo -e "${blue}\rUndercloud VM has IP $UNDERCLOUD${reset}"
-  fi
-
-  CNT=10
-  echo -en "${blue}\rValidating Undercloud VM connectivity${reset}"
-  while ! ping -c 1 $UNDERCLOUD > /dev/null && [ $CNT -gt 0 ]; do
-      echo -n "."
-      sleep 3
-      CNT=$((CNT-1))
-  done
-  if [ "$CNT" -eq 0 ]; then
-      echo "Failed to contact Undercloud. Can Not Continue"
-      exit 1
-  fi
-  CNT=10
-  while ! ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "echo ''" 2>&1> /dev/null && [ $CNT -gt 0 ]; do
-      echo -n "."
-      sleep 3
-      CNT=$((CNT-1))
-  done
-  if [ "$CNT" -eq 0 ]; then
-      echo "Failed to connect to Undercloud. Can Not Continue"
-      exit 1
-  fi
-
-  # extra space to overwrite the previous connectivity output
-  echo -e "${blue}\r                                                                 ${reset}"
-  sleep 1
-
-  # ssh key fix for stack user
-  ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "restorecon -r /home/stack"
-}
-
-##Create virtual nodes in virsh
-##params: vcpus, ramsize
-function setup_virtual_baremetal {
-  local vcpus ramsize
-  if [ -z "$1" ]; then
-    vcpus=4
-    ramsize=8192
-  elif [ -z "$2" ]; then
-    vcpus=$1
-    ramsize=8192
-  else
-    vcpus=$1
-    ramsize=$(($2*1024))
-  fi
-  #start by generating the opening json for instackenv.json
-  cat > $CONFIG/instackenv-virt.json << EOF
-{
-  "nodes": [
-EOF
-
-  # next create the virtual machines and add their definitions to the file
-  if [ "$ha_enabled" == "False" ]; then
-      # 1 controller + computes
-      # zero based so just pass compute count
-      vm_index=$VM_COMPUTES
-  else
-      # 3 controller + computes
-      # zero based so add 2 to compute count
-      vm_index=$((2+$VM_COMPUTES))
-  fi
-
-  for i in $(seq 0 $vm_index); do
-    if ! virsh list --all | grep baremetal${i} > /dev/null; then
-      define_vm baremetal${i} network 41 'admin_network' $vcpus $ramsize
-      for n in private_network public_network storage_network api_network; do
-        if [[ $enabled_network_list =~ $n ]]; then
-          echo -n "$n "
-          virsh attach-interface --domain baremetal${i} --type network --source $n --model virtio --config
-        fi
-      done
-    else
-      echo "Found Baremetal ${i} VM, using existing VM"
-    fi
-    #virsh vol-list default | grep baremetal${i} 2>&1> /dev/null || virsh vol-create-as default baremetal${i}.qcow2 41G --format qcow2
-    mac=$(virsh domiflist baremetal${i} | grep admin_network | awk '{ print $5 }')
-
-    if [ "$VM_COMPUTES" -gt 0 ]; then
-      capability="profile:compute"
-      VM_COMPUTES=$((VM_COMPUTES - 1))
-    else
-      capability="profile:control"
-    fi
-
-    cat >> $CONFIG/instackenv-virt.json << EOF
-    {
-      "pm_addr": "192.168.122.1",
-      "pm_user": "root",
-      "pm_password": "INSERT_STACK_USER_PRIV_KEY",
-      "pm_type": "pxe_ssh",
-      "mac": [
-        "$mac"
-      ],
-      "cpu": "$vcpus",
-      "memory": "$ramsize",
-      "disk": "41",
-      "arch": "x86_64",
-      "capabilities": "$capability"
-    },
-EOF
-  done
-
-  #truncate the last line to remove the comma behind the bracket
-  tail -n 1 $CONFIG/instackenv-virt.json | wc -c | xargs -I {} truncate $CONFIG/instackenv-virt.json -s -{}
-
-  #finally reclose the bracket and close the instackenv.json file
-  cat >> $CONFIG/instackenv-virt.json << EOF
-    }
-  ],
-  "arch": "x86_64",
-  "host-ip": "192.168.122.1",
-  "power_manager": "nova.virt.baremetal.virtual_power_driver.VirtualPowerManager",
-  "seed-ip": "",
-  "ssh-key": "INSERT_STACK_USER_PRIV_KEY",
-  "ssh-user": "root"
-}
-EOF
-  #Overwrite the tripleo-inclubator domain.xml with our own, keeping a backup.
-  if [ ! -f /usr/share/tripleo/templates/domain.xml.bak ]; then
-    /usr/bin/mv -f /usr/share/tripleo/templates/domain.xml /usr/share/tripleo/templates/domain.xml.bak
-  fi
-
-  /usr/bin/cp -f $LIB/installer/domain.xml /usr/share/tripleo/templates/domain.xml
-}
-
-##Create virtual nodes in virsh
-##params: name - String: libvirt name for VM
-##        bootdev - String: boot device for the VM
-##        disksize - Number: size of the disk in GB
-##        ovs_bridges: - List: list of ovs bridges
-##        vcpus - Number of VCPUs to use (defaults to 4)
-##        ramsize - Size of RAM for VM in MB (defaults to 8192)
-function define_vm () {
-  local vcpus ramsize
-
-  if [ -z "$5" ]; then
-    vcpus=4
-    ramsize=8388608
-  elif [ -z "$6" ]; then
-    vcpus=$5
-    ramsize=8388608
-  else
-    vcpus=$5
-    ramsize=$(($6*1024))
-  fi
-
-  # Create the libvirt storage volume
-  if virsh vol-list default | grep ${1}.qcow2 2>&1> /dev/null; then
-    volume_path=$(virsh vol-path --pool default ${1}.qcow2 || echo "/var/lib/libvirt/images/${1}.qcow2")
-    echo "Volume ${1} exists. Deleting Existing Volume $volume_path"
-    virsh vol-dumpxml ${1}.qcow2 --pool default > /dev/null || echo '' #ok for this to fail
-    touch $volume_path
-    virsh vol-delete ${1}.qcow2 --pool default
-  fi
-  virsh vol-create-as default ${1}.qcow2 ${3}G --format qcow2
-  volume_path=$(virsh vol-path --pool default ${1}.qcow2)
-  if [ ! -f $volume_path ]; then
-      echo "$volume_path Not created successfully... Aborting"
-      exit 1
-  fi
-
-  # create the VM
-  /usr/libexec/openstack-tripleo/configure-vm --name $1 \
-                                              --bootdev $2 \
-                                              --image "$volume_path" \
-                                              --diskbus sata \
-                                              --arch x86_64 \
-                                              --cpus $vcpus \
-                                              --memory $ramsize \
-                                              --libvirt-nic-driver virtio \
-                                              --baremetal-interface $4
-}
-
-##Copy over the glance images and instackenv json file
-##params: none
-function configure_undercloud {
-  local controller_nic_template compute_nic_template
-  echo
-  echo "Copying configuration files to Undercloud"
-  if [[ "$net_isolation_enabled" == "TRUE" ]]; then
-    echo -e "${blue}Network Environment set for Deployment: ${reset}"
-    cat /tmp/network-environment.yaml
-    scp ${SSH_OPTIONS[@]} /tmp/network-environment.yaml "stack@$UNDERCLOUD":
-
-    # check for ODL L3/ONOS
-    if [ "${deploy_options_array['sdn_l3']}" == 'True' ]; then
-      ext_net_type=br-ex
-    fi
-
-    if [ "${deploy_options_array['dataplane']}" == 'ovs_dpdk' ]; then
-      ovs_dpdk_bridge='br-phy'
-    else
-      ovs_dpdk_bridge=''
-    fi
-
-    if ! controller_nic_template=$(python3.4 -B $LIB/python/apex_python_utils.py nic-template -r controller -s $NETSETS -i $net_isolation_enabled -t $CONFIG/nics-template.yaml.jinja2 -n "$enabled_network_list" -e "br-ex" -af $ip_addr_family); then
-      echo -e "${red}ERROR: Failed to generate controller NIC heat template ${reset}"
-      exit 1
-    fi
-
-    if ! compute_nic_template=$(python3.4 -B $LIB/python/apex_python_utils.py nic-template -r compute -s $NETSETS -i $net_isolation_enabled -t $CONFIG/nics-template.yaml.jinja2 -n "$enabled_network_list" -e $ext_net_type -af $ip_addr_family -d "$ovs_dpdk_bridge"); then
-      echo -e "${red}ERROR: Failed to generate compute NIC heat template ${reset}"
-      exit 1
-    fi
-    ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" << EOI
-mkdir nics/
-cat > nics/controller.yaml << EOF
-$controller_nic_template
-EOF
-cat > nics/compute.yaml << EOF
-$compute_nic_template
-EOF
-EOI
-  fi
-
-  # ensure stack user on Undercloud machine has an ssh key
-  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "if [ ! -e ~/.ssh/id_rsa.pub ]; then ssh-keygen -t rsa -N '' -f ~/.ssh/id_rsa; fi"
-
-  if [ "$virtual" == "TRUE" ]; then
-
-      # copy the Undercloud VM's stack user's pub key to
-      # root's auth keys so that Undercloud can control
-      # vm power on the hypervisor
-      ssh ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "cat /home/stack/.ssh/id_rsa.pub" >> /root/.ssh/authorized_keys
-
-      DEPLOY_OPTIONS+=" --libvirt-type qemu"
-      INSTACKENV=$CONFIG/instackenv-virt.json
-
-      # upload instackenv file to Undercloud for virtual deployment
-      scp ${SSH_OPTIONS[@]} $INSTACKENV "stack@$UNDERCLOUD":instackenv.json
-  fi
-
-  # allow stack to control power management on the hypervisor via sshkey
-  # only if this is a virtual deployment
-  if [ "$virtual" == "TRUE" ]; then
-      ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
-while read -r line; do
-  stack_key=\${stack_key}\\\\\\\\n\${line}
-done < <(cat ~/.ssh/id_rsa)
-stack_key=\$(echo \$stack_key | sed 's/\\\\\\\\n//')
-sed -i 's~INSERT_STACK_USER_PRIV_KEY~'"\$stack_key"'~' instackenv.json
-EOI
-  fi
-
-  # copy stack's ssh key to this users authorized keys
-  ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "cat /home/stack/.ssh/id_rsa.pub" >> ~/.ssh/authorized_keys
-
-  # disable requiretty for sudo
-  ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "sed -i 's/Defaults\s*requiretty//'" /etc/sudoers
-
-  # configure undercloud on Undercloud VM
-  echo "Running undercloud configuration."
-  echo "Logging undercloud configuration to undercloud:/home/stack/apex-undercloud-install.log"
-  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" << EOI
-if [[ "$net_isolation_enabled" == "TRUE" ]]; then
-  sed -i 's/#local_ip/local_ip/' undercloud.conf
-  sed -i 's/#network_gateway/network_gateway/' undercloud.conf
-  sed -i 's/#network_cidr/network_cidr/' undercloud.conf
-  sed -i 's/#dhcp_start/dhcp_start/' undercloud.conf
-  sed -i 's/#dhcp_end/dhcp_end/' undercloud.conf
-  sed -i 's/#inspection_iprange/inspection_iprange/' undercloud.conf
-  sed -i 's/#undercloud_debug/undercloud_debug/' undercloud.conf
-
-  openstack-config --set undercloud.conf DEFAULT local_ip ${admin_network_provisioner_ip}/${admin_network_cidr##*/}
-  openstack-config --set undercloud.conf DEFAULT network_gateway ${admin_network_provisioner_ip}
-  openstack-config --set undercloud.conf DEFAULT network_cidr ${admin_network_cidr}
-  openstack-config --set undercloud.conf DEFAULT dhcp_start ${admin_network_dhcp_range%%,*}
-  openstack-config --set undercloud.conf DEFAULT dhcp_end ${admin_network_dhcp_range##*,}
-  openstack-config --set undercloud.conf DEFAULT inspection_iprange ${admin_network_introspection_range}
-  openstack-config --set undercloud.conf DEFAULT undercloud_debug false
-
-fi
-
-sudo sed -i '/CephClusterFSID:/c\\  CephClusterFSID: \\x27$(cat /proc/sys/kernel/random/uuid)\\x27' /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml
-sudo sed -i '/CephMonKey:/c\\  CephMonKey: \\x27'"\$(ceph-authtool --gen-print-key)"'\\x27' /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml
-sudo sed -i '/CephAdminKey:/c\\  CephAdminKey: \\x27'"\$(ceph-authtool --gen-print-key)"'\\x27' /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml
-
-# we assume that packages will not need to be updated with undercloud install
-# and that it will be used only to configure the undercloud
-# packages updates would need to be handled manually with yum update
-sudo cp -f /usr/share/diskimage-builder/elements/yum/bin/install-packages /usr/share/diskimage-builder/elements/yum/bin/install-packages.bak
-cat << 'EOF' | sudo tee /usr/share/diskimage-builder/elements/yum/bin/install-packages > /dev/null
-#!/bin/sh
-exit 0
-EOF
-
-openstack undercloud install &> apex-undercloud-install.log || {
-    # cat the undercloud install log incase it fails
-    echo "ERROR: openstack undercloud install has failed. Dumping Log:"
-    cat apex-undercloud-install.log
-    exit 1
-}
-
-sleep 30
-sudo systemctl restart openstack-glance-api
-sudo systemctl restart openstack-nova-conductor
-sudo systemctl restart openstack-nova-compute
-
-sudo sed -i '/num_engine_workers/c\num_engine_workers = 2' /etc/heat/heat.conf
-sudo sed -i '/#workers\s=/c\workers = 2' /etc/heat/heat.conf
-sudo systemctl restart openstack-heat-engine
-sudo systemctl restart openstack-heat-api
-EOI
-
-# configure external network
-  ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" << EOI
-if [[ "$public_network_vlan" != "native" ]]; then
-  cat <<EOF > /etc/sysconfig/network-scripts/ifcfg-vlan${public_network_vlan}
-DEVICE=vlan${public_network_vlan}
-ONBOOT=yes
-DEVICETYPE=ovs
-TYPE=OVSIntPort
-BOOTPROTO=static
-IPADDR=${public_network_provisioner_ip}
-PREFIX=${public_network_cidr##*/}
-OVS_BRIDGE=br-ctlplane
-OVS_OPTIONS="tag=${public_network_vlan}"
-EOF
-  ifup vlan${public_network_vlan}
-else
-  if ! ip a s eth2 | grep ${public_network_provisioner_ip} > /dev/null; then
-      ip a a ${public_network_provisioner_ip}/${public_network_cidr##*/} dev eth2
-      ip link set up dev eth2
-  fi
-fi
-EOI
-
-# WORKAROUND: must restart the above services to fix sync problem with nova compute manager
-# TODO: revisit and file a bug if necessary. This should eventually be removed
-# as well as glance api problem
-echo -e "${blue}INFO: Sleeping 15 seconds while services come back from restart${reset}"
-sleep 15
-
-}
-
-##preping it for deployment and launch the deploy
-##params: none
-function undercloud_prep_overcloud_deploy {
-  if [[ "${#deploy_options_array[@]}" -eq 0 || "${deploy_options_array['sdn_controller']}" == 'opendaylight' ]]; then
-    if [ "${deploy_options_array['sdn_l3']}" == 'True' ]; then
-      DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight_l3.yaml"
-    elif [ "${deploy_options_array['sfc']}" == 'True' ]; then
-      DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight_sfc.yaml"
-    elif [ "${deploy_options_array['vpn']}" == 'True' ]; then
-      DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight_sdnvpn.yaml"
-    elif [ "${deploy_options_array['vpp']}" == 'True' ]; then
-      DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight_fdio.yaml"
-    else
-      DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight.yaml"
-    fi
-    SDN_IMAGE=opendaylight
-    if [ "${deploy_options_array['sfc']}" == 'True' ]; then
-      SDN_IMAGE+=-sfc
-      if [ ! -f $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 ]; then
-          echo "${red} $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 is required to execute an SFC deployment."
-          echo "Please install the opnfv-apex-opendaylight-sfc package to provide this overcloud image for deployment.${reset}"
-          exit 1
-      fi
-    fi
-  elif [ "${deploy_options_array['sdn_controller']}" == 'opendaylight-external' ]; then
-    DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight-external.yaml"
-    SDN_IMAGE=opendaylight
-  elif [ "${deploy_options_array['sdn_controller']}" == 'onos' ]; then
-    DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/onos.yaml"
-    SDN_IMAGE=onos
-  elif [ "${deploy_options_array['sdn_controller']}" == 'opencontrail' ]; then
-    echo -e "${red}ERROR: OpenContrail is currently unsupported...exiting${reset}"
-    exit 1
-  elif [[ -z "${deploy_options_array['sdn_controller']}" || "${deploy_options_array['sdn_controller']}" == 'False' ]]; then
-    echo -e "${blue}INFO: SDN Controller disabled...will deploy nosdn scenario${reset}"
-    SDN_IMAGE=opendaylight
-  else
-    echo "${red}Invalid sdn_controller: ${deploy_options_array['sdn_controller']}${reset}"
-    echo "${red}Valid choices are opendaylight, opendaylight-external, onos, opencontrail, False, or null${reset}"
-    exit 1
-  fi
-
-
-
-  # Make sure the correct overcloud image is available
-  if [ ! -f $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 ]; then
-      echo "${red} $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 is required to execute your deployment."
-      echo "Both ONOS and OpenDaylight are currently deployed from this image."
-      echo "Please install the opnfv-apex package to provide this overcloud image for deployment.${reset}"
-      exit 1
-  fi
-
-  echo "Copying overcloud image to Undercloud"
-  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "rm -f overcloud-full.qcow2"
-  scp ${SSH_OPTIONS[@]} $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 "stack@$UNDERCLOUD":overcloud-full.qcow2
-
-  # Install ovs-dpdk inside the overcloud image if it is enabled.
-  if [ "${deploy_options_array['dataplane']}" == 'ovs_dpdk' ]; then
-    # install dpdk packages before ovs
-    echo -e "${blue}INFO: Enabling kernel modules for dpdk inside overcloud image${reset}"
-
-    ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
-      cat << EOF > vfio_pci.modules
-#!/bin/bash
-exec /sbin/modprobe vfio_pci >/dev/null 2>&1
-EOF
-
-      cat << EOF > uio_pci_generic.modules
-#!/bin/bash
-exec /sbin/modprobe uio_pci_generic >/dev/null 2>&1
-EOF
-
-      LIBGUESTFS_BACKEND=direct virt-customize --upload vfio_pci.modules:/etc/sysconfig/modules/ \
-                                               --upload uio_pci_generic.modules:/etc/sysconfig/modules/ \
-                                               --run-command "chmod 0755 /etc/sysconfig/modules/vfio_pci.modules" \
-                                               --run-command "chmod 0755 /etc/sysconfig/modules/uio_pci_generic.modules" \
-                                               --run-command "yum install -y /root/dpdk_rpms/*" \
-                                               -a overcloud-full.qcow2
-EOI
-  elif [ "${deploy_options_array['dataplane']}" != 'ovs' ]; then
-    echo "${red}${deploy_options_array['dataplane']} not supported${reset}"
-    exit 1
-  fi
-
-  # Set ODL version accordingly
-  if [[ "${deploy_options_array['sdn_controller']}" == 'opendaylight' && "${deploy_options_array['odl_version']}" == 'boron' ]]; then
-    ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
-      LIBGUESTFS_BACKEND=direct virt-customize --run-command "yum -y remove opendaylight" \
-                                               --run-command "yum -y install /root/boron/*" \
-                                               -a overcloud-full.qcow2
-EOI
-  fi
-
-  # Add performance deploy options if they have been set
-  if [ ! -z "${deploy_options_array['performance']}" ]; then
-
-    # Remove previous kernel args files per role
-    ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "rm -f Compute-kernel_params.txt"
-    ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "rm -f Controller-kernel_params.txt"
-
-    # Push performance options to subscript to modify per-role images as needed
-    for option in "${performance_options[@]}" ; do
-      echo -e "${blue}Setting performance option $option${reset}"
-      ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "bash build_perf_image.sh $option"
-    done
-
-    # Build IPA kernel option ramdisks
-    ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" <<EOI
-/bin/cp -f /home/stack/ironic-python-agent.initramfs /root/
-mkdir -p ipa/
-pushd ipa
-gunzip -c ../ironic-python-agent.initramfs | cpio -i
-if [ ! -f /home/stack/Compute-kernel_params.txt ]; then
-  touch /home/stack/Compute-kernel_params.txt
-  chown stack /home/stack/Compute-kernel_params.txt
-fi
-/bin/cp -f /home/stack/Compute-kernel_params.txt tmp/kernel_params.txt
-echo "Compute params set: "
-cat tmp/kernel_params.txt
-/bin/cp -f /root/image.py usr/lib/python2.7/site-packages/ironic_python_agent/extensions/image.py
-/bin/cp -f /root/image.pyc usr/lib/python2.7/site-packages/ironic_python_agent/extensions/image.pyc
-find . | cpio -o -H newc | gzip > /home/stack/Compute-ironic-python-agent.initramfs
-chown stack /home/stack/Compute-ironic-python-agent.initramfs
-if [ ! -f /home/stack/Controller-kernel_params.txt ]; then
-  touch /home/stack/Controller-kernel_params.txt
-  chown stack /home/stack/Controller-kernel_params.txt
-fi
-/bin/cp -f /home/stack/Controller-kernel_params.txt tmp/kernel_params.txt
-echo "Controller params set: "
-cat tmp/kernel_params.txt
-find . | cpio -o -H newc | gzip > /home/stack/Controller-ironic-python-agent.initramfs
-chown stack /home/stack/Controller-ironic-python-agent.initramfs
-popd
-/bin/rm -rf ipa/
-EOI
-
-    DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/numa.yaml"
-  fi
-
-  # make sure ceph is installed
-  DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml"
-
-  # scale compute nodes according to inventory
-  total_nodes=$(ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "cat /home/stack/instackenv.json | grep -c memory")
-
-  # check if HA is enabled
-  if [[ "$ha_enabled" == "True" ]]; then
-     DEPLOY_OPTIONS+=" --control-scale 3"
-     compute_nodes=$((total_nodes - 3))
-     DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/puppet-pacemaker.yaml"
-  else
-     compute_nodes=$((total_nodes - 1))
-  fi
-
-  if [ "$compute_nodes" -le 0 ]; then
-    echo -e "${red}ERROR: Invalid number of compute nodes: ${compute_nodes}. Check your inventory file.${reset}"
-    exit 1
-  else
-    echo -e "${blue}INFO: Number of compute nodes set for deployment: ${compute_nodes}${reset}"
-    DEPLOY_OPTIONS+=" --compute-scale ${compute_nodes}"
-  fi
-
-  if [[ "$net_isolation_enabled" == "TRUE" ]]; then
-     #DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/network-isolation.yaml"
-     DEPLOY_OPTIONS+=" -e network-environment.yaml"
-  fi
-
-  if [[ "$ha_enabled" == "True" ]] || [[ "$net_isolation_enabled" == "TRUE" ]]; then
-     DEPLOY_OPTIONS+=" --ntp-server $ntp_server"
-  fi
-
-  if [[ ! "$virtual" == "TRUE" ]]; then
-     DEPLOY_OPTIONS+=" --control-flavor control --compute-flavor compute"
-  else
-     DEPLOY_OPTIONS+=" -e virtual-environment.yaml"
-  fi
-
-  DEPLOY_OPTIONS+=" -e opnfv-environment.yaml"
-
-  echo -e "${blue}INFO: Deploy options set:\n${DEPLOY_OPTIONS}${reset}"
-
-  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
-if [ "$debug" == 'TRUE' ]; then
-    LIBGUESTFS_BACKEND=direct virt-customize -a overcloud-full.qcow2 --root-password password:opnfvapex
-fi
-
-source stackrc
-set -o errexit
-echo "Uploading overcloud glance images"
-openstack overcloud image upload
-
-echo "Configuring undercloud and discovering nodes"
-openstack baremetal import --json instackenv.json
-openstack baremetal configure boot
-bash -x set_perf_images.sh ${performance_roles[@]}
-#if [[ -z "$virtual" ]]; then
-#  openstack baremetal introspection bulk start
-#fi
-echo "Configuring flavors"
-for flavor in baremetal control compute; do
-  echo -e "${blue}INFO: Updating flavor: \${flavor}${reset}"
-  if openstack flavor list | grep \${flavor}; then
-    openstack flavor delete \${flavor}
-  fi
-  openstack flavor create --id auto --ram 4096 --disk 39 --vcpus 1 \${flavor}
-  if ! openstack flavor list | grep \${flavor}; then
-    echo -e "${red}ERROR: Unable to create flavor \${flavor}${reset}"
-  fi
-done
-openstack flavor set --property "cpu_arch"="x86_64" --property "capabilities:boot_option"="local" baremetal
-openstack flavor set --property "cpu_arch"="x86_64" --property "capabilities:boot_option"="local" --property "capabilities:profile"="control" control
-openstack flavor set --property "cpu_arch"="x86_64" --property "capabilities:boot_option"="local" --property "capabilities:profile"="compute" compute
-echo "Configuring nameserver on ctlplane network"
-dns_server_ext=''
-for dns_server in ${dns_servers}; do
-  dns_server_ext="\${dns_server_ext} --dns-nameserver \${dns_server}"
-done
-neutron subnet-update \$(neutron subnet-list | grep -Ev "id|tenant|external|storage" | grep -v \\\\-\\\\- | awk {'print \$2'}) \${dns_server_ext}
-echo "Executing overcloud deployment, this should run for an extended period without output."
-sleep 60 #wait for Hypervisor stats to check-in to nova
-# save deploy command so it can be used for debugging
-cat > deploy_command << EOF
-openstack overcloud deploy --templates $DEPLOY_OPTIONS --timeout 90
-EOF
-EOI
-
-  if [ "$interactive" == "TRUE" ]; then
-    if ! prompt_user "Overcloud Deployment"; then
-      echo -e "${blue}INFO: User requests exit${reset}"
-      exit 0
-    fi
-  fi
-
-  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
-source stackrc
-openstack overcloud deploy --templates $DEPLOY_OPTIONS --timeout 90
-if ! heat stack-list | grep CREATE_COMPLETE 1>/dev/null; then
-  $(typeset -f debug_stack)
-  debug_stack
-  exit 1
-fi
-EOI
-
-  # Configure DPDK
-  if [ "${deploy_options_array['dataplane']}" == 'ovs_dpdk' ]; then
-    ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI || (echo "DPDK config failed, exiting..."; exit 1)
-source stackrc
-set -o errexit
-for node in \$(nova list | grep novacompute | grep -Eo "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+"); do
-echo "Running DPDK test app on \$node"
-ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
-set -o errexit
-sudo dpdk_helloworld --no-pci
-sudo dpdk_nic_bind -s
-EOF
-done
-EOI
-  fi
-
-  if [ "$debug" == 'TRUE' ]; then
-      ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
-source overcloudrc
-echo "Keystone Endpoint List:"
-openstack endpoint list
-echo "Keystone Service List"
-openstack service list
-cinder quota-show \$(openstack project list | grep admin | awk {'print \$2'})
-EOI
-  fi
-}
-
-##Post configuration after install
-##params: none
-function configure_post_install {
-  local opnfv_attach_networks ovs_ip ip_range net_cidr tmp_ip
-  opnfv_attach_networks="admin_network public_network"
-
-  echo -e "${blue}INFO: Post Install Configuration Running...${reset}"
-
-  echo -e "${blue}INFO: Configuring ssh for root to overcloud nodes...${reset}"
-  # copy host key to instack
-  scp ${SSH_OPTIONS[@]} /root/.ssh/id_rsa.pub "stack@$UNDERCLOUD":jumphost_id_rsa.pub
-
-  # add host key to overcloud nodes authorized keys
-  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" << EOI
-source stackrc
-nodes=\$(nova list | grep -Eo "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+")
-for node in \$nodes; do
-cat ~/jumphost_id_rsa.pub | ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" 'cat >> ~/.ssh/authorized_keys'
-done
-EOI
-
-  if [ "${deploy_options_array['dataplane']}" == 'ovs_dpdk' ]; then
-    echo -e "${blue}INFO: Bringing up br-phy and ovs-agent for dpdk compute nodes...${reset}"
-    compute_nodes=$(undercloud_connect stack "source stackrc; nova list | grep compute | wc -l")
-    i=0
-    while [ "$i" -lt "$compute_nodes" ]; do
-      overcloud_connect compute${i} "sudo ifup br-phy; sudo systemctl restart neutron-openvswitch-agent"
-      i=$((i + 1))
-    done
-  fi
-
-  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
-source overcloudrc
-set -o errexit
-echo "Configuring Neutron external network"
-neutron net-create external --router:external=True --tenant-id \$(openstack project show service | grep id | awk '{ print \$4 }')
-neutron subnet-create --name external-net --tenant-id \$(openstack project show service | grep id | awk '{ print \$4 }') --disable-dhcp external --gateway ${public_network_gateway} --allocation-pool start=${public_network_floating_ip_range%%,*},end=${public_network_floating_ip_range##*,} ${public_network_cidr}
-
-echo "Removing sahara endpoint and service"
-sahara_service_id=\$(openstack service list | grep sahara | cut -d ' ' -f 2)
-sahara_endpoint_id=\$(openstack endpoint list | grep sahara | cut -d ' ' -f 2)
-openstack endpoint delete \$sahara_endpoint_id
-openstack service delete \$sahara_service_id
-
-echo "Removing swift endpoint and service"
-swift_service_id=\$(openstack service list | grep swift | cut -d ' ' -f 2)
-swift_endpoint_id=\$(openstack endpoint list | grep swift | cut -d ' ' -f 2)
-openstack endpoint delete \$swift_endpoint_id
-openstack service delete \$swift_service_id
-
-if [ "${deploy_options_array['congress']}" == 'True' ]; then
-    for s in nova neutronv2 ceilometer cinder glancev2 keystone; do
-        openstack congress datasource create \$s "\$s" \\
-            --config username=\$OS_USERNAME \\
-            --config tenant_name=\$OS_TENANT_NAME \\
-            --config password=\$OS_PASSWORD \\
-            --config auth_url=\$OS_AUTH_URL
-    done
-fi
-EOI
-
-  echo -e "${blue}INFO: Checking if OVS bridges have IP addresses...${reset}"
-  for network in ${opnfv_attach_networks}; do
-    ovs_ip=$(find_ip ${NET_MAP[$network]})
-    tmp_ip=''
-    if [ -n "$ovs_ip" ]; then
-      echo -e "${blue}INFO: OVS Bridge ${NET_MAP[$network]} has IP address ${ovs_ip}${reset}"
-    else
-      echo -e "${blue}INFO: OVS Bridge ${NET_MAP[$network]} missing IP, will configure${reset}"
-      # use last IP of allocation pool
-      eval "ip_range=\${${network}_usable_ip_range}"
-      ovs_ip=${ip_range##*,}
-      eval "net_cidr=\${${network}_cidr}"
-      sudo ip addr add ${ovs_ip}/${net_cidr##*/} dev ${NET_MAP[$network]}
-      sudo ip link set up ${NET_MAP[$network]}
-      tmp_ip=$(find_ip ${NET_MAP[$network]})
-      if [ -n "$tmp_ip" ]; then
-        echo -e "${blue}INFO: OVS Bridge ${NET_MAP[$network]} IP set: ${tmp_ip}${reset}"
-        continue
-      else
-        echo -e "${red}ERROR: Unable to set OVS Bridge ${NET_MAP[$network]} with IP: ${ovs_ip}${reset}"
-        return 1
-      fi
-    fi
-  done
-
-  # for virtual, we NAT public network through Undercloud
-  if [ "$virtual" == "TRUE" ]; then
-    if ! configure_undercloud_nat ${public_network_cidr}; then
-      echo -e "${red}ERROR: Unable to NAT undercloud with external net: ${public_network_cidr}${reset}"
-      exit 1
-    else
-      echo -e "${blue}INFO: Undercloud VM has been setup to NAT Overcloud public network${reset}"
-    fi
-  fi
-
-  # for sfc deployments we need the vxlan workaround
-  if [ "${deploy_options_array['sfc']}" == 'True' ]; then
-      ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
-source stackrc
-set -o errexit
-for node in \$(nova list | grep -Eo "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+"); do
-ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
-sudo ifconfig br-int up
-sudo ip route add 123.123.123.0/24 dev br-int
-EOF
-done
-EOI
-  fi
-
-  # Collect deployment logs
-  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
-mkdir -p ~/deploy_logs
-rm -rf deploy_logs/*
-source stackrc
-set -o errexit
-for node in \$(nova list | grep -Eo "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+"); do
- ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
- sudo cp /var/log/messages /home/heat-admin/messages.log
- sudo chown heat-admin /home/heat-admin/messages.log
-EOF
-scp ${SSH_OPTIONS[@]} heat-admin@\$node:/home/heat-admin/messages.log ~/deploy_logs/\$node.messages.log
-if [ "$debug" == "TRUE" ]; then
-    nova list --ip \$node
-    echo "---------------------------"
-    echo "-----/var/log/messages-----"
-    echo "---------------------------"
-    cat ~/deploy_logs/\$node.messages.log
-    echo "---------------------------"
-    echo "----------END LOG----------"
-    echo "---------------------------"
-fi
- ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
- sudo rm -f /home/heat-admin/messages.log
-EOF
-done
-
-# Print out the undercloud IP and dashboard URL
-source stackrc
-echo "Undercloud IP: $UNDERCLOUD, please connect by doing 'opnfv-util undercloud'"
-echo "Overcloud dashboard available at http://\$(heat output-show overcloud PublicVip | sed 's/"//g')/dashboard"
-EOI
-
-}
-
 display_usage() {
   echo -e "Usage:\n$0 [arguments] \n"
   echo -e "   -d|--deploy-settings : Full path to deploy settings yaml file. Optional.  Defaults to null"
@@ -1195,8 +209,6 @@ parse_cmdline() {
 
 }
 
-##END FUNCTIONS
-
 main() {
   parse_cmdline "$@"
   echo -e "${blue}INFO: Parsing network settings file...${reset}"
@@ -1216,7 +228,7 @@ main() {
     parse_inventory_file
   fi
   configure_undercloud
-  undercloud_prep_overcloud_deploy
+  overcloud_deploy
   if [ "$post_config" == "TRUE" ]; then
     if ! configure_post_install; then
       echo -e "${red}ERROR:Post Install Configuration Failed, Exiting.${reset}"
index 365f8e3..2ace997 100644 (file)
@@ -257,3 +257,33 @@ function prompt_user {
     fi
   done
 }
+
+##checks if prefix exists in string
+##params: string, prefix
+##usage: contains_prefix "deploy_setting_launcher=1" "deploy_setting"
+contains_prefix() {
+  local mystr=$1
+  local prefix=$2
+  if echo $mystr | grep -E "^$prefix.*$" > /dev/null; then
+    return 0
+  else
+    return 1
+  fi
+}
+
+##verify internet connectivity
+#params: none
+function verify_internet {
+  if ping -c 2 $ping_site > /dev/null; then
+    if ping -c 2 www.google.com > /dev/null; then
+      echo "${blue}Internet connectivity detected${reset}"
+      return 0
+    else
+      echo "${red}Internet connectivity detected, but DNS lookup failed${reset}"
+      return 1
+    fi
+  else
+    echo "${red}No internet connectivity detected${reset}"
+    return 1
+  fi
+}
diff --git a/lib/configure-deps-functions.sh b/lib/configure-deps-functions.sh
new file mode 100755 (executable)
index 0000000..06a4c72
--- /dev/null
@@ -0,0 +1,144 @@
+#!/usr/bin/env bash
+##############################################################################
+# Copyright (c) 2015 Tim Rozet (Red Hat), Dan Radez (Red Hat) and others.
+#
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
+
+##download dependencies if missing and configure host
+#params: none
+function configure_deps {
+  if ! verify_internet; then
+    echo "${red}Will not download dependencies${reset}"
+    internet=false
+  fi
+
+  # verify ip forwarding
+  if sysctl net.ipv4.ip_forward | grep 0; then
+    sudo sysctl -w net.ipv4.ip_forward=1
+    sudo sh -c "echo 'net.ipv4.ip_forward = 1' >> /etc/sysctl.conf"
+  fi
+
+  # ensure no dhcp server is running on jumphost
+  if ! sudo systemctl status dhcpd | grep dead; then
+    echo "${red}WARN: DHCP Server detected on jumphost, disabling...${reset}"
+    sudo systemctl stop dhcpd
+    sudo systemctl disable dhcpd
+  fi
+
+  # ensure networks are configured
+  systemctl status libvirtd || systemctl start libvirtd
+  systemctl status openvswitch || systemctl start openvswitch
+
+  # If flat we only use admin network
+  if [[ "$net_isolation_enabled" == "FALSE" ]]; then
+    virsh_enabled_networks="admin_network"
+    enabled_network_list="admin_network"
+  # For baremetal we only need to create/attach Undercloud to admin and public
+  elif [ "$virtual" == "FALSE" ]; then
+    virsh_enabled_networks="admin_network public_network"
+  else
+    virsh_enabled_networks=$enabled_network_list
+  fi
+
+  # ensure default network is configured correctly
+  libvirt_dir="/usr/share/libvirt/networks"
+  virsh net-list --all | grep default || virsh net-define ${libvirt_dir}/default.xml
+  virsh net-list --all | grep -E "default\s+active" > /dev/null || virsh net-start default
+  virsh net-list --all | grep -E "default\s+active\s+yes" > /dev/null || virsh net-autostart --network default
+
+  if [[ -z "$virtual" || "$virtual" == "FALSE" ]]; then
+    for network in ${enabled_network_list}; do
+      echo "${blue}INFO: Creating Virsh Network: $network & OVS Bridge: ${NET_MAP[$network]}${reset}"
+      ovs-vsctl list-br | grep "^${NET_MAP[$network]}$" > /dev/null || ovs-vsctl add-br ${NET_MAP[$network]}
+      virsh net-list --all | grep $network > /dev/null || (cat > ${libvirt_dir}/apex-virsh-net.xml && virsh net-define ${libvirt_dir}/apex-virsh-net.xml) << EOF
+<network>
+  <name>$network</name>
+  <forward mode='bridge'/>
+  <bridge name='${NET_MAP[$network]}'/>
+  <virtualport type='openvswitch'/>
+</network>
+EOF
+      if ! (virsh net-list --all | grep $network > /dev/null); then
+          echo "${red}ERROR: unable to create network: ${network}${reset}"
+          exit 1;
+      fi
+      rm -f ${libvirt_dir}/apex-virsh-net.xml &> /dev/null;
+      virsh net-list | grep -E "$network\s+active" > /dev/null || virsh net-start $network
+      virsh net-list | grep -E "$network\s+active\s+yes" > /dev/null || virsh net-autostart --network $network
+    done
+
+    echo -e "${blue}INFO: Bridges set: ${reset}"
+    ovs-vsctl list-br
+
+    # bridge interfaces to correct OVS instances for baremetal deployment
+    for network in ${enabled_network_list}; do
+      if [[ "$network" != "admin_network" && "$network" != "public_network" ]]; then
+        continue
+      fi
+      this_interface=$(eval echo \${${network}_bridged_interface})
+      # check if this a bridged interface for this network
+      if [[ ! -z "$this_interface" || "$this_interface" != "none" ]]; then
+        if ! attach_interface_to_ovs ${NET_MAP[$network]} ${this_interface} ${network}; then
+          echo -e "${red}ERROR: Unable to bridge interface ${this_interface} to bridge ${NET_MAP[$network]} for enabled network: ${network}${reset}"
+          exit 1
+        else
+          echo -e "${blue}INFO: Interface ${this_interface} bridged to bridge ${NET_MAP[$network]} for enabled network: ${network}${reset}"
+        fi
+      else
+        echo "${red}ERROR: Unable to determine interface to bridge to for enabled network: ${network}${reset}"
+        exit 1
+      fi
+    done
+  else
+    for network in ${OPNFV_NETWORK_TYPES}; do
+      echo "${blue}INFO: Creating Virsh Network: $network${reset}"
+      virsh net-list --all | grep $network > /dev/null || (cat > ${libvirt_dir}/apex-virsh-net.xml && virsh net-define ${libvirt_dir}/apex-virsh-net.xml) << EOF
+<network ipv6='yes'>
+<name>$network</name>
+<bridge name='${NET_MAP[$network]}'/>
+</network>
+EOF
+      if ! (virsh net-list --all | grep $network > /dev/null); then
+          echo "${red}ERROR: unable to create network: ${network}${reset}"
+          exit 1;
+      fi
+      rm -f ${libvirt_dir}/apex-virsh-net.xml &> /dev/null;
+      virsh net-list | grep -E "$network\s+active" > /dev/null || virsh net-start $network
+      virsh net-list | grep -E "$network\s+active\s+yes" > /dev/null || virsh net-autostart --network $network
+    done
+
+    echo -e "${blue}INFO: Bridges set: ${reset}"
+    brctl show
+  fi
+
+  echo -e "${blue}INFO: virsh networks set: ${reset}"
+  virsh net-list
+
+  # ensure storage pool exists and is started
+  virsh pool-list --all | grep default > /dev/null || virsh pool-define-as --name default dir --target /var/lib/libvirt/images
+  virsh pool-list | grep -Eo "default\s+active" > /dev/null || (virsh pool-autostart default; virsh pool-start default)
+
+  if ! egrep '^flags.*(vmx|svm)' /proc/cpuinfo > /dev/null; then
+    echo "${red}virtualization extensions not found, kvm kernel module insertion may fail.\n  \
+Are you sure you have enabled vmx in your bios or hypervisor?${reset}"
+  fi
+
+  if ! lsmod | grep kvm > /dev/null; then modprobe kvm; fi
+  if ! lsmod | grep kvm_intel > /dev/null; then modprobe kvm_intel; fi
+
+  if ! lsmod | grep kvm > /dev/null; then
+    echo "${red}kvm kernel modules not loaded!${reset}"
+    return 1
+  fi
+
+  ##sshkeygen for root
+  if [ ! -e ~/.ssh/id_rsa.pub ]; then
+    ssh-keygen -t rsa -N "" -f ~/.ssh/id_rsa
+  fi
+
+  echo "${blue}All dependencies installed and running${reset}"
+}
diff --git a/lib/overcloud-deploy-functions.sh b/lib/overcloud-deploy-functions.sh
new file mode 100755 (executable)
index 0000000..067a641
--- /dev/null
@@ -0,0 +1,280 @@
+#!/usr/bin/env bash
+##############################################################################
+# Copyright (c) 2015 Tim Rozet (Red Hat), Dan Radez (Red Hat) and others.
+#
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
+
+##preping it for deployment and launch the deploy
+##params: none
+function overcloud_deploy {
+  if [[ "${#deploy_options_array[@]}" -eq 0 || "${deploy_options_array['sdn_controller']}" == 'opendaylight' ]]; then
+    if [ "${deploy_options_array['sdn_l3']}" == 'True' ]; then
+      DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight_l3.yaml"
+    elif [ "${deploy_options_array['sfc']}" == 'True' ]; then
+      DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight_sfc.yaml"
+    elif [ "${deploy_options_array['vpn']}" == 'True' ]; then
+      DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight_sdnvpn.yaml"
+    elif [ "${deploy_options_array['vpp']}" == 'True' ]; then
+      DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight_fdio.yaml"
+    else
+      DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight.yaml"
+    fi
+    SDN_IMAGE=opendaylight
+    if [ "${deploy_options_array['sfc']}" == 'True' ]; then
+      SDN_IMAGE+=-sfc
+      if [ ! -f $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 ]; then
+          echo "${red} $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 is required to execute an SFC deployment."
+          echo "Please install the opnfv-apex-opendaylight-sfc package to provide this overcloud image for deployment.${reset}"
+          exit 1
+      fi
+    fi
+  elif [ "${deploy_options_array['sdn_controller']}" == 'opendaylight-external' ]; then
+    DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight-external.yaml"
+    SDN_IMAGE=opendaylight
+  elif [ "${deploy_options_array['sdn_controller']}" == 'onos' ]; then
+    DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/onos.yaml"
+    SDN_IMAGE=onos
+  elif [ "${deploy_options_array['sdn_controller']}" == 'opencontrail' ]; then
+    echo -e "${red}ERROR: OpenContrail is currently unsupported...exiting${reset}"
+    exit 1
+  elif [[ -z "${deploy_options_array['sdn_controller']}" || "${deploy_options_array['sdn_controller']}" == 'False' ]]; then
+    echo -e "${blue}INFO: SDN Controller disabled...will deploy nosdn scenario${reset}"
+    SDN_IMAGE=opendaylight
+  else
+    echo "${red}Invalid sdn_controller: ${deploy_options_array['sdn_controller']}${reset}"
+    echo "${red}Valid choices are opendaylight, opendaylight-external, onos, opencontrail, False, or null${reset}"
+    exit 1
+  fi
+
+
+
+  # Make sure the correct overcloud image is available
+  if [ ! -f $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 ]; then
+      echo "${red} $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 is required to execute your deployment."
+      echo "Both ONOS and OpenDaylight are currently deployed from this image."
+      echo "Please install the opnfv-apex package to provide this overcloud image for deployment.${reset}"
+      exit 1
+  fi
+
+  echo "Copying overcloud image to Undercloud"
+  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "rm -f overcloud-full.qcow2"
+  scp ${SSH_OPTIONS[@]} $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 "stack@$UNDERCLOUD":overcloud-full.qcow2
+
+  # Install ovs-dpdk inside the overcloud image if it is enabled.
+  if [ "${deploy_options_array['dataplane']}" == 'ovs_dpdk' ]; then
+    # install dpdk packages before ovs
+    echo -e "${blue}INFO: Enabling kernel modules for dpdk inside overcloud image${reset}"
+
+    ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+      cat << EOF > vfio_pci.modules
+#!/bin/bash
+exec /sbin/modprobe vfio_pci >/dev/null 2>&1
+EOF
+
+      cat << EOF > uio_pci_generic.modules
+#!/bin/bash
+exec /sbin/modprobe uio_pci_generic >/dev/null 2>&1
+EOF
+
+      LIBGUESTFS_BACKEND=direct virt-customize --upload vfio_pci.modules:/etc/sysconfig/modules/ \
+                                               --upload uio_pci_generic.modules:/etc/sysconfig/modules/ \
+                                               --run-command "chmod 0755 /etc/sysconfig/modules/vfio_pci.modules" \
+                                               --run-command "chmod 0755 /etc/sysconfig/modules/uio_pci_generic.modules" \
+                                               --run-command "yum install -y /root/dpdk_rpms/*" \
+                                               -a overcloud-full.qcow2
+EOI
+  elif [ "${deploy_options_array['dataplane']}" != 'ovs' ]; then
+    echo "${red}${deploy_options_array['dataplane']} not supported${reset}"
+    exit 1
+  fi
+
+  # Set ODL version accordingly
+  if [[ "${deploy_options_array['sdn_controller']}" == 'opendaylight' && "${deploy_options_array['odl_version']}" == 'boron' ]]; then
+    ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+      LIBGUESTFS_BACKEND=direct virt-customize --run-command "yum -y remove opendaylight" \
+                                               --run-command "yum -y install /root/boron/*" \
+                                               -a overcloud-full.qcow2
+EOI
+  fi
+
+  # Add performance deploy options if they have been set
+  if [ ! -z "${deploy_options_array['performance']}" ]; then
+
+    # Remove previous kernel args files per role
+    ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "rm -f Compute-kernel_params.txt"
+    ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "rm -f Controller-kernel_params.txt"
+
+    # Push performance options to subscript to modify per-role images as needed
+    for option in "${performance_options[@]}" ; do
+      echo -e "${blue}Setting performance option $option${reset}"
+      ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "bash build_perf_image.sh $option"
+    done
+
+    # Build IPA kernel option ramdisks
+    ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" <<EOI
+/bin/cp -f /home/stack/ironic-python-agent.initramfs /root/
+mkdir -p ipa/
+pushd ipa
+gunzip -c ../ironic-python-agent.initramfs | cpio -i
+if [ ! -f /home/stack/Compute-kernel_params.txt ]; then
+  touch /home/stack/Compute-kernel_params.txt
+  chown stack /home/stack/Compute-kernel_params.txt
+fi
+/bin/cp -f /home/stack/Compute-kernel_params.txt tmp/kernel_params.txt
+echo "Compute params set: "
+cat tmp/kernel_params.txt
+/bin/cp -f /root/image.py usr/lib/python2.7/site-packages/ironic_python_agent/extensions/image.py
+/bin/cp -f /root/image.pyc usr/lib/python2.7/site-packages/ironic_python_agent/extensions/image.pyc
+find . | cpio -o -H newc | gzip > /home/stack/Compute-ironic-python-agent.initramfs
+chown stack /home/stack/Compute-ironic-python-agent.initramfs
+if [ ! -f /home/stack/Controller-kernel_params.txt ]; then
+  touch /home/stack/Controller-kernel_params.txt
+  chown stack /home/stack/Controller-kernel_params.txt
+fi
+/bin/cp -f /home/stack/Controller-kernel_params.txt tmp/kernel_params.txt
+echo "Controller params set: "
+cat tmp/kernel_params.txt
+find . | cpio -o -H newc | gzip > /home/stack/Controller-ironic-python-agent.initramfs
+chown stack /home/stack/Controller-ironic-python-agent.initramfs
+popd
+/bin/rm -rf ipa/
+EOI
+
+    DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/numa.yaml"
+  fi
+
+  # make sure ceph is installed
+  DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml"
+
+  # scale compute nodes according to inventory
+  total_nodes=$(ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "cat /home/stack/instackenv.json | grep -c memory")
+
+  # check if HA is enabled
+  if [[ "$ha_enabled" == "True" ]]; then
+     DEPLOY_OPTIONS+=" --control-scale 3"
+     compute_nodes=$((total_nodes - 3))
+     DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/puppet-pacemaker.yaml"
+  else
+     compute_nodes=$((total_nodes - 1))
+  fi
+
+  if [ "$compute_nodes" -le 0 ]; then
+    echo -e "${red}ERROR: Invalid number of compute nodes: ${compute_nodes}. Check your inventory file.${reset}"
+    exit 1
+  else
+    echo -e "${blue}INFO: Number of compute nodes set for deployment: ${compute_nodes}${reset}"
+    DEPLOY_OPTIONS+=" --compute-scale ${compute_nodes}"
+  fi
+
+  if [[ "$net_isolation_enabled" == "TRUE" ]]; then
+     #DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/network-isolation.yaml"
+     DEPLOY_OPTIONS+=" -e network-environment.yaml"
+  fi
+
+  if [[ "$ha_enabled" == "True" ]] || [[ "$net_isolation_enabled" == "TRUE" ]]; then
+     DEPLOY_OPTIONS+=" --ntp-server $ntp_server"
+  fi
+
+  if [[ ! "$virtual" == "TRUE" ]]; then
+     DEPLOY_OPTIONS+=" --control-flavor control --compute-flavor compute"
+  else
+     DEPLOY_OPTIONS+=" -e virtual-environment.yaml"
+  fi
+
+  DEPLOY_OPTIONS+=" -e opnfv-environment.yaml"
+
+  echo -e "${blue}INFO: Deploy options set:\n${DEPLOY_OPTIONS}${reset}"
+
+  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+if [ "$debug" == 'TRUE' ]; then
+    LIBGUESTFS_BACKEND=direct virt-customize -a overcloud-full.qcow2 --root-password password:opnfvapex
+fi
+
+source stackrc
+set -o errexit
+echo "Uploading overcloud glance images"
+openstack overcloud image upload
+
+echo "Configuring undercloud and discovering nodes"
+openstack baremetal import --json instackenv.json
+openstack baremetal configure boot
+bash -x set_perf_images.sh ${performance_roles[@]}
+#if [[ -z "$virtual" ]]; then
+#  openstack baremetal introspection bulk start
+#fi
+echo "Configuring flavors"
+for flavor in baremetal control compute; do
+  echo -e "${blue}INFO: Updating flavor: \${flavor}${reset}"
+  if openstack flavor list | grep \${flavor}; then
+    openstack flavor delete \${flavor}
+  fi
+  openstack flavor create --id auto --ram 4096 --disk 39 --vcpus 1 \${flavor}
+  if ! openstack flavor list | grep \${flavor}; then
+    echo -e "${red}ERROR: Unable to create flavor \${flavor}${reset}"
+  fi
+done
+openstack flavor set --property "cpu_arch"="x86_64" --property "capabilities:boot_option"="local" baremetal
+openstack flavor set --property "cpu_arch"="x86_64" --property "capabilities:boot_option"="local" --property "capabilities:profile"="control" control
+openstack flavor set --property "cpu_arch"="x86_64" --property "capabilities:boot_option"="local" --property "capabilities:profile"="compute" compute
+echo "Configuring nameserver on ctlplane network"
+dns_server_ext=''
+for dns_server in ${dns_servers}; do
+  dns_server_ext="\${dns_server_ext} --dns-nameserver \${dns_server}"
+done
+neutron subnet-update \$(neutron subnet-list | grep -Ev "id|tenant|external|storage" | grep -v \\\\-\\\\- | awk {'print \$2'}) \${dns_server_ext}
+echo "Executing overcloud deployment, this should run for an extended period without output."
+sleep 60 #wait for Hypervisor stats to check-in to nova
+# save deploy command so it can be used for debugging
+cat > deploy_command << EOF
+openstack overcloud deploy --templates $DEPLOY_OPTIONS --timeout 90
+EOF
+EOI
+
+  if [ "$interactive" == "TRUE" ]; then
+    if ! prompt_user "Overcloud Deployment"; then
+      echo -e "${blue}INFO: User requests exit${reset}"
+      exit 0
+    fi
+  fi
+
+  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+source stackrc
+openstack overcloud deploy --templates $DEPLOY_OPTIONS --timeout 90
+if ! heat stack-list | grep CREATE_COMPLETE 1>/dev/null; then
+  $(typeset -f debug_stack)
+  debug_stack
+  exit 1
+fi
+EOI
+
+  # Configure DPDK
+  if [ "${deploy_options_array['dataplane']}" == 'ovs_dpdk' ]; then
+    ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI || (echo "DPDK config failed, exiting..."; exit 1)
+source stackrc
+set -o errexit
+for node in \$(nova list | grep novacompute | grep -Eo "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+"); do
+echo "Running DPDK test app on \$node"
+ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
+set -o errexit
+sudo dpdk_helloworld --no-pci
+sudo dpdk_nic_bind -s
+EOF
+done
+EOI
+  fi
+
+  if [ "$debug" == 'TRUE' ]; then
+      ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+source overcloudrc
+echo "Keystone Endpoint List:"
+openstack endpoint list
+echo "Keystone Service List"
+openstack service list
+cinder quota-show \$(openstack project list | grep admin | awk {'print \$2'})
+EOI
+  fi
+}
diff --git a/lib/post-install-functions.sh b/lib/post-install-functions.sh
new file mode 100755 (executable)
index 0000000..f52e28e
--- /dev/null
@@ -0,0 +1,153 @@
+#!/usr/bin/env bash
+##############################################################################
+# Copyright (c) 2015 Tim Rozet (Red Hat), Dan Radez (Red Hat) and others.
+#
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
+
+##Post configuration after install
+##params: none
+function configure_post_install {
+  local opnfv_attach_networks ovs_ip ip_range net_cidr tmp_ip
+  opnfv_attach_networks="admin_network public_network"
+
+  echo -e "${blue}INFO: Post Install Configuration Running...${reset}"
+
+  echo -e "${blue}INFO: Configuring ssh for root to overcloud nodes...${reset}"
+  # copy host key to instack
+  scp ${SSH_OPTIONS[@]} /root/.ssh/id_rsa.pub "stack@$UNDERCLOUD":jumphost_id_rsa.pub
+
+  # add host key to overcloud nodes authorized keys
+  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" << EOI
+source stackrc
+nodes=\$(nova list | grep -Eo "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+")
+for node in \$nodes; do
+cat ~/jumphost_id_rsa.pub | ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" 'cat >> ~/.ssh/authorized_keys'
+done
+EOI
+
+  if [ "${deploy_options_array['dataplane']}" == 'ovs_dpdk' ]; then
+    echo -e "${blue}INFO: Bringing up br-phy and ovs-agent for dpdk compute nodes...${reset}"
+    compute_nodes=$(undercloud_connect stack "source stackrc; nova list | grep compute | wc -l")
+    i=0
+    while [ "$i" -lt "$compute_nodes" ]; do
+      overcloud_connect compute${i} "sudo ifup br-phy; sudo systemctl restart neutron-openvswitch-agent"
+      i=$((i + 1))
+    done
+  fi
+
+  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+source overcloudrc
+set -o errexit
+echo "Configuring Neutron external network"
+neutron net-create external --router:external=True --tenant-id \$(openstack project show service | grep id | awk '{ print \$4 }')
+neutron subnet-create --name external-net --tenant-id \$(openstack project show service | grep id | awk '{ print \$4 }') --disable-dhcp external --gateway ${public_network_gateway} --allocation-pool start=${public_network_floating_ip_range%%,*},end=${public_network_floating_ip_range##*,} ${public_network_cidr}
+
+echo "Removing sahara endpoint and service"
+sahara_service_id=\$(openstack service list | grep sahara | cut -d ' ' -f 2)
+sahara_endpoint_id=\$(openstack endpoint list | grep sahara | cut -d ' ' -f 2)
+openstack endpoint delete \$sahara_endpoint_id
+openstack service delete \$sahara_service_id
+
+echo "Removing swift endpoint and service"
+swift_service_id=\$(openstack service list | grep swift | cut -d ' ' -f 2)
+swift_endpoint_id=\$(openstack endpoint list | grep swift | cut -d ' ' -f 2)
+openstack endpoint delete \$swift_endpoint_id
+openstack service delete \$swift_service_id
+
+if [ "${deploy_options_array['congress']}" == 'True' ]; then
+    for s in nova neutronv2 ceilometer cinder glancev2 keystone; do
+        openstack congress datasource create \$s "\$s" \\
+            --config username=\$OS_USERNAME \\
+            --config tenant_name=\$OS_TENANT_NAME \\
+            --config password=\$OS_PASSWORD \\
+            --config auth_url=\$OS_AUTH_URL
+    done
+fi
+EOI
+
+  echo -e "${blue}INFO: Checking if OVS bridges have IP addresses...${reset}"
+  for network in ${opnfv_attach_networks}; do
+    ovs_ip=$(find_ip ${NET_MAP[$network]})
+    tmp_ip=''
+    if [ -n "$ovs_ip" ]; then
+      echo -e "${blue}INFO: OVS Bridge ${NET_MAP[$network]} has IP address ${ovs_ip}${reset}"
+    else
+      echo -e "${blue}INFO: OVS Bridge ${NET_MAP[$network]} missing IP, will configure${reset}"
+      # use last IP of allocation pool
+      eval "ip_range=\${${network}_usable_ip_range}"
+      ovs_ip=${ip_range##*,}
+      eval "net_cidr=\${${network}_cidr}"
+      sudo ip addr add ${ovs_ip}/${net_cidr##*/} dev ${NET_MAP[$network]}
+      sudo ip link set up ${NET_MAP[$network]}
+      tmp_ip=$(find_ip ${NET_MAP[$network]})
+      if [ -n "$tmp_ip" ]; then
+        echo -e "${blue}INFO: OVS Bridge ${NET_MAP[$network]} IP set: ${tmp_ip}${reset}"
+        continue
+      else
+        echo -e "${red}ERROR: Unable to set OVS Bridge ${NET_MAP[$network]} with IP: ${ovs_ip}${reset}"
+        return 1
+      fi
+    fi
+  done
+
+  # for virtual, we NAT public network through Undercloud
+  if [ "$virtual" == "TRUE" ]; then
+    if ! configure_undercloud_nat ${public_network_cidr}; then
+      echo -e "${red}ERROR: Unable to NAT undercloud with external net: ${public_network_cidr}${reset}"
+      exit 1
+    else
+      echo -e "${blue}INFO: Undercloud VM has been setup to NAT Overcloud public network${reset}"
+    fi
+  fi
+
+  # for sfc deployments we need the vxlan workaround
+  if [ "${deploy_options_array['sfc']}" == 'True' ]; then
+      ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+source stackrc
+set -o errexit
+for node in \$(nova list | grep -Eo "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+"); do
+ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
+sudo ifconfig br-int up
+sudo ip route add 123.123.123.0/24 dev br-int
+EOF
+done
+EOI
+  fi
+
+  # Collect deployment logs
+  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+mkdir -p ~/deploy_logs
+rm -rf deploy_logs/*
+source stackrc
+set -o errexit
+for node in \$(nova list | grep -Eo "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+"); do
+ ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
+ sudo cp /var/log/messages /home/heat-admin/messages.log
+ sudo chown heat-admin /home/heat-admin/messages.log
+EOF
+scp ${SSH_OPTIONS[@]} heat-admin@\$node:/home/heat-admin/messages.log ~/deploy_logs/\$node.messages.log
+if [ "$debug" == "TRUE" ]; then
+    nova list --ip \$node
+    echo "---------------------------"
+    echo "-----/var/log/messages-----"
+    echo "---------------------------"
+    cat ~/deploy_logs/\$node.messages.log
+    echo "---------------------------"
+    echo "----------END LOG----------"
+    echo "---------------------------"
+fi
+ ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
+ sudo rm -f /home/heat-admin/messages.log
+EOF
+done
+
+# Print out the undercloud IP and dashboard URL
+source stackrc
+echo "Undercloud IP: $UNDERCLOUD, please connect by doing 'opnfv-util undercloud'"
+echo "Overcloud dashboard available at http://\$(heat output-show overcloud PublicVip | sed 's/"//g')/dashboard"
+EOI
+}
diff --git a/lib/undercloud-functions.sh b/lib/undercloud-functions.sh
new file mode 100755 (executable)
index 0000000..f829e98
--- /dev/null
@@ -0,0 +1,277 @@
+#!/usr/bin/env bash
+##############################################################################
+# Copyright (c) 2015 Tim Rozet (Red Hat), Dan Radez (Red Hat) and others.
+#
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
+
+##verify vm exists, an has a dhcp lease assigned to it
+##params: none
+function setup_undercloud_vm {
+  if ! virsh list --all | grep undercloud > /dev/null; then
+      undercloud_nets="default admin_network"
+      if [[ $enabled_network_list =~ "public_network" ]]; then
+        undercloud_nets+=" public_network"
+      fi
+      define_vm undercloud hd 30 "$undercloud_nets" 4 12288
+
+      ### this doesn't work for some reason I was getting hangup events so using cp instead
+      #virsh vol-upload --pool default --vol undercloud.qcow2 --file $CONFIG/stack/undercloud.qcow2
+      #2015-12-05 12:57:20.569+0000: 8755: info : libvirt version: 1.2.8, package: 16.el7_1.5 (CentOS BuildSystem <http://bugs.centos.org>, 2015-11-03-13:56:46, worker1.bsys.centos.org)
+      #2015-12-05 12:57:20.569+0000: 8755: warning : virKeepAliveTimerInternal:143 : No response from client 0x7ff1e231e630 after 6 keepalive messages in 35 seconds
+      #2015-12-05 12:57:20.569+0000: 8756: warning : virKeepAliveTimerInternal:143 : No response from client 0x7ff1e231e630 after 6 keepalive messages in 35 seconds
+      #error: cannot close volume undercloud.qcow2
+      #error: internal error: received hangup / error event on socket
+      #error: Reconnected to the hypervisor
+
+      local undercloud_dst=/var/lib/libvirt/images/undercloud.qcow2
+      cp -f $RESOURCES/undercloud.qcow2 $undercloud_dst
+
+      # resize Undercloud machine
+      echo "Checking if Undercloud needs to be resized..."
+      undercloud_size=$(LIBGUESTFS_BACKEND=direct virt-filesystems --long -h --all -a $undercloud_dst |grep device | grep -Eo "[0-9\.]+G" | sed -n 's/\([0-9][0-9]*\).*/\1/p')
+      if [ "$undercloud_size" -lt 30 ]; then
+        qemu-img resize /var/lib/libvirt/images/undercloud.qcow2 +25G
+        LIBGUESTFS_BACKEND=direct virt-resize --expand /dev/sda1 $RESOURCES/undercloud.qcow2 $undercloud_dst
+        LIBGUESTFS_BACKEND=direct virt-customize -a $undercloud_dst --run-command 'xfs_growfs -d /dev/sda1 || true'
+        new_size=$(LIBGUESTFS_BACKEND=direct virt-filesystems --long -h --all -a $undercloud_dst |grep filesystem | grep -Eo "[0-9\.]+G" | sed -n 's/\([0-9][0-9]*\).*/\1/p')
+        if [ "$new_size" -lt 30 ]; then
+          echo "Error resizing Undercloud machine, disk size is ${new_size}"
+          exit 1
+        else
+          echo "Undercloud successfully resized"
+        fi
+      else
+        echo "Skipped Undercloud resize, upstream is large enough"
+      fi
+
+  else
+      echo "Found existing Undercloud VM, exiting."
+      exit 1
+  fi
+
+  # if the VM is not running update the authkeys and start it
+  if ! virsh list | grep undercloud > /dev/null; then
+    echo "Injecting ssh key to Undercloud VM"
+    LIBGUESTFS_BACKEND=direct virt-customize -a $undercloud_dst --run-command "mkdir -p /root/.ssh/" \
+        --upload ~/.ssh/id_rsa.pub:/root/.ssh/authorized_keys \
+        --run-command "chmod 600 /root/.ssh/authorized_keys && restorecon /root/.ssh/authorized_keys" \
+        --run-command "cp /root/.ssh/authorized_keys /home/stack/.ssh/" \
+        --run-command "chown stack:stack /home/stack/.ssh/authorized_keys && chmod 600 /home/stack/.ssh/authorized_keys"
+    virsh start undercloud
+  fi
+
+  sleep 10 # let undercloud get started up
+
+  # get the undercloud VM IP
+  CNT=10
+  echo -n "${blue}Waiting for Undercloud's dhcp address${reset}"
+  undercloud_mac=$(virsh domiflist undercloud | grep default | awk '{ print $5 }')
+  while ! $(arp -e | grep ${undercloud_mac} > /dev/null) && [ $CNT -gt 0 ]; do
+      echo -n "."
+      sleep 10
+      CNT=$((CNT-1))
+  done
+  UNDERCLOUD=$(arp -e | grep ${undercloud_mac} | awk {'print $1'})
+
+  if [ -z "$UNDERCLOUD" ]; then
+    echo "\n\nCan't get IP for Undercloud. Can Not Continue."
+    exit 1
+  else
+     echo -e "${blue}\rUndercloud VM has IP $UNDERCLOUD${reset}"
+  fi
+
+  CNT=10
+  echo -en "${blue}\rValidating Undercloud VM connectivity${reset}"
+  while ! ping -c 1 $UNDERCLOUD > /dev/null && [ $CNT -gt 0 ]; do
+      echo -n "."
+      sleep 3
+      CNT=$((CNT-1))
+  done
+  if [ "$CNT" -eq 0 ]; then
+      echo "Failed to contact Undercloud. Can Not Continue"
+      exit 1
+  fi
+  CNT=10
+  while ! ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "echo ''" 2>&1> /dev/null && [ $CNT -gt 0 ]; do
+      echo -n "."
+      sleep 3
+      CNT=$((CNT-1))
+  done
+  if [ "$CNT" -eq 0 ]; then
+      echo "Failed to connect to Undercloud. Can Not Continue"
+      exit 1
+  fi
+
+  # extra space to overwrite the previous connectivity output
+  echo -e "${blue}\r                                                                 ${reset}"
+  sleep 1
+
+  # ssh key fix for stack user
+  ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "restorecon -r /home/stack"
+}
+
+##Copy over the glance images and instackenv json file
+##params: none
+function configure_undercloud {
+  local controller_nic_template compute_nic_template
+  echo
+  echo "Copying configuration files to Undercloud"
+  if [[ "$net_isolation_enabled" == "TRUE" ]]; then
+    echo -e "${blue}Network Environment set for Deployment: ${reset}"
+    cat /tmp/network-environment.yaml
+    scp ${SSH_OPTIONS[@]} /tmp/network-environment.yaml "stack@$UNDERCLOUD":
+
+    # check for ODL L3/ONOS
+    if [ "${deploy_options_array['sdn_l3']}" == 'True' ]; then
+      ext_net_type=br-ex
+    fi
+
+    if [ "${deploy_options_array['dataplane']}" == 'ovs_dpdk' ]; then
+      ovs_dpdk_bridge='br-phy'
+    else
+      ovs_dpdk_bridge=''
+    fi
+
+    if ! controller_nic_template=$(python3.4 -B $LIB/python/apex_python_utils.py nic-template -r controller -s $NETSETS -i $net_isolation_enabled -t $CONFIG/nics-template.yaml.jinja2 -n "$enabled_network_list" -e "br-ex" -af $ip_addr_family); then
+      echo -e "${red}ERROR: Failed to generate controller NIC heat template ${reset}"
+      exit 1
+    fi
+
+    if ! compute_nic_template=$(python3.4 -B $LIB/python/apex_python_utils.py nic-template -r compute -s $NETSETS -i $net_isolation_enabled -t $CONFIG/nics-template.yaml.jinja2 -n "$enabled_network_list" -e $ext_net_type -af $ip_addr_family -d "$ovs_dpdk_bridge"); then
+      echo -e "${red}ERROR: Failed to generate compute NIC heat template ${reset}"
+      exit 1
+    fi
+    ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" << EOI
+mkdir nics/
+cat > nics/controller.yaml << EOF
+$controller_nic_template
+EOF
+cat > nics/compute.yaml << EOF
+$compute_nic_template
+EOF
+EOI
+  fi
+
+  # ensure stack user on Undercloud machine has an ssh key
+  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "if [ ! -e ~/.ssh/id_rsa.pub ]; then ssh-keygen -t rsa -N '' -f ~/.ssh/id_rsa; fi"
+
+  if [ "$virtual" == "TRUE" ]; then
+
+      # copy the Undercloud VM's stack user's pub key to
+      # root's auth keys so that Undercloud can control
+      # vm power on the hypervisor
+      ssh ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" "cat /home/stack/.ssh/id_rsa.pub" >> /root/.ssh/authorized_keys
+
+      DEPLOY_OPTIONS+=" --libvirt-type qemu"
+      INSTACKENV=$CONFIG/instackenv-virt.json
+
+      # upload instackenv file to Undercloud for virtual deployment
+      scp ${SSH_OPTIONS[@]} $INSTACKENV "stack@$UNDERCLOUD":instackenv.json
+  fi
+
+  # allow stack to control power management on the hypervisor via sshkey
+  # only if this is a virtual deployment
+  if [ "$virtual" == "TRUE" ]; then
+      ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+while read -r line; do
+  stack_key=\${stack_key}\\\\\\\\n\${line}
+done < <(cat ~/.ssh/id_rsa)
+stack_key=\$(echo \$stack_key | sed 's/\\\\\\\\n//')
+sed -i 's~INSERT_STACK_USER_PRIV_KEY~'"\$stack_key"'~' instackenv.json
+EOI
+  fi
+
+  # copy stack's ssh key to this users authorized keys
+  ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "cat /home/stack/.ssh/id_rsa.pub" >> ~/.ssh/authorized_keys
+
+  # disable requiretty for sudo
+  ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" "sed -i 's/Defaults\s*requiretty//'" /etc/sudoers
+
+  # configure undercloud on Undercloud VM
+  echo "Running undercloud configuration."
+  echo "Logging undercloud configuration to undercloud:/home/stack/apex-undercloud-install.log"
+  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" << EOI
+if [[ "$net_isolation_enabled" == "TRUE" ]]; then
+  sed -i 's/#local_ip/local_ip/' undercloud.conf
+  sed -i 's/#network_gateway/network_gateway/' undercloud.conf
+  sed -i 's/#network_cidr/network_cidr/' undercloud.conf
+  sed -i 's/#dhcp_start/dhcp_start/' undercloud.conf
+  sed -i 's/#dhcp_end/dhcp_end/' undercloud.conf
+  sed -i 's/#inspection_iprange/inspection_iprange/' undercloud.conf
+  sed -i 's/#undercloud_debug/undercloud_debug/' undercloud.conf
+
+  openstack-config --set undercloud.conf DEFAULT local_ip ${admin_network_provisioner_ip}/${admin_network_cidr##*/}
+  openstack-config --set undercloud.conf DEFAULT network_gateway ${admin_network_provisioner_ip}
+  openstack-config --set undercloud.conf DEFAULT network_cidr ${admin_network_cidr}
+  openstack-config --set undercloud.conf DEFAULT dhcp_start ${admin_network_dhcp_range%%,*}
+  openstack-config --set undercloud.conf DEFAULT dhcp_end ${admin_network_dhcp_range##*,}
+  openstack-config --set undercloud.conf DEFAULT inspection_iprange ${admin_network_introspection_range}
+  openstack-config --set undercloud.conf DEFAULT undercloud_debug false
+
+fi
+
+sudo sed -i '/CephClusterFSID:/c\\  CephClusterFSID: \\x27$(cat /proc/sys/kernel/random/uuid)\\x27' /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml
+sudo sed -i '/CephMonKey:/c\\  CephMonKey: \\x27'"\$(ceph-authtool --gen-print-key)"'\\x27' /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml
+sudo sed -i '/CephAdminKey:/c\\  CephAdminKey: \\x27'"\$(ceph-authtool --gen-print-key)"'\\x27' /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml
+
+# we assume that packages will not need to be updated with undercloud install
+# and that it will be used only to configure the undercloud
+# packages updates would need to be handled manually with yum update
+sudo cp -f /usr/share/diskimage-builder/elements/yum/bin/install-packages /usr/share/diskimage-builder/elements/yum/bin/install-packages.bak
+cat << 'EOF' | sudo tee /usr/share/diskimage-builder/elements/yum/bin/install-packages > /dev/null
+#!/bin/sh
+exit 0
+EOF
+
+openstack undercloud install &> apex-undercloud-install.log || {
+    # cat the undercloud install log incase it fails
+    echo "ERROR: openstack undercloud install has failed. Dumping Log:"
+    cat apex-undercloud-install.log
+    exit 1
+}
+
+sleep 30
+sudo systemctl restart openstack-glance-api
+sudo systemctl restart openstack-nova-conductor
+sudo systemctl restart openstack-nova-compute
+
+sudo sed -i '/num_engine_workers/c\num_engine_workers = 2' /etc/heat/heat.conf
+sudo sed -i '/#workers\s=/c\workers = 2' /etc/heat/heat.conf
+sudo systemctl restart openstack-heat-engine
+sudo systemctl restart openstack-heat-api
+EOI
+
+# configure external network
+  ssh -T ${SSH_OPTIONS[@]} "root@$UNDERCLOUD" << EOI
+if [[ "$public_network_vlan" != "native" ]]; then
+  cat <<EOF > /etc/sysconfig/network-scripts/ifcfg-vlan${public_network_vlan}
+DEVICE=vlan${public_network_vlan}
+ONBOOT=yes
+DEVICETYPE=ovs
+TYPE=OVSIntPort
+BOOTPROTO=static
+IPADDR=${public_network_provisioner_ip}
+PREFIX=${public_network_cidr##*/}
+OVS_BRIDGE=br-ctlplane
+OVS_OPTIONS="tag=${public_network_vlan}"
+EOF
+  ifup vlan${public_network_vlan}
+else
+  if ! ip a s eth2 | grep ${public_network_provisioner_ip} > /dev/null; then
+      ip a a ${public_network_provisioner_ip}/${public_network_cidr##*/} dev eth2
+      ip link set up dev eth2
+  fi
+fi
+EOI
+
+# WORKAROUND: must restart the above services to fix sync problem with nova compute manager
+# TODO: revisit and file a bug if necessary. This should eventually be removed
+# as well as glance api problem
+echo -e "${blue}INFO: Sleeping 15 seconds while services come back from restart${reset}"
+sleep 15
+
+}
diff --git a/lib/virtual-setup-functions.sh b/lib/virtual-setup-functions.sh
new file mode 100755 (executable)
index 0000000..e7410da
--- /dev/null
@@ -0,0 +1,151 @@
+#!/usr/bin/env bash
+##############################################################################
+# Copyright (c) 2015 Tim Rozet (Red Hat), Dan Radez (Red Hat) and others.
+#
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
+
+##Create virtual nodes in virsh
+##params: vcpus, ramsize
+function setup_virtual_baremetal {
+  local vcpus ramsize
+  if [ -z "$1" ]; then
+    vcpus=4
+    ramsize=8192
+  elif [ -z "$2" ]; then
+    vcpus=$1
+    ramsize=8192
+  else
+    vcpus=$1
+    ramsize=$(($2*1024))
+  fi
+  #start by generating the opening json for instackenv.json
+  cat > $CONFIG/instackenv-virt.json << EOF
+{
+  "nodes": [
+EOF
+
+  # next create the virtual machines and add their definitions to the file
+  if [ "$ha_enabled" == "False" ]; then
+      # 1 controller + computes
+      # zero based so just pass compute count
+      vm_index=$VM_COMPUTES
+  else
+      # 3 controller + computes
+      # zero based so add 2 to compute count
+      vm_index=$((2+$VM_COMPUTES))
+  fi
+
+  for i in $(seq 0 $vm_index); do
+    if ! virsh list --all | grep baremetal${i} > /dev/null; then
+      define_vm baremetal${i} network 41 'admin_network' $vcpus $ramsize
+      for n in private_network public_network storage_network api_network; do
+        if [[ $enabled_network_list =~ $n ]]; then
+          echo -n "$n "
+          virsh attach-interface --domain baremetal${i} --type network --source $n --model virtio --config
+        fi
+      done
+    else
+      echo "Found Baremetal ${i} VM, using existing VM"
+    fi
+    #virsh vol-list default | grep baremetal${i} 2>&1> /dev/null || virsh vol-create-as default baremetal${i}.qcow2 41G --format qcow2
+    mac=$(virsh domiflist baremetal${i} | grep admin_network | awk '{ print $5 }')
+
+    if [ "$VM_COMPUTES" -gt 0 ]; then
+      capability="profile:compute"
+      VM_COMPUTES=$((VM_COMPUTES - 1))
+    else
+      capability="profile:control"
+    fi
+
+    cat >> $CONFIG/instackenv-virt.json << EOF
+    {
+      "pm_addr": "192.168.122.1",
+      "pm_user": "root",
+      "pm_password": "INSERT_STACK_USER_PRIV_KEY",
+      "pm_type": "pxe_ssh",
+      "mac": [
+        "$mac"
+      ],
+      "cpu": "$vcpus",
+      "memory": "$ramsize",
+      "disk": "41",
+      "arch": "x86_64",
+      "capabilities": "$capability"
+    },
+EOF
+  done
+
+  #truncate the last line to remove the comma behind the bracket
+  tail -n 1 $CONFIG/instackenv-virt.json | wc -c | xargs -I {} truncate $CONFIG/instackenv-virt.json -s -{}
+
+  #finally reclose the bracket and close the instackenv.json file
+  cat >> $CONFIG/instackenv-virt.json << EOF
+    }
+  ],
+  "arch": "x86_64",
+  "host-ip": "192.168.122.1",
+  "power_manager": "nova.virt.baremetal.virtual_power_driver.VirtualPowerManager",
+  "seed-ip": "",
+  "ssh-key": "INSERT_STACK_USER_PRIV_KEY",
+  "ssh-user": "root"
+}
+EOF
+  #Overwrite the tripleo-inclubator domain.xml with our own, keeping a backup.
+  if [ ! -f /usr/share/tripleo/templates/domain.xml.bak ]; then
+    /usr/bin/mv -f /usr/share/tripleo/templates/domain.xml /usr/share/tripleo/templates/domain.xml.bak
+  fi
+
+  /usr/bin/cp -f $LIB/installer/domain.xml /usr/share/tripleo/templates/domain.xml
+}
+
+##Create virtual nodes in virsh
+##params: name - String: libvirt name for VM
+##        bootdev - String: boot device for the VM
+##        disksize - Number: size of the disk in GB
+##        ovs_bridges: - List: list of ovs bridges
+##        vcpus - Number of VCPUs to use (defaults to 4)
+##        ramsize - Size of RAM for VM in MB (defaults to 8192)
+function define_vm () {
+  local vcpus ramsize
+
+  if [ -z "$5" ]; then
+    vcpus=4
+    ramsize=8388608
+  elif [ -z "$6" ]; then
+    vcpus=$5
+    ramsize=8388608
+  else
+    vcpus=$5
+    ramsize=$(($6*1024))
+  fi
+
+  # Create the libvirt storage volume
+  if virsh vol-list default | grep ${1}.qcow2 2>&1> /dev/null; then
+    volume_path=$(virsh vol-path --pool default ${1}.qcow2 || echo "/var/lib/libvirt/images/${1}.qcow2")
+    echo "Volume ${1} exists. Deleting Existing Volume $volume_path"
+    virsh vol-dumpxml ${1}.qcow2 --pool default > /dev/null || echo '' #ok for this to fail
+    touch $volume_path
+    virsh vol-delete ${1}.qcow2 --pool default
+  fi
+  virsh vol-create-as default ${1}.qcow2 ${3}G --format qcow2
+  volume_path=$(virsh vol-path --pool default ${1}.qcow2)
+  if [ ! -f $volume_path ]; then
+      echo "$volume_path Not created successfully... Aborting"
+      exit 1
+  fi
+
+  # create the VM
+  /usr/libexec/openstack-tripleo/configure-vm --name $1 \
+                                              --bootdev $2 \
+                                              --image "$volume_path" \
+                                              --diskbus sata \
+                                              --arch x86_64 \
+                                              --cpus $vcpus \
+                                              --memory $ramsize \
+                                              --libvirt-nic-driver virtio \
+                                              --baremetal-interface $4
+}