Merge "Adding error checking"
[apex.git] / ci / deploy.sh
index bcd28f8..31b5d2f 100755 (executable)
@@ -30,7 +30,6 @@ else
 fi
 
 vm_index=4
-#ha_enabled="TRUE"
 interactive="FALSE"
 ping_site="8.8.8.8"
 ntp_server="pool.ntp.org"
@@ -195,7 +194,7 @@ for Auto-detection${reset}"
     done
     echo -e "${blue}INFO: Detecting Network Specific settings for: ${enabled_network}${reset}"
     # detect network specific settings
-    if [ -n $(eval echo \${${network}_optional_settings}) ]; then
+    if [ -n $(eval echo \${${enabled_network}_optional_settings}) ]; then
       eval "network_specific_settings=\${${enabled_network}_optional_settings}"
       for setting in ${network_specific_settings}; do
         eval "setting_value=\${${enabled_network}_${setting}}"
@@ -304,7 +303,7 @@ parse_inventory_file() {
     exit 1
   fi
 
-  eval $(parse_yaml $INVENTORY_FILE)
+  eval $(parse_yaml $INVENTORY_FILE) || echo "${red}Failed to parse inventory.yaml. Aborting.${reset}" && exit 1
 
   instack_env_output="
 {
@@ -317,7 +316,7 @@ parse_inventory_file() {
     node_output="
         {
           \"pm_password\": \"$(eval echo \${${node}ipmi_pass})\",
-          \"pm_type\": \"pxe_ipmitool\",
+          \"pm_type\": \"$(eval echo \${${node}pm_type})\",
           \"mac\": [
             \"$(eval echo \${${node}mac_address})\"
           ],
@@ -389,7 +388,8 @@ function configure_deps {
   fi
 
   # ensure networks are configured
-  systemctl start openvswitch
+  systemctl status libvirtd || systemctl start libvirtd
+  systemctl status openvswitch || systemctl start openvswitch
 
   # If flat we only use admin network
   if [[ "$net_isolation_enabled" == "FALSE" ]]; then
@@ -401,10 +401,15 @@ function configure_deps {
     virsh_enabled_networks=$enabled_network_list
   fi
 
+  virsh net-list | grep default || virsh net-define /usr/share/libvirt/networks/default.xml
+  virsh net-list | grep -E "default\s+active" > /dev/null || virsh net-start default
+  virsh net-list | grep -E "default\s+active\s+yes" > /dev/null || virsh net-autostart --network default
+
   for network in ${OPNFV_NETWORK_TYPES}; do
     ovs-vsctl list-br | grep ${NET_MAP[$network]} > /dev/null || ovs-vsctl add-br ${NET_MAP[$network]}
     virsh net-list --all | grep ${NET_MAP[$network]} > /dev/null || virsh net-define $CONFIG/${NET_MAP[$network]}-net.xml
     virsh net-list | grep -E "${NET_MAP[$network]}\s+active" > /dev/null || virsh net-start ${NET_MAP[$network]}
+    virsh net-list | grep -E "${NET_MAP[$network]}\s+active\s+yes" > /dev/null || virsh net-autostart --network ${NET_MAP[$network]}
   done
 
   echo -e "${blue}INFO: Bridges set: ${reset}"
@@ -492,7 +497,7 @@ function setup_instack_vm {
       #error: internal error: received hangup / error event on socket
       #error: Reconnected to the hypervisor
 
-      instack_dst=/var/lib/libvirt/images/instack.qcow2
+      local instack_dst=/var/lib/libvirt/images/instack.qcow2
       cp -f $RESOURCES/instack.qcow2 $instack_dst
 
       # resize instack machine
@@ -500,8 +505,8 @@ function setup_instack_vm {
       instack_size=$(LIBGUESTFS_BACKEND=direct virt-filesystems --long -h --all -a $instack_dst |grep device | grep -Eo "[0-9\.]+G" | sed -n 's/\([0-9][0-9]*\).*/\1/p')
       if [ "$instack_size" -lt 30 ]; then
         qemu-img resize /var/lib/libvirt/images/instack.qcow2 +25G
-       LIBGUESTFS_BACKEND=direct virt-resize --expand /dev/sda1 $RESOURCES/instack.qcow2 $instack_dst
-       LIBGUESTFS_BACKEND=direct virt-customize -a $instack_dst --run-command 'xfs_growfs -d /dev/sda1 || true'
+        LIBGUESTFS_BACKEND=direct virt-resize --expand /dev/sda1 $RESOURCES/instack.qcow2 $instack_dst
+        LIBGUESTFS_BACKEND=direct virt-customize -a $instack_dst --run-command 'xfs_growfs -d /dev/sda1 || true'
         new_size=$(LIBGUESTFS_BACKEND=direct virt-filesystems --long -h --all -a $instack_dst |grep filesystem | grep -Eo "[0-9\.]+G" | sed -n 's/\([0-9][0-9]*\).*/\1/p')
         if [ "$new_size" -lt 30 ]; then
           echo "Error resizing instack machine, disk size is ${new_size}"
@@ -520,7 +525,7 @@ function setup_instack_vm {
   # if the VM is not running update the authkeys and start it
   if ! virsh list | grep instack > /dev/null; then
     echo "Injecting ssh key to instack VM"
-    virt-customize -c qemu:///system -d instack --run-command "mkdir -p /root/.ssh/" \
+    LIBGUESTFS_BACKEND=direct virt-customize -a $instack_dst --run-command "mkdir -p /root/.ssh/" \
         --upload ~/.ssh/id_rsa.pub:/root/.ssh/authorized_keys \
         --run-command "chmod 600 /root/.ssh/authorized_keys && restorecon /root/.ssh/authorized_keys" \
         --run-command "cp /root/.ssh/authorized_keys /home/stack/.ssh/" \
@@ -762,6 +767,15 @@ sudo sed -i '/CephClusterFSID:/c\\  CephClusterFSID: \\x27$(cat /proc/sys/kernel
 sudo sed -i '/CephMonKey:/c\\  CephMonKey: \\x27'"\$(ceph-authtool --gen-print-key)"'\\x27' /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml
 sudo sed -i '/CephAdminKey:/c\\  CephAdminKey: \\x27'"\$(ceph-authtool --gen-print-key)"'\\x27' /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml
 
+# we assume that packages will not need to be updated with undercloud install
+# and that it will be used only to configure the undercloud
+# packages updates would need to be handled manually with yum update
+sudo cp -f /usr/share/diskimage-builder/elements/yum/bin/install-packages /usr/share/diskimage-builder/elements/yum/bin/install-packages.bak
+cat << 'EOF' | sudo tee /usr/share/diskimage-builder/elements/yum/bin/install-packages > /dev/null
+#!/bin/sh
+exit 0
+EOF
+
 openstack undercloud install &> apex-undercloud-install.log
 sleep 30
 sudo systemctl restart openstack-glance-api
@@ -790,6 +804,11 @@ function undercloud_prep_overcloud_deploy {
     SDN_IMAGE=opendaylight
     if [ "${deploy_options_array['sfc']}" == 'true' ]; then
       SDN_IMAGE+=-sfc
+      if [ ! -f $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 ]; then
+          echo "${red} $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 is required to execute an SFC deployment."
+          echo "Please install the opnfv-apex-opendaylight-sfc package to provide this overcloud image for deployment.${reset}"
+          exit 1
+      fi
     fi
   elif [ "${deploy_options_array['sdn_controller']}" == 'opendaylight-external' ]; then
     DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/opendaylight-external.yaml"
@@ -809,6 +828,14 @@ function undercloud_prep_overcloud_deploy {
     exit 1
   fi
 
+  # Make sure the correct overcloud image is available
+  if [ ! -f $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 ]; then
+      echo "${red} $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 is required to execute your deployment."
+      echo "Both ONOS and OpenDaylight are currently deployed from this image."
+      echo "Please install the opnfv-apex package to provide this overcloud image for deployment.${reset}"
+      exit 1
+  fi
+
   echo "Copying overcloud image to instack"
   scp ${SSH_OPTIONS[@]} $RESOURCES/overcloud-full-${SDN_IMAGE}.qcow2 "stack@$UNDERCLOUD":overcloud-full.qcow2
 
@@ -824,7 +851,7 @@ function undercloud_prep_overcloud_deploy {
      compute_nodes=$((total_nodes - 3))
      DEPLOY_OPTIONS+=" -e /usr/share/openstack-tripleo-heat-templates/environments/puppet-pacemaker.yaml"
   else
-     compute_nodes=$((total_nodes - 1))
+     compute_nodes=1
   fi
 
   if [ "$compute_nodes" -le 0 ]; then
@@ -851,6 +878,9 @@ function undercloud_prep_overcloud_deploy {
   echo -e "${blue}INFO: Deploy options set:\n${DEPLOY_OPTIONS}${reset}"
 
   ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+if [ "$debug" == 'TRUE' ]; then
+    LIBGUESTFS_BACKEND=direct virt-customize -a overcloud-full.qcow2 --root-password password:opnfvapex
+fi
 source stackrc
 set -o errexit
 echo "Uploading overcloud glance images"
@@ -921,9 +951,10 @@ function configure_post_install {
   ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
 source overcloudrc
 set -o errexit
+service_tenant_id="\$(keystone tenant-get service | grep id | awk '{ print \$4 }')"
 echo "Configuring Neutron external network"
-neutron net-create external --router:external=True
-neutron subnet-create --name external-net --disable-dhcp external --gateway ${public_network_gateway} --allocation-pool start=${public_network_floating_ip_range%%,*},end=${public_network_floating_ip_range##*,} ${public_network_cidr}
+neutron net-create external --router:external=True --tenant-id \$service_tenant_id
+neutron subnet-create --name external-net --tenant-id \$service_tenant_id --disable-dhcp external --gateway ${public_network_gateway} --allocation-pool start=${public_network_floating_ip_range%%,*},end=${public_network_floating_ip_range##*,} ${public_network_cidr}
 EOI
 
   echo -e "${blue}INFO: Checking if OVS bridges have IP addresses...${reset}"
@@ -974,6 +1005,40 @@ EOF
 done
 EOI
   fi
+
+  # Collect deployment logs
+  ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+mkdir -p ~/deploy_logs
+rm -rf deploy_logs/*
+source stackrc
+set -o errexit
+for node in \$(nova list | grep -Eo "[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+"); do
+ ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
+ sudo cp /var/log/messages /home/heat-admin/messages.log
+ sudo chown heat-admin /home/heat-admin/messages.log
+EOF
+scp ${SSH_OPTIONS[@]} heat-admin@\$node:/home/heat-admin/messages.log ~/deploy_logs/\$node.messages.log
+if [ "$debug" == "TRUE" ]; then
+    nova list --ip \$node
+    echo "---------------------------"
+    echo "-----/var/log/messages-----"
+    echo "---------------------------"
+    cat ~/deploy_logs/\$node.messages.log
+    echo "---------------------------"
+    echo "----------END LOG----------"
+    echo "---------------------------"
+fi
+ ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
+ sudo rm -f /home/heat-admin/messages.log
+EOF
+done
+
+# Print out the dashboard URL
+source stackrc
+publicvip=\$(heat output-show overcloud PublicVip | sed 's/"//g')
+echo "Overcloud dashboard available at http://\$publicvip/dashboard"
+EOI
+
 }
 
 display_usage() {