-#!/bin/bash
-# shellcheck disable=SC2034,SC2154,SC1091
-set -ex
+#!/bin/bash -e
+# shellcheck disable=SC2034,SC2154,SC1090,SC1091
##############################################################################
# Copyright (c) 2017 Ericsson AB, Mirantis Inc., Enea AB and others.
# jonas.bjurel@ericsson.com
# BEGIN of Exit handlers
#
do_exit () {
- clean
- echo "Exiting ..."
+ local RC=$?
+ cleanup_mounts > /dev/null 2>&1
+ if [ ${RC} -eq 0 ]; then
+ notify_n "[OK] MCP: Openstack installation finished succesfully!" 2
+ else
+ notify_n "[ERROR] MCP: Openstack installation threw a fatal error!"
+ fi
}
#
# End of Exit handlers
$(notify "$(basename "$0"): Deploy the Fuel@OPNFV MCP stack" 3)
$(notify "USAGE:" 2)
- $(basename "$0") -b base-uri -l lab-name -p pod-name -s deploy-scenario \\
- [-B PXE Bridge [-B Mgmt Bridge [-B Internal Bridge [-B Public Bridge]]]] \\
- [-S storage-dir] [-L /path/to/log/file.tar.gz]
+ $(basename "$0") -l lab-name -p pod-name -s deploy-scenario \\
+ [-b Lab Config Base URI] \\
+ [-S storage-dir] [-L /path/to/log/file.tar.gz] \\
+ [-f[f]] [-F] [-e | -E[E]] [-d] [-D]
$(notify "OPTIONS:" 2)
-b Base-uri for the stack-configuration structure
- -B Bridge(s): 1st usage = PXE, 2nd = Mgmt, 3rd = Internal, 4th = Public
+ -d Dry-run
+ -D Debug logging
+ -e Do not launch environment deployment
+ -E Remove existing VCP VMs (use twice to redeploy baremetal nodes)
+ -f Deploy on existing Salt master (use twice to also skip config sync)
+ -F Do only create a Salt master
-h Print this message and exit
-l Lab-name
-p Pod-name
+ -P Skip installation of package dependencies
-s Deploy-scenario short-name
-S Storage dir for VM images
-L Deployment log path and file name
-$(notify "DISABLED OPTIONS (not yet supported with MCP):" 3)
- -d (disabled) Dry-run
- -e (disabled) Do not launch environment deployment
- -f (disabled) Deploy on existing Salt master
- -F (disabled) Do only create a Salt master
- -i (disabled) iso url
- -T (disabled) Timeout, in minutes, for the deploy.
-
-$(notify "Description:" 2)
+$(notify_i "Description:" 2)
Deploys the Fuel@OPNFV stack on the indicated lab resource.
This script provides the Fuel@OPNFV deployment abstraction.
It depends on the OPNFV official configuration directory/file structure
and provides a fairly simple mechanism to execute a deployment.
-$(notify "Input parameters to the build script are:" 2)
--b Base URI to the configuration directory (needs to be provided in a URI
- style, it can be a local resource: file:// or a remote resource http(s)://)
--B Bridges to be used by deploy script. It can be specified several times,
- or as a comma separated list of bridges, or both: -B br1 -B br2,br3
- First occurence sets PXE Brige, next Mgmt, then Internal and Public.
- For an empty value, the deploy script will use virsh to create the default
- expected network (e.g. -B pxe,,,public will use existing "pxe" and "public"
- bridges, respectively create "mgmt" and "internal").
- Note that a virtual network "mcpcontrol" is always created. For virtual
- deploys, "mcpcontrol" is also used for PXE, leaving the PXE bridge unused.
- For baremetal deploys, PXE bridge is used for baremetal node provisioning,
- while "mcpcontrol" is used to provision the infrastructure VMs only.
- The default is 'pxebr'.
+$(notify_i "Input parameters to the build script are:" 2)
+-b Base URI to the configuration directory (needs to be provided in URI style,
+ it can be a local resource: file:// or a remote resource http(s)://).
+ A POD Descriptor File (PDF) and its Installer Descriptor File (IDF)
+ companion should be available at:
+ <base-uri>/labs/<lab-name>/<pod-name>.yaml
+ <base-uri>/labs/<lab-name>/idf-<pod-name>.yaml
+ The default is using the git submodule tracking 'OPNFV Pharos' in
+ <./mcp/scripts/pharos>.
+ An example config is provided inside current repo in
+ <./mcp/config>, automatically linked as <./mcp/scripts/pharos/labs/local>.
+-d Dry-run - Produce deploy config files, but do not execute deploy
+-D Debug logging - Enable extra logging in sh deploy scripts (set -x)
+-e Do not launch environment deployment
+-E Remove existing VCP VMs. It will destroy and undefine all VCP VMs
+ currently defined on cluster KVM nodes. If specified twice (e.g. -E -E),
+ baremetal nodes (VCP too, implicitly) will be removed, then reprovisioned.
+ Only applicable for baremetal deploys.
+-f Deploy on existing Salt master. It will skip infrastructure VM creation,
+ but it will still sync reclass configuration from current repo to Salt
+ Master node. If specified twice (e.g. -f -f), config sync will also be
+ skipped.
+-F Do only create a Salt master
-h Print this message and exit
-L Deployment log path and name, eg. -L /home/jenkins/job.log.tar.gz
-l Lab name as defined in the configuration directory, e.g. lf
--p POD name as defined in the configuration directory, e.g. pod-1
+-p POD name as defined in the configuration directory, e.g. pod2
+-P Skip installing dependency distro packages on current host
+ This flag should only be used if you have kept back older packages that
+ would be upgraded and that is undesirable on the current system.
+ Note that without the required packages, deploy will fail.
-s Deployment-scenario, this points to a short deployment scenario name, which
has to be defined in config directory (e.g. os-odl-nofeature-ha).
-S Storage dir for VM images, default is mcp/deploy/images
-$(notify "Disabled input parameters (not yet supported with MCP):" 3)
--d (disabled) Dry-run - Produce deploy config files, but do not execute deploy
--f (disabled) Deploy on existing Salt master
--e (disabled) Do not launch environment deployment
--F (disabled) Do only create a Salt master
--T (disabled) Timeout, in minutes, for the deploy.
- It defaults to using the DEPLOY_TIMEOUT environment variable when defined.
--i (disabled) .iso image to be deployed (needs to be provided in a URI
- style, it can be a local resource: file:// or a remote resource http(s)://)
-
-$(notify "[NOTE] sudo & virsh priviledges are needed for this script to run" 3)
+$(notify_i "[NOTE] sudo & virsh priviledges are needed for this script to run" 3)
Example:
-$(notify "sudo $(basename "$0") \\
- -b file:///home/jenkins/lab-config \\
+$(notify_i "sudo $(basename "$0") \\
+ -b file:///home/jenkins/securedlab \\
-l lf -p pod2 \\
-s os-odl-nofeature-ha" 2)
EOF
# END of usage description
##############################################################################
-##############################################################################
-# BEGIN of colored notification wrapper
-#
-notify() {
- tput setaf "${2:-1}" || true
- echo -en "${1:-"[WARN] Unsupported opt arg: $3\\n"}"
- tput sgr0
-}
-#
-# END of colored notification wrapper
-##############################################################################
-
-##############################################################################
-# BEGIN of deployment clean-up
-#
-clean() {
- echo "Cleaning up deploy tmp directories"
-}
-#
-# END of deployment clean-up
-##############################################################################
-
##############################################################################
# BEGIN of variables to customize
#
-SCRIPT_PATH=$(readlink -f "$(dirname "${BASH_SOURCE[0]}")")
-DEPLOY_DIR=$(cd "${SCRIPT_PATH}/../mcp/scripts"; pwd)
-STORAGE_DIR=$(cd "${SCRIPT_PATH}/../mcp/deploy/images"; pwd)
-RECLASS_CLUSTER_DIR=$(cd "${SCRIPT_PATH}/../mcp/reclass/classes/cluster"; pwd)
-DEPLOY_TYPE='baremetal'
+CI_DEBUG=${CI_DEBUG:-0}; [[ "${CI_DEBUG}" =~ (false|0) ]] || set -x
+REPO_ROOT_PATH=$(readlink -f "$(dirname "${BASH_SOURCE[0]}")/..")
+DEPLOY_DIR=$(cd "${REPO_ROOT_PATH}/mcp/scripts"; pwd)
+STORAGE_DIR=$(cd "${REPO_ROOT_PATH}/mcp/deploy/images"; pwd)
+BR_NAMES=('admin' 'mgmt' 'private' 'public')
OPNFV_BRIDGES=('pxebr' 'mgmt' 'internal' 'public')
URI_REGEXP='(file|https?|ftp)://.*'
+BASE_CONFIG_URI="file://${REPO_ROOT_PATH}/mcp/scripts/pharos"
-export SSH_KEY=${SSH_KEY:-"/var/lib/opnfv/mcp.rsa"}
-export SALT_MASTER=${INSTALLER_IP:-10.20.0.2}
-export SALT_MASTER_USER=${SALT_MASTER_USER:-ubuntu}
-export MAAS_IP=${MAAS_IP:-${SALT_MASTER%.*}.3}
-export MAAS_PXE_NETWORK=${MAAS_PXE_NETWORK:-192.168.11.0}
+# Customize deploy workflow
+DRY_RUN=${DRY_RUN:-0}
+USE_EXISTING_PKGS=${USE_EXISTING_PKGS:-0}
+USE_EXISTING_INFRA=${USE_EXISTING_INFRA:-0}
+INFRA_CREATION_ONLY=${INFRA_CREATION_ONLY:-0}
+NO_DEPLOY_ENVIRONMENT=${NO_DEPLOY_ENVIRONMENT:-0}
+ERASE_ENV=${ERASE_ENV:-0}
-# Derivated from above global vars
-export MCP_CTRL_NETWORK_ROOTSTR=${SALT_MASTER%.*}
-export MAAS_PXE_NETWORK_ROOTSTR=${MAAS_PXE_NETWORK%.*}
-export SSH_OPTS="-o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -i ${SSH_KEY}"
-export SSH_SALT="${SALT_MASTER_USER}@${SALT_MASTER}"
+source "${DEPLOY_DIR}/globals.sh"
+source "${DEPLOY_DIR}/lib.sh"
-# Variables below are disabled for now, to be re-introduced or removed later
-set +x
-USE_EXISTING_FUEL=''
-FUEL_CREATION_ONLY=''
-NO_DEPLOY_ENVIRONMENT=''
-DRY_RUN=0
-if ! [ -z "${DEPLOY_TIMEOUT}" ]; then
- DEPLOY_TIMEOUT="-dt ${DEPLOY_TIMEOUT}"
-else
- DEPLOY_TIMEOUT=""
-fi
-set -x
#
# END of variables to customize
##############################################################################
# BEGIN of main
#
set +x
-OPNFV_BRIDGE_IDX=0
-while getopts "b:B:dfFl:L:p:s:S:T:i:he" OPTION
+while getopts "b:dDfEFl:L:p:Ps:S:he" OPTION
do
case $OPTION in
b)
BASE_CONFIG_URI=${OPTARG}
if [[ ! $BASE_CONFIG_URI =~ ${URI_REGEXP} ]]; then
- notify "[ERROR] -b $BASE_CONFIG_URI - invalid URI\n"
+ notify "[ERROR] -b $BASE_CONFIG_URI - invalid URI"
usage
exit 1
fi
;;
- B)
- OIFS=${IFS}
- IFS=','
- OPT_BRIDGES=($OPTARG)
- for bridge in "${OPT_BRIDGES[@]}"; do
- if [ -n "${bridge}" ]; then
- OPNFV_BRIDGES[${OPNFV_BRIDGE_IDX}]="${bridge}"
- fi
- OPNFV_BRIDGE_IDX=$((OPNFV_BRIDGE_IDX + 1))
- done
- IFS=${OIFS}
- ;;
d)
- notify '' 3 "${OPTION}"; continue
DRY_RUN=1
;;
+ D)
+ CI_DEBUG=1
+ ;;
f)
- notify '' 3 "${OPTION}"; continue
- USE_EXISTING_FUEL='-nf'
+ ((USE_EXISTING_INFRA+=1))
;;
F)
- notify '' 3 "${OPTION}"; continue
- FUEL_CREATION_ONLY='-fo'
+ INFRA_CREATION_ONLY=1
;;
e)
- notify '' 3 "${OPTION}"; continue
- NO_DEPLOY_ENVIRONMENT='-nde'
+ NO_DEPLOY_ENVIRONMENT=1
+ ;;
+ E)
+ ((ERASE_ENV+=1))
;;
l)
TARGET_LAB=${OPTARG}
;;
p)
TARGET_POD=${OPTARG}
- if [[ "${TARGET_POD}" =~ "virtual" ]]; then
- DEPLOY_TYPE='virtual'
+ if [[ "${TARGET_POD}" =~ virtual ]]; then
+ # All vPODs will use 'local-virtual1' PDF/IDF for now
+ TARGET_LAB='local'
+ TARGET_POD='virtual1'
fi
;;
+ P)
+ USE_EXISTING_PKGS=1
+ ;;
s)
DEPLOY_SCENARIO=${OPTARG}
;;
STORAGE_DIR="${OPTARG}"
fi
;;
- T)
- notify '' 3 "${OPTION}"; continue
- DEPLOY_TIMEOUT="-dt ${OPTARG}"
- ;;
- i)
- notify '' 3 "${OPTION}"; continue
- ISO=${OPTARG}
- if [[ ! $ISO =~ ${URI_REGEXP} ]]; then
- notify "[ERROR] -i $ISO - invalid URI\n"
- usage
- exit 1
- fi
- ;;
h)
usage
exit 0
;;
*)
- notify "[ERROR] Arguments not according to new argument style\n"
- exit 1
+ notify_e "[ERROR] Unsupported arg, see -h for help"
;;
esac
done
if [[ "$(sudo whoami)" != 'root' ]]; then
- notify "This script requires sudo rights\n" 1>&2
- exit 1
-fi
-
-if ! virsh list >/dev/null 2>&1; then
- notify "This script requires hypervisor access\n" 1>&2
- exit 1
+ notify_e "[ERROR] This script requires sudo rights"
fi
# Validate mandatory arguments are set
if [ -z "${TARGET_LAB}" ] || [ -z "${TARGET_POD}" ] || \
[ -z "${DEPLOY_SCENARIO}" ]; then
- notify "[ERROR] At least one of the mandatory args is missing!\n" 1>&2
usage
- exit 1
+ notify_e "[ERROR] At least one of the mandatory args is missing!"
fi
-set -x
+[[ "${CI_DEBUG}" =~ (false|0) ]] || set -x
# Enable the automatic exit trap
trap do_exit SIGINT SIGTERM EXIT
# Set no restrictive umask so that Jenkins can remove any residuals
umask 0000
-clean
-
pushd "${DEPLOY_DIR}" > /dev/null
# Prepare the deploy config files based on lab/pod information, deployment
# scenario, etc.
-# Install required packages
-[ -n "$(command -v apt-get)" ] && sudo apt-get install -y \
- git make rsync mkisofs curl virtinst cpu-checker qemu-kvm
-[ -n "$(command -v yum)" ] && sudo yum install -y --skip-broken \
- git make rsync genisoimage curl virt-install qemu-kvm
+# Install required packages on jump server
+if [ ${USE_EXISTING_PKGS} -eq 1 ]; then
+ notify "[NOTE] Skipping distro pkg installation" 2
+else
+ notify "[NOTE] Installing required distro pkgs" 2
+ if [ -n "$(command -v apt-get)" ]; then
+ pkg_type='deb'; pkg_cmd='sudo apt-get install -y'
+ else
+ pkg_type='rpm'; pkg_cmd='sudo yum install -y --skip-broken'
+ fi
+ eval "$(parse_yaml "./requirements_${pkg_type}.yaml")"
+ for section in 'common' "$(uname -m)"; do
+ section_var="requirements_pkg_${section}[*]"
+ pkg_list+=" ${!section_var}"
+ done
+ # shellcheck disable=SC2086
+ ${pkg_cmd} ${pkg_list}
+fi
-if [ "$(uname -i)" = "aarch64" ]; then
- [ -n "$(command -v apt-get)" ] && sudo apt-get install -y vgabios && \
- sudo ln -sf /usr/share/vgabios/vgabios.bin /usr/share/qemu/vgabios-stdvga.bin
- [ -n "$(command -v yum)" ] && sudo yum install -y --skip-broken vgabios
+if ! virsh list >/dev/null 2>&1; then
+ notify_e "[ERROR] This script requires hypervisor access"
fi
+# Collect jump server system information for deploy debugging
+./sysinfo_print.sh
+
+# Clone git submodules and apply our patches
+make -C "${REPO_ROOT_PATH}/mcp/patches" deepclean patches-import
+
+# Expand scenario files, pod_config based on PDF
+SCENARIO_DIR="$(readlink -f "../config/scenario")"
+do_templates "${REPO_ROOT_PATH}" "${STORAGE_DIR}" "${TARGET_LAB}" \
+ "${TARGET_POD}" "${BASE_CONFIG_URI}" "${SCENARIO_DIR}"
+
# Check scenario file existence
-SCENARIO_DIR="../config/scenario"
-if [ ! -f "${SCENARIO_DIR}/${DEPLOY_TYPE}/${DEPLOY_SCENARIO}.yaml" ]; then
- notify "[WARN] ${DEPLOY_SCENARIO}.yaml not found! \
- Setting simplest scenario (os-nosdn-nofeature-noha)\n" 3
- DEPLOY_SCENARIO='os-nosdn-nofeature-noha'
- if [ ! -f "${SCENARIO_DIR}/${DEPLOY_TYPE}/${DEPLOY_SCENARIO}.yaml" ]; then
- notify "[ERROR] Scenario definition file is missing!\n" 1>&2
- exit 1
- fi
+if [ ! -f "${SCENARIO_DIR}/${DEPLOY_SCENARIO}.yaml" ]; then
+ notify_e "[ERROR] Scenario definition file is missing!"
fi
# Check defaults file existence
if [ ! -f "${SCENARIO_DIR}/defaults-$(uname -i).yaml" ]; then
- notify "[ERROR] Scenario defaults file is missing!\n" 1>&2
- exit 1
+ notify_e "[ERROR] Scenario defaults file is missing!"
fi
-# Get required infra deployment data
-source lib.sh
+# Get scenario data and (jumpserver) arch defaults
eval "$(parse_yaml "${SCENARIO_DIR}/defaults-$(uname -i).yaml")"
-eval "$(parse_yaml "${SCENARIO_DIR}/${DEPLOY_TYPE}/${DEPLOY_SCENARIO}.yaml")"
-
+eval "$(parse_yaml "${SCENARIO_DIR}/${DEPLOY_SCENARIO}.yaml")"
export CLUSTER_DOMAIN=${cluster_domain}
-declare -A virtual_nodes_ram virtual_nodes_vcpus
+# Expand jinja2 templates based on PDF data and env vars
+do_templates "${REPO_ROOT_PATH}" "${STORAGE_DIR}" "${TARGET_LAB}" \
+ "${TARGET_POD}" "${BASE_CONFIG_URI}"
+
+# Get required infra deployment data based on PDF/IDF (after template parsing)
+set +x
+eval "$(parse_yaml "${STORAGE_DIR}/pod_config.yml")"
+[[ "${CI_DEBUG}" =~ (false|0) ]] || set -x
+
+# Serialize vnode data as '<name0>,<ram0>,<vcpu0>|<name1>,<ram1>,<vcpu1>[...]'
for node in "${virtual_nodes[@]}"; do
virtual_custom_ram="virtual_${node}_ram"
virtual_custom_vcpus="virtual_${node}_vcpus"
- virtual_nodes_ram[$node]=${!virtual_custom_ram:-$virtual_default_ram}
- virtual_nodes_vcpus[$node]=${!virtual_custom_vcpus:-$virtual_default_vcpus}
+ virtual_nodes_data+="${node},"
+ virtual_nodes_data+="${!virtual_custom_ram:-$virtual_default_ram},"
+ virtual_nodes_data+="${!virtual_custom_vcpus:-$virtual_default_vcpus}|"
+done
+virtual_nodes_data=${virtual_nodes_data%|}
+
+# Serialize repos, packages to (pre-)install/remove for:
+# - foundation node VM base image (virtual: all VMs, baremetal: cfg01|mas01)
+# - virtualized control plane VM base image (only when VCP is used)
+base_image_flavors=common
+if [[ "${cluster_states[*]}" =~ virtual_control ]]; then
+ base_image_flavors+=" control"
+fi
+for sc in ${base_image_flavors}; do
+ for va in apt_keys apt_repos pkg_install pkg_remove; do
+ key=virtual_${sc}_${va}
+ eval "${key}=\${${key}[@]// /|}"
+ eval "${key}=\${${key}// /,}"
+ virtual_repos_pkgs+="${!key}^"
+ done
done
+virtual_repos_pkgs=${virtual_repos_pkgs%^}
-# Expand reclass and virsh network templates
-for tp in "${RECLASS_CLUSTER_DIR}/all-mcp-ocata-common/opnfv/"*.template \
- net_*.template; do envsubst < "${tp}" > "${tp%.template}"; done
+# Determine 'admin', 'mgmt', 'private' and 'public' bridge names based on IDF
+for ((i = 0; i < ${#BR_NAMES[@]}; i++)); do
+ br_jump=$(eval echo "\$parameters__param_opnfv_jump_bridge_${BR_NAMES[i]}")
+ if [ -n "${br_jump}" ] && [ "${br_jump}" != 'None' ]; then
+ OPNFV_BRIDGES[${i}]="${br_jump}"
+ fi
+done
+notify "[NOTE] Using bridges: ${OPNFV_BRIDGES[*]}" 2
-# Infra setup
-generate_ssh_key
-prepare_vms virtual_nodes "${base_image}" "${STORAGE_DIR}"
-create_networks OPNFV_BRIDGES
-create_vms virtual_nodes virtual_nodes_ram virtual_nodes_vcpus \
- OPNFV_BRIDGES "${STORAGE_DIR}"
-update_mcpcontrol_network
-start_vms virtual_nodes
-check_connection
+# Jumpserver prerequisites check
+jumpserver_check_requirements "${virtual_nodes[*]}" "${OPNFV_BRIDGES[@]}"
-./salt.sh
+# Infra setup
+if [ ${DRY_RUN} -eq 1 ]; then
+ notify "[NOTE] Dry run, skipping all deployment tasks" 2
+ exit 0
+elif [ ${USE_EXISTING_INFRA} -gt 0 ]; then
+ notify "[NOTE] Use existing infra" 2
+ check_connection
+else
+ generate_ssh_key
+ prepare_vms "${base_image}" "${STORAGE_DIR}" "${virtual_repos_pkgs}" \
+ "${virtual_nodes[@]}"
+ create_networks "${OPNFV_BRIDGES[@]}"
+ do_sysctl_cfg
+ create_vms "${STORAGE_DIR}" "${virtual_nodes_data}" "${OPNFV_BRIDGES[@]}"
+ update_mcpcontrol_network
+ start_vms "${virtual_nodes[@]}"
+ check_connection
+fi
+if [ ${USE_EXISTING_INFRA} -lt 2 ]; then
+ wait_for 5 "./salt.sh ${STORAGE_DIR}/pod_config.yml ${virtual_nodes[*]}"
+fi
# Openstack cluster setup
-for state in "${cluster_states[@]}"; do
- notify "STATE: ${state}\n" 2
- # shellcheck disable=SC2086,2029
- ssh ${SSH_OPTS} "ubuntu@${SALT_MASTER}" \
- sudo "/root/fuel/mcp/config/states/${state} || true"
-done
+set +x
+if [ ${INFRA_CREATION_ONLY} -eq 1 ] || [ ${NO_DEPLOY_ENVIRONMENT} -eq 1 ]; then
+ notify "[NOTE] Skip openstack cluster setup" 2
+else
+ for state in "${cluster_states[@]}"; do
+ notify "[STATE] Applying state: ${state}" 2
+ # shellcheck disable=SC2086,2029
+ wait_for 5 "ssh ${SSH_OPTS} ${SSH_SALT} sudo \
+ CI_DEBUG=$CI_DEBUG ERASE_ENV=$ERASE_ENV \
+ /root/fuel/mcp/config/states/${state}"
+ done
+fi
./log.sh "${DEPLOY_LOG}"