local RC=$?
cleanup_mounts > /dev/null 2>&1
if [ ${RC} -eq 0 ]; then
- notify "\n[OK] MCP: Openstack installation finished succesfully!\n\n" 2
+ notify_n "[OK] MCP: Openstack installation finished succesfully!" 2
else
- notify "\n[ERROR] MCP: Openstack installation threw a fatal error!\n\n"
+ notify_n "[ERROR] MCP: Openstack installation threw a fatal error!"
fi
}
#
-S Storage dir for VM images
-L Deployment log path and file name
-$(notify "Description:" 2)
+$(notify_i "Description:" 2)
Deploys the Fuel@OPNFV stack on the indicated lab resource.
This script provides the Fuel@OPNFV deployment abstraction.
It depends on the OPNFV official configuration directory/file structure
and provides a fairly simple mechanism to execute a deployment.
-$(notify "Input parameters to the build script are:" 2)
+$(notify_i "Input parameters to the build script are:" 2)
-b Base URI to the configuration directory (needs to be provided in URI style,
it can be a local resource: file:// or a remote resource http(s)://).
A POD Descriptor File (PDF) and its Installer Descriptor File (IDF)
has to be defined in config directory (e.g. os-odl-nofeature-ha).
-S Storage dir for VM images, default is mcp/deploy/images
-$(notify "[NOTE] sudo & virsh priviledges are needed for this script to run" 3)
+$(notify_i "[NOTE] sudo & virsh priviledges are needed for this script to run" 3)
Example:
-$(notify "sudo $(basename "$0") \\
+$(notify_i "sudo $(basename "$0") \\
-b file:///home/jenkins/securedlab \\
-l lf -p pod2 \\
-s os-odl-nofeature-ha" 2)
# END of usage description
##############################################################################
-##############################################################################
-# BEGIN of colored notification wrapper
-#
-notify() {
- tput setaf "${2:-1}" || true
- echo -en "${1:-"[WARN] Unsupported opt arg: $3\\n"}"
- tput sgr0
-}
-#
-# END of colored notification wrapper
-##############################################################################
-
##############################################################################
# BEGIN of variables to customize
#
REPO_ROOT_PATH=$(readlink -f "$(dirname "${BASH_SOURCE[0]}")/..")
DEPLOY_DIR=$(cd "${REPO_ROOT_PATH}/mcp/scripts"; pwd)
STORAGE_DIR=$(cd "${REPO_ROOT_PATH}/mcp/deploy/images"; pwd)
-RECLASS_CLUSTER_DIR=$(cd "${REPO_ROOT_PATH}/mcp/reclass/classes/cluster"; pwd)
-DEPLOY_TYPE='baremetal'
BR_NAMES=('admin' 'mgmt' 'private' 'public')
OPNFV_BRIDGES=('pxebr' 'mgmt' 'internal' 'public')
URI_REGEXP='(file|https?|ftp)://.*'
b)
BASE_CONFIG_URI=${OPTARG}
if [[ ! $BASE_CONFIG_URI =~ ${URI_REGEXP} ]]; then
- notify "[ERROR] -b $BASE_CONFIG_URI - invalid URI\n"
+ notify "[ERROR] -b $BASE_CONFIG_URI - invalid URI"
usage
exit 1
fi
p)
TARGET_POD=${OPTARG}
if [[ "${TARGET_POD}" =~ virtual ]]; then
- DEPLOY_TYPE='virtual'
# All vPODs will use 'local-virtual1' PDF/IDF for now
TARGET_LAB='local'
TARGET_POD='virtual1'
exit 0
;;
*)
- notify "[ERROR] Arguments not according to new argument style\n"
- exit 1
+ notify_e "[ERROR] Unsupported arg, see -h for help"
;;
esac
done
if [[ "$(sudo whoami)" != 'root' ]]; then
- notify "[ERROR] This script requires sudo rights\n" 1>&2
- exit 1
+ notify_e "[ERROR] This script requires sudo rights"
fi
# Validate mandatory arguments are set
if [ -z "${TARGET_LAB}" ] || [ -z "${TARGET_POD}" ] || \
[ -z "${DEPLOY_SCENARIO}" ]; then
- notify "[ERROR] At least one of the mandatory args is missing!\n" 1>&2
usage
- exit 1
+ notify_e "[ERROR] At least one of the mandatory args is missing!"
fi
[[ "${CI_DEBUG}" =~ (false|0) ]] || set -x
# Install required packages on jump server
if [ ${USE_EXISTING_PKGS} -eq 1 ]; then
- notify "[NOTE] Skipping distro pkg installation\n" 2 1>&2
+ notify "[NOTE] Skipping distro pkg installation" 2
else
- notify "[NOTE] Installing required distro pkgs\n" 2 1>&2
+ notify "[NOTE] Installing required distro pkgs" 2
if [ -n "$(command -v apt-get)" ]; then
pkg_type='deb'; pkg_cmd='sudo apt-get install -y'
else
pkg_type='rpm'; pkg_cmd='sudo yum install -y --skip-broken'
fi
eval "$(parse_yaml "./requirements_${pkg_type}.yaml")"
- for section in 'common' "${DEPLOY_TYPE}" "$(uname -m)"; do
+ for section in 'common' "$(uname -m)"; do
section_var="requirements_pkg_${section}[*]"
pkg_list+=" ${!section_var}"
done
fi
if ! virsh list >/dev/null 2>&1; then
- notify "[ERROR] This script requires hypervisor access\n" 1>&2
- exit 1
+ notify_e "[ERROR] This script requires hypervisor access"
fi
# Collect jump server system information for deploy debugging
# Clone git submodules and apply our patches
make -C "${REPO_ROOT_PATH}/mcp/patches" deepclean patches-import
-# Convert Pharos-compatible POD Descriptor File (PDF) to reclass model input
-PHAROS_GEN_CONFIG_SCRIPT="./pharos/config/utils/generate_config.py"
-PHAROS_INSTALLER_ADAPTER="./pharos/config/installers/fuel/pod_config.yml.j2"
-BASE_CONFIG_PDF="${BASE_CONFIG_URI}/labs/${TARGET_LAB}/${TARGET_POD}.yaml"
-BASE_CONFIG_IDF="${BASE_CONFIG_URI}/labs/${TARGET_LAB}/idf-${TARGET_POD}.yaml"
-LOCAL_PDF="${STORAGE_DIR}/$(basename "${BASE_CONFIG_PDF}")"
-LOCAL_IDF="${STORAGE_DIR}/$(basename "${BASE_CONFIG_IDF}")"
-LOCAL_PDF_RECLASS="${STORAGE_DIR}/pod_config.yml"
-rm -f "${LOCAL_PDF_RECLASS}"
-ln -sf "$(readlink -f "../config/labs/local")" "./pharos/labs/"
-if ! curl --create-dirs -o "${LOCAL_PDF}" "${BASE_CONFIG_PDF}"; then
- notify "[ERROR] Could not retrieve PDF (Pod Descriptor File)!\n" 1>&2
- exit 1
-elif ! curl -o "${LOCAL_IDF}" "${BASE_CONFIG_IDF}"; then
- notify "[ERROR] Could not retrieve IDF (Installer Descriptor File)!\n" 1>&2
- exit 1
-elif ! "${PHAROS_GEN_CONFIG_SCRIPT}" -y "${LOCAL_PDF}" \
- -j "${PHAROS_INSTALLER_ADAPTER}" > "${LOCAL_PDF_RECLASS}"; then
- notify "[ERROR] Could not convert PDF+IDF to reclass model input!\n" 1>&2
- exit 1
-fi
+# Expand scenario files, pod_config based on PDF
+SCENARIO_DIR="$(readlink -f "../config/scenario")"
+do_templates "${REPO_ROOT_PATH}" "${STORAGE_DIR}" "${TARGET_LAB}" \
+ "${TARGET_POD}" "${BASE_CONFIG_URI}" "${SCENARIO_DIR}"
# Check scenario file existence
-SCENARIO_DIR="../config/scenario"
-if [ ! -f "${SCENARIO_DIR}/${DEPLOY_TYPE}/${DEPLOY_SCENARIO}.yaml" ]; then
- notify "[ERROR] Scenario definition file is missing!\n" 1>&2
- exit 1
+if [ ! -f "${SCENARIO_DIR}/${DEPLOY_SCENARIO}.yaml" ]; then
+ notify_e "[ERROR] Scenario definition file is missing!"
fi
# Check defaults file existence
if [ ! -f "${SCENARIO_DIR}/defaults-$(uname -i).yaml" ]; then
- notify "[ERROR] Scenario defaults file is missing!\n" 1>&2
- exit 1
+ notify_e "[ERROR] Scenario defaults file is missing!"
fi
-# Get required infra deployment data
-set +x
+# Get scenario data and (jumpserver) arch defaults
eval "$(parse_yaml "${SCENARIO_DIR}/defaults-$(uname -i).yaml")"
-eval "$(parse_yaml "${SCENARIO_DIR}/${DEPLOY_TYPE}/${DEPLOY_SCENARIO}.yaml")"
-eval "$(parse_yaml "${LOCAL_PDF_RECLASS}")"
-[[ "${CI_DEBUG}" =~ (false|0) ]] || set -x
-
+eval "$(parse_yaml "${SCENARIO_DIR}/${DEPLOY_SCENARIO}.yaml")"
export CLUSTER_DOMAIN=${cluster_domain}
+# Expand jinja2 templates based on PDF data and env vars
+do_templates "${REPO_ROOT_PATH}" "${STORAGE_DIR}" "${TARGET_LAB}" \
+ "${TARGET_POD}" "${BASE_CONFIG_URI}"
+
+# Get required infra deployment data based on PDF/IDF (after template parsing)
+set +x
+eval "$(parse_yaml "${STORAGE_DIR}/pod_config.yml")"
+[[ "${CI_DEBUG}" =~ (false|0) ]] || set -x
+
# Serialize vnode data as '<name0>,<ram0>,<vcpu0>|<name1>,<ram1>,<vcpu1>[...]'
for node in "${virtual_nodes[@]}"; do
virtual_custom_ram="virtual_${node}_ram"
done
virtual_repos_pkgs=${virtual_repos_pkgs%^}
-# Expand reclass and virsh network templates
-for tp in "${RECLASS_CLUSTER_DIR}/all-mcp-arch-common/opnfv/"*.template \
- net_*.template; do
- eval "cat <<-EOF
- $(<"${tp}")
- EOF" 2> /dev/null > "${tp%.template}"
-done
-
-# Convert Pharos-compatible PDF to reclass network definitions
-find "${RECLASS_CLUSTER_DIR}" -name '*.j2' | while read -r tp
-do
- if ! "${PHAROS_GEN_CONFIG_SCRIPT}" -y "${LOCAL_PDF}" \
- -j "${tp}" > "${tp%.j2}"; then
- notify "[ERROR] Could not convert PDF to reclass network defs!\n"
- exit 1
- fi
-done
-
# Determine 'admin', 'mgmt', 'private' and 'public' bridge names based on IDF
for ((i = 0; i < ${#BR_NAMES[@]}; i++)); do
br_jump=$(eval echo "\$parameters__param_opnfv_jump_bridge_${BR_NAMES[i]}")
OPNFV_BRIDGES[${i}]="${br_jump}"
fi
done
-notify "[NOTE] Using bridges: ${OPNFV_BRIDGES[*]}\n" 2
+notify "[NOTE] Using bridges: ${OPNFV_BRIDGES[*]}" 2
+
+# Jumpserver prerequisites check
+jumpserver_check_requirements "${virtual_nodes[*]}" "${OPNFV_BRIDGES[@]}"
# Infra setup
if [ ${DRY_RUN} -eq 1 ]; then
- notify "[NOTE] Dry run, skipping all deployment tasks\n" 2 1>&2
+ notify "[NOTE] Dry run, skipping all deployment tasks" 2
exit 0
elif [ ${USE_EXISTING_INFRA} -gt 0 ]; then
- notify "[NOTE] Use existing infra\n" 2 1>&2
+ notify "[NOTE] Use existing infra" 2
check_connection
else
generate_ssh_key
check_connection
fi
if [ ${USE_EXISTING_INFRA} -lt 2 ]; then
- wait_for 5 "./salt.sh ${LOCAL_PDF_RECLASS}"
+ wait_for 5 "./salt.sh ${STORAGE_DIR}/pod_config.yml ${virtual_nodes[*]}"
fi
# Openstack cluster setup
set +x
if [ ${INFRA_CREATION_ONLY} -eq 1 ] || [ ${NO_DEPLOY_ENVIRONMENT} -eq 1 ]; then
- notify "[NOTE] Skip openstack cluster setup\n" 2
+ notify "[NOTE] Skip openstack cluster setup" 2
else
for state in "${cluster_states[@]}"; do
- notify "[STATE] Applying state: ${state}\n" 2
+ notify "[STATE] Applying state: ${state}" 2
# shellcheck disable=SC2086,2029
wait_for 5 "ssh ${SSH_OPTS} ${SSH_SALT} sudo \
CI_DEBUG=$CI_DEBUG ERASE_ENV=$ERASE_ENV \