X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=ci%2Fdeploy.sh;h=a61946e6c06ab8ee9e2bc6c11c1c73fe054d4fc9;hb=aa31e99c06c7c003ed5339ada7b962689d16c030;hp=10b639e3c2a67ec9908c9d35378d5c168df95e82;hpb=d151d4342aa828e0f4c94e7390262080686fcfa3;p=fuel.git diff --git a/ci/deploy.sh b/ci/deploy.sh index 10b639e3c..a61946e6c 100755 --- a/ci/deploy.sh +++ b/ci/deploy.sh @@ -32,13 +32,13 @@ usage () { cat << EOF xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx -$(notify "$(basename "$0"): Deploy the Fuel@OPNFV MCP stack" 3) +$(notify "$(basename "$0"): Deploy the OPNFV Fuel MCP stack" 3) $(notify "USAGE:" 2) $(basename "$0") -l lab-name -p pod-name -s deploy-scenario \\ [-b Lab Config Base URI] \\ [-S storage-dir] [-L /path/to/log/file.tar.gz] \\ - [-f[f]] [-F] [-e | -E[E]] [-d] [-D] [-N] + [-f] [-F] [-e | -E[E]] [-d] [-D] [-N] [-m] $(notify "OPTIONS:" 2) -b Base-uri for the stack-configuration structure @@ -46,21 +46,22 @@ $(notify "OPTIONS:" 2) -D Debug logging -e Do not launch environment deployment -E Remove existing VCP VMs (use twice to redeploy baremetal nodes) - -f Deploy on existing Salt master (use twice to also skip config sync) - -F Do only create a Salt master + -f Deploy on existing Salt master (use twice or more to skip states) + -F Same as -e, do not launch environment deployment (legacy option) -h Print this message and exit -l Lab-name -p Pod-name -P Skip installation of package dependencies -s Deploy-scenario short-name - -S Storage dir for VM images + -S Storage dir for VM images and other deploy artifacts -L Deployment log path and file name + -m Use single socket CPU compute nodes (only affects virtual computes) -N Experimental: Do not virtualize control plane (novcp) $(notify_i "Description:" 2) -Deploys the Fuel@OPNFV stack on the indicated lab resource. +Deploys the OPNFV Fuel stack on the indicated lab resource. -This script provides the Fuel@OPNFV deployment abstraction. +This script provides the OPNFV Fuel deployment abstraction. It depends on the OPNFV official configuration directory/file structure and provides a fairly simple mechanism to execute a deployment. @@ -73,8 +74,6 @@ $(notify_i "Input parameters to the build script are:" 2) /labs//idf-.yaml The default is using the git submodule tracking 'OPNFV Pharos' in <./mcp/scripts/pharos>. - An example config is provided inside current repo in - <./mcp/config>, automatically linked as <./mcp/scripts/pharos/labs/local>. -d Dry-run - Produce deploy config files, but do not execute deploy -D Debug logging - Enable extra logging in sh deploy scripts (set -x) -e Do not launch environment deployment @@ -84,16 +83,17 @@ $(notify_i "Input parameters to the build script are:" 2) Only applicable for baremetal deploys. -f Deploy on existing Salt master. It will skip infrastructure VM creation, but it will still sync reclass configuration from current repo to Salt - Master node. If specified twice (e.g. -f -f), config sync will also be - skipped. --F Do only create a Salt master + Master node. + Each additional use skips one more state file. For example, -fff would + skip the first 3 state files (e.g. virtual_init, maas, baremetal_init). +-F Same as -e, do not launch environment deployment (legacy option) -h Print this message and exit -L Deployment log path and name, eg. -L /home/jenkins/job.log.tar.gz -l Lab name as defined in the configuration directory, e.g. lf - For the sample configuration in <./mcp/config>, lab name is 'local'. -p POD name as defined in the configuration directory, e.g. pod2 - For the sample configuration in <./mcp/config>, POD name is 'virtual1' - for virtual deployments or 'pod1' for baremetal (based on lf-pod2). +-m Use single socket compute nodes. Instead of using default NUMA-enabled + topology for virtual compute nodes created via libvirt, configure a + single guest CPU socket. -N Experimental: Instead of virtualizing the control plane (VCP), deploy control plane directly on baremetal nodes -P Skip installing dependency distro packages on current host @@ -102,7 +102,9 @@ $(notify_i "Input parameters to the build script are:" 2) Note that without the required packages, deploy will fail. -s Deployment-scenario, this points to a short deployment scenario name, which has to be defined in config directory (e.g. os-odl-nofeature-ha). --S Storage dir for VM images, default is mcp/deploy/images +-S Storage dir for VM images, default is /var/lib/opnfv/tmpdir + It is recommended to store the deploy artifacts on a fast disk, outside of + the current git repository (so clean operations won't erase it). $(notify_i "[NOTE] sudo & virsh priviledges are needed for this script to run" 3) @@ -111,7 +113,8 @@ Example: $(notify_i "sudo $(basename "$0") \\ -b file:///home/jenkins/securedlab \\ -l lf -p pod2 \\ - -s os-odl-nofeature-ha" 2) + -s os-odl-nofeature-ha \\ + -S /home/jenkins/tmpdir" 2) EOF } @@ -123,24 +126,27 @@ EOF # BEGIN of variables to customize # CI_DEBUG=${CI_DEBUG:-0}; [[ "${CI_DEBUG}" =~ (false|0) ]] || set -x -REPO_ROOT_PATH=$(readlink -f "$(dirname "${BASH_SOURCE[0]}")/..") -DEPLOY_DIR=$(cd "${REPO_ROOT_PATH}/mcp/scripts"; pwd) -STORAGE_DIR=$(cd "${REPO_ROOT_PATH}/mcp/deploy/images"; pwd) +MCP_REPO_ROOT_PATH=$(readlink -f "$(dirname "${BASH_SOURCE[0]}")/..") +DEPLOY_DIR=$(cd "${MCP_REPO_ROOT_PATH}/mcp/scripts"; pwd) +MCP_STORAGE_DIR='/var/lib/opnfv/tmpdir' URI_REGEXP='(file|https?|ftp)://.*' -BASE_CONFIG_URI="file://${REPO_ROOT_PATH}/mcp/scripts/pharos" +BASE_CONFIG_URI="file://${MCP_REPO_ROOT_PATH}/mcp/scripts/pharos" # Customize deploy workflow DRY_RUN=${DRY_RUN:-0} USE_EXISTING_PKGS=${USE_EXISTING_PKGS:-0} USE_EXISTING_INFRA=${USE_EXISTING_INFRA:-0} -INFRA_CREATION_ONLY=${INFRA_CREATION_ONLY:-0} NO_DEPLOY_ENVIRONMENT=${NO_DEPLOY_ENVIRONMENT:-0} ERASE_ENV=${ERASE_ENV:-0} MCP_VCP=${MCP_VCP:-1} +MCP_DOCKER_TAG=${MCP_DOCKER_TAG:-latest} +MCP_CMP_SS=${MCP_CMP_SS:-0} source "${DEPLOY_DIR}/globals.sh" source "${DEPLOY_DIR}/lib.sh" source "${DEPLOY_DIR}/lib_template.sh" +source "${DEPLOY_DIR}/lib_jump_common.sh" +source "${DEPLOY_DIR}/lib_jump_deploy.sh" # # END of variables to customize @@ -170,10 +176,7 @@ do f) ((USE_EXISTING_INFRA+=1)) ;; - F) - INFRA_CREATION_ONLY=1 - ;; - e) + F|e) NO_DEPLOY_ENVIRONMENT=1 ;; E) @@ -185,6 +188,9 @@ do L) DEPLOY_LOG="${OPTARG}" ;; + m) + MCP_CMP_SS=1 + ;; N) MCP_VCP=0 ;; @@ -199,7 +205,7 @@ do ;; S) if [[ ${OPTARG} ]]; then - STORAGE_DIR="${OPTARG}" + MCP_STORAGE_DIR="${OPTARG}" fi ;; h) @@ -236,14 +242,18 @@ pushd "${DEPLOY_DIR}" > /dev/null # scenario, etc. # Install required packages on jump server +sudo mkdir -p "${MCP_STORAGE_DIR}" +sudo chown -R "${USER}:${USER}" "${MCP_STORAGE_DIR}" if [ ${USE_EXISTING_PKGS} -eq 1 ]; then notify "[NOTE] Skipping distro pkg installation" 2 else notify "[NOTE] Installing required distro pkgs" 2 - jumpserver_pkg_install + jumpserver_pkg_install 'deploy' + docker_install "${MCP_STORAGE_DIR}" + virtinst_install "${MCP_STORAGE_DIR}" fi -if ! virsh list >/dev/null 2>&1; then +if ! ${VIRSH} list >/dev/null 2>&1; then notify_e "[ERROR] This script requires hypervisor access" fi @@ -251,7 +261,7 @@ fi ./sysinfo_print.sh # Clone git submodules and apply our patches -make -C "${REPO_ROOT_PATH}/mcp/patches" deepclean patches-import +make -C "${MCP_REPO_ROOT_PATH}/mcp/patches" deepclean patches-import # Check scenario file existence SCENARIO_DIR="$(readlink -f "../config/scenario")" @@ -264,15 +274,16 @@ fi generate_ssh_key export MAAS_SSH_KEY="$(cat "$(basename "${SSH_KEY}").pub")" -MCP_DPDK_MODE=$([[ "$DEPLOY_SCENARIO" =~ ovs ]] && echo 1 || echo 0) # Expand jinja2 templates based on PDF data and env vars -export MCP_VCP MCP_DPDK_MODE MCP_JUMP_ARCH=$(uname -i) -do_templates_scenario "${STORAGE_DIR}" "${TARGET_LAB}" "${TARGET_POD}" \ - "${BASE_CONFIG_URI}" "${SCENARIO_DIR}" -do_templates_cluster "${STORAGE_DIR}" "${TARGET_LAB}" "${TARGET_POD}" \ - "${REPO_ROOT_PATH}" \ - "${SCENARIO_DIR}/defaults.yaml" \ +[[ "${DEPLOY_SCENARIO}" =~ -ha$ ]] || MCP_VCP=0 +export MCP_REPO_ROOT_PATH MCP_VCP MCP_STORAGE_DIR MCP_DOCKER_TAG MCP_CMP_SS \ + MCP_JUMP_ARCH=$(uname -i) MCP_DEPLOY_SCENARIO="${DEPLOY_SCENARIO}" +do_templates_scenario "${MCP_STORAGE_DIR}" "${TARGET_LAB}" "${TARGET_POD}" \ + "${BASE_CONFIG_URI}" "${SCENARIO_DIR}" \ "${SCENARIO_DIR}/${DEPLOY_SCENARIO}.yaml" +do_templates_cluster "${MCP_STORAGE_DIR}" "${TARGET_LAB}" "${TARGET_POD}" \ + "${MCP_REPO_ROOT_PATH}" \ + "${SCENARIO_DIR}/defaults.yaml" # Determine additional data (e.g. jump bridge names) based on XDF source "${DEPLOY_DIR}/xdf_data.sh" @@ -286,37 +297,47 @@ if [ ${DRY_RUN} -eq 1 ]; then notify "[NOTE] Dry run, skipping all deployment tasks" 2 exit 0 elif [ ${USE_EXISTING_INFRA} -gt 0 ]; then - notify "[NOTE] Use existing infra" 2 - check_connection + notify "[NOTE] Use existing infra: skip first ${USE_EXISTING_INFRA} states" 2 + notify "[STATE] Skipping: ${cluster_states[*]::${USE_EXISTING_INFRA}}" 2 else - prepare_vms "${base_image}" "${STORAGE_DIR}" "${virtual_repos_pkgs}" \ + prepare_vms "${base_image}" "${MCP_STORAGE_DIR}" "${virtual_repos_pkgs}" \ "${virtual_nodes[@]}" create_networks "${OPNFV_BRIDGES[@]}" do_sysctl_cfg - create_vms "${STORAGE_DIR}" "${virtual_nodes_data}" "${OPNFV_BRIDGES[@]}" + do_udev_cfg + create_vms "${MCP_STORAGE_DIR}" "${virtual_nodes_data}" "${OPNFV_BRIDGES[@]}" update_mcpcontrol_network start_vms "${virtual_nodes[@]}" - check_connection -fi -if [ ${USE_EXISTING_INFRA} -lt 2 ]; then - wait_for 5 "./salt.sh ${STORAGE_DIR}/pod_config.yml ${virtual_nodes[*]}" + + # https://github.com/docker/libnetwork/issues/1743 + # rm -f /var/lib/docker/network/files/local-kv.db + sudo systemctl restart docker + prepare_containers "${MCP_STORAGE_DIR}" fi +start_containers "${MCP_STORAGE_DIR}" +check_connection + # Openstack cluster setup set +x -if [ ${INFRA_CREATION_ONLY} -eq 1 ] || [ ${NO_DEPLOY_ENVIRONMENT} -eq 1 ]; then +if [ ${NO_DEPLOY_ENVIRONMENT} -eq 1 ]; then notify "[NOTE] Skip openstack cluster setup" 2 else - for state in "${cluster_states[@]}"; do + for state in "${cluster_states[@]:${USE_EXISTING_INFRA}}"; do notify "[STATE] Applying state: ${state}" 2 # shellcheck disable=SC2086,2029 wait_for 5 "ssh ${SSH_OPTS} ${SSH_SALT} sudo \ CI_DEBUG=$CI_DEBUG ERASE_ENV=$ERASE_ENV \ /root/fuel/mcp/config/states/${state}" + if [ "${state}" = 'maas' ]; then + # For hybrid PODs (virtual + baremetal nodes), the virtual nodes + # should be reset to force a DHCP request from MaaS DHCP + reset_vms "${virtual_nodes[@]}" + fi done -fi -./log.sh "${DEPLOY_LOG}" + ./log.sh "${DEPLOY_LOG}" +fi popd > /dev/null