##Post configuration after install
##params: none
function configure_post_install {
- local opnfv_attach_networks ovs_ip ip_range net_cidr tmp_ip
- opnfv_attach_networks="admin_network public_network"
+ local opnfv_attach_networks ovs_ip ip_range net_cidr tmp_ip af external_network_ipv6
+ external_network_ipv6=False
+ opnfv_attach_networks="admin"
+ if [[ $enabled_network_list =~ "external" ]]; then
+ opnfv_attach_networks+=' external'
+ fi
echo -e "${blue}INFO: Post Install Configuration Running...${reset}"
else
echo -e "${blue}INFO: OVS Bridge ${NET_MAP[$network]} missing IP, will configure${reset}"
# use last IP of allocation pool
- eval "ip_range=\${${network}_usable_ip_range}"
+ eval "ip_range=\${${network}_overcloud_ip_range}"
ovs_ip=${ip_range##*,}
eval "net_cidr=\${${network}_cidr}"
+ if [[ $ovs_ip =~ ^[0-9]+\.[0-9]+\.[0-9]+\.[0-9]+$ ]]; then
+ af=4
+ else
+ af=6
+ if [ "$network" == "external" ]; then
+ ublic_network_ipv6=True
+ fi
+ #enable ipv6 on bridge interface
+ echo 0 > /proc/sys/net/ipv6/conf/${NET_MAP[$network]}/disable_ipv6
+ fi
sudo ip addr add ${ovs_ip}/${net_cidr##*/} dev ${NET_MAP[$network]}
sudo ip link set up ${NET_MAP[$network]}
- tmp_ip=$(find_ip ${NET_MAP[$network]})
+ tmp_ip=$(find_ip ${NET_MAP[$network]} $af)
if [ -n "$tmp_ip" ]; then
echo -e "${blue}INFO: OVS Bridge ${NET_MAP[$network]} IP set: ${tmp_ip}${reset}"
continue
# TODO fix this when HA SDN controllers are supported
if [ "${deploy_options_array['sdn_controller']}" != 'False' ]; then
echo -e "${blue}INFO: Finding SDN Controller IP for overcloudrc...${reset}"
- sdn_controller_ip=$(overcloud_connect controller0 "facter ipaddress_br_ex")
+ sdn_controller_ip=$(undercloud_connect stack "source stackrc;nova list | grep controller-0 | cut -d '|' -f 7 | grep -Eo [0-9]+\.[0-9]+\.[0-9]+\.[0-9]+")
echo -e "${blue}INFO: SDN Controller IP is ${sdn_controller_ip} ${reset}"
undercloud_connect stack "echo 'export SDN_CONTROLLER_IP=${sdn_controller_ip}' >> /home/stack/overcloudrc"
fi
source overcloudrc
set -o errexit
echo "Configuring Neutron external network"
-neutron net-create external --router:external=True --tenant-id \$(openstack project show service | grep id | awk '{ print \$4 }')
-neutron subnet-create --name external-net --tenant-id \$(openstack project show service | grep id | awk '{ print \$4 }') --disable-dhcp external --gateway ${public_network_gateway} --allocation-pool start=${public_network_floating_ip_range%%,*},end=${public_network_floating_ip_range##*,} ${public_network_cidr}
+if [[ -n "$external_nic_mapping_compute_vlan" && "$external_nic_mapping_compute_vlan" != 'native' ]]; then
+ neutron net-create external --router:external=True --tenant-id \$(openstack project show service | grep id | awk '{ print \$4 }') --provider:network_type vlan --provider:segmentation_id ${external_nic_mapping_compute_vlan} --provider:physical_network datacentre
+else
+ neutron net-create external --router:external=True --tenant-id \$(openstack project show service | grep id | awk '{ print \$4 }') --provider:network_type flat --provider:physical_network datacentre
+fi
+if [ "$external_network_ipv6" == "True" ]; then
+ neutron subnet-create --name external-net --tenant-id \$(openstack project show service | grep id | awk '{ print \$4 }') external --ip_version 6 --ipv6_ra_mode slaac --ipv6_address_mode slaac --gateway ${external_gateway} --allocation-pool start=${external_floating_ip_range%%,*},end=${external_floating_ip_range##*,} ${external_cidr}
+elif [[ "$enabled_network_list" =~ "external" ]]; then
+ neutron subnet-create --name external-net --tenant-id \$(openstack project show service | grep id | awk '{ print \$4 }') --disable-dhcp external --gateway ${external_gateway} --allocation-pool start=${external_floating_ip_range%%,*},end=${external_floating_ip_range##*,} ${external_cidr}
+else
+ # we re-use the introspection range for floating ips with single admin network
+ neutron subnet-create --name external-net --tenant-id \$(openstack project show service | grep id | awk '{ print \$4 }') --disable-dhcp external --gateway ${admin_gateway} --allocation-pool start=${admin_introspection_range%%,*},end=${admin_introspection_range##*,} ${admin_cidr}
+fi
+
+if [ "${deploy_options_array['gluon']}" == 'True' ]; then
+ echo "Creating Gluon dummy network and subnet"
+ neutron net-create --shared --provider:network_type vxlan GluonNetwork
+ neutron subnet-create --name GluonSubnet --no-gateway --disable-dhcp GluonNetwork 0.0.0.0/1
+fi
echo "Removing sahara endpoint and service"
sahara_service_id=\$(openstack service list | grep sahara | cut -d ' ' -f 2)
sahara_endpoint_id=\$(openstack endpoint list | grep sahara | cut -d ' ' -f 2)
-openstack endpoint delete \$sahara_endpoint_id
-openstack service delete \$sahara_service_id
+[[ -n "\$sahara_endpoint_id" ]] && openstack endpoint delete \$sahara_endpoint_id
+[[ -n "\$sahara_service_id" ]] && openstack service delete \$sahara_service_id
echo "Removing swift endpoint and service"
swift_service_id=\$(openstack service list | grep swift | cut -d ' ' -f 2)
swift_endpoint_id=\$(openstack endpoint list | grep swift | cut -d ' ' -f 2)
-openstack endpoint delete \$swift_endpoint_id
-openstack service delete \$swift_service_id
+[[ -n "\$swift_endpoint_id" ]] && openstack endpoint delete \$swift_endpoint_id
+[[ -n "\$swift_service_id" ]] && openstack service delete \$swift_service_id
+
+if [ "${deploy_options_array['dataplane']}" == 'fdio' ] || [ "${deploy_options_array['dataplane']}" == 'ovs_dpdk' ]; then
+ for flavor in \$(openstack flavor list -c Name -f value); do
+ echo "INFO: Configuring \$flavor to use hugepage"
+ nova flavor-key \$flavor set hw:mem_page_size=large
+ done
+fi
if [ "${deploy_options_array['congress']}" == 'True' ]; then
- for s in nova neutronv2 ceilometer cinder glancev2 keystone; do
- openstack congress datasource create \$s "\$s" \\
- --config username=\$OS_USERNAME \\
- --config tenant_name=\$OS_TENANT_NAME \\
- --config password=\$OS_PASSWORD \\
- --config auth_url=\$OS_AUTH_URL
+ ds_configs="--config username=\$OS_USERNAME
+ --config tenant_name=\$OS_PROJECT_NAME
+ --config password=\$OS_PASSWORD
+ --config auth_url=\$OS_AUTH_URL"
+ for s in nova neutronv2 cinder glancev2 keystone; do
+ ds_extra_configs=""
+ if [ "\$s" == "nova" ]; then
+ # nova's latest version is 2.38 but congress relies on nova to do
+ # floating ip operation instead of neutron. fip support in nova
+ # was depricated as of 2.35. Hard coding 2.34 for danube.
+ # Carlos.Goncalves working on fixes for upstream congress that
+ # should be ready for ocata.
+ nova_micro_version="2.34"
+ #nova_micro_version=\$(nova version-list | grep CURRENT | awk '{print \$10}')
+ ds_extra_configs+="--config api_version=\$nova_micro_version"
+ fi
+ if openstack congress datasource create \$s "\$s" \$ds_configs \$ds_extra_configs; then
+ echo "INFO: Datasource: \$s created"
+ else
+ echo "WARN: Datasource: \$s could NOT be created"
+ fi
done
- openstack congress datasource create doctor "doctor"
+ if openstack congress datasource create doctor "doctor"; then
+ echo "INFO: Datasource: doctor created"
+ else
+ echo "WARN: Datsource: doctor could NOT be created"
+ fi
+fi
+
+
+# Fix project_id and os_tenant_name not in overcloudrc
+# Deprecated openstack client does not need project_id
+# and os_tenant_name anymore but glance client and
+# Rally in generall does need it.
+# REMOVE when not needed in Rally/glance-client anymore.
+if ! grep -q "OS_PROJECT_ID" ./overcloudrc;then
+ project_id=\$(openstack project list |grep admin|awk '{print \$2}')
+ echo "export OS_PROJECT_ID=\$project_id" >> ./overcloudrc
fi
+if ! grep -q "OS_TENANT_NAME" ./overcloudrc;then
+ echo "export OS_TENANT_NAME=admin" >> ./overcloudrc
+fi
+
+
EOI
- # for virtual, we NAT public network through Undercloud
- if [ "$virtual" == "TRUE" ]; then
- if ! configure_undercloud_nat ${public_network_cidr}; then
- echo -e "${red}ERROR: Unable to NAT undercloud with external net: ${public_network_cidr}${reset}"
+ # we need to restart neutron-server in Gluon deployments to allow the Gluon core
+ # plugin to correctly register itself with Neutron
+ if [ "${deploy_options_array['gluon']}" == 'True' ]; then
+ echo "Restarting neutron-server to finalize Gluon installation"
+ overcloud_connect "controller0" "sudo systemctl restart neutron-server"
+ fi
+
+ # for virtual, we NAT external network through Undercloud
+ # same goes for baremetal if only jumphost has external connectivity
+ if [ "$virtual" == "TRUE" ] || ! test_overcloud_connectivity && [ "$external_network_ipv6" != "True" ]; then
+ if [[ "$enabled_network_list" =~ "external" ]]; then
+ nat_cidr=${external_cidr}
+ else
+ nat_cidr=${admin_cidr}
+ fi
+ if ! configure_undercloud_nat ${nat_cidr}; then
+ echo -e "${red}ERROR: Unable to NAT undercloud with external net: ${nat_cidr}${reset}"
exit 1
else
- echo -e "${blue}INFO: Undercloud VM has been setup to NAT Overcloud public network${reset}"
+ echo -e "${blue}INFO: Undercloud VM has been setup to NAT Overcloud external network${reset}"
fi
fi
EOI
fi
+ ### VSPERF ###
+ if [[ "${deploy_options_array['vsperf']}" == 'True' ]]; then
+ echo "${blue}\nVSPERF enabled, running build_base_machine.sh\n${reset}"
+ overcloud_connect "compute0" "sudo sh -c 'cd /var/opt/vsperf/systems/ && ./build_base_machine.sh 2>&1 > /var/log/vsperf.log'"
+ fi
+
+ # install docker
+ if [ "${deploy_options_array['yardstick']}" == 'True' ] || [ "${deploy_options_array['dovetail']}" == 'True' ]; then
+ ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+sudo yum install docker -y
+sudo systemctl start docker
+sudo systemctl enable docker
+EOI
+ fi
+
+ # pull yardstick image
+ if [ "${deploy_options_array['yardstick']}" == 'True' ]; then
+ ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+sudo docker pull opnfv/yardstick
+EOI
+ fi
+
+ # pull dovetail image
+ if [ "${deploy_options_array['dovetail']}" == 'True' ]; then
+ ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+sudo docker pull opnfv/dovetail
+EOI
+ fi
+
# Collect deployment logs
ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
mkdir -p ~/deploy_logs
echo "---------------------------"
echo "----------END LOG----------"
echo "---------------------------"
+
+ ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
+echo "$node"
+sudo openstack-status
+EOF
fi
ssh -T ${SSH_OPTIONS[@]} "heat-admin@\$node" <<EOF
sudo rm -f /home/heat-admin/messages.log
# Print out the undercloud IP and dashboard URL
source stackrc
echo "Undercloud IP: $UNDERCLOUD, please connect by doing 'opnfv-util undercloud'"
-echo "Overcloud dashboard available at http://\$(heat output-show overcloud PublicVip | sed 's/"//g')/dashboard"
+echo "Overcloud dashboard available at http://\$(openstack stack output show overcloud PublicVip -f json | jq -r .output_value)/dashboard"
EOI
+
+if [[ "$ha_enabled" == 'True' ]]; then
+ if [ "$debug" == "TRUE" ]; then
+ echo "${blue}\nChecking pacemaker service status\n${reset}"
+ fi
+ overcloud_connect "controller0" "for i in \$(sudo pcs status | grep '^* ' | cut -d ' ' -f 2 | cut -d '_' -f 1 | uniq); do echo \"WARNING: Service: \$i not running\"; done"
+fi
+
+if [ "${deploy_options_array['vpn']}" == 'True' ]; then
+ # Check zrpcd is started
+ overcloud_connect "controller0" "sudo systemctl status zrpcd > /dev/null || echo 'WARNING: zrpcd is not running on controller0'"
+fi
}