#PmdCoreList: 1
#OvsDpdkCoreList: 2
#OvsDpdkSocketMemory: 1024
+ #ControllerExtraConfig:
+ #NovaComputeExtraConfig:
ExtraConfig:
tripleo::ringbuilder::build_ring: False
nova::nova_public_key:
key: 'os_compute_api:servers:show:host_status'
value: 'rule:admin_or_owner'
nova::api::default_floating_pool: 'external'
- #VPP routing node, used for odl_l3-fdio only.
+ #VPP routing node, used for odl-fdio only.
#opendaylight::vpp_routing_node: overcloud-novacompute-0.opnfvlf.org
ControllerServices:
- OS::TripleO::Services::CACerts
- OS::TripleO::Services::Congress
- OS::TripleO::Services::NeutronVppAgent
- OS::TripleO::Services::OVNDBs
+ - OS::TripleO::Services::Vpp
ComputeServices:
- OS::TripleO::Services::CACerts
- OS::TripleO::Services::CephClient
- OS::TripleO::Services::VipHosts
- OS::TripleO::Services::NeutronHoneycombAgent
- OS::TripleO::Services::NeutronVppAgent
+ - OS::TripleO::Services::Vpp
packages and configuration for an OPNFV deployment to execute. Once the disk
images have been written to node's disks the nodes will boot locally and
execute cloud-init which will execute the final node configuration. This
-configuration is largly completed by executing a puppet apply on each node.
+configuration is largely completed by executing a puppet apply on each node.
Installation High-Level Overview - VM Deployment
================================================
control node and 1 or more compute nodes for a Non-HA Deployment) will be
defined for the target OPNFV deployment. The part of the toolchain that
executes IPMI power instructions calls into libvirt instead of the IPMI
-interfaces on baremetal servers to operate the power managment. These VMs are
+interfaces on baremetal servers to operate the power management. These VMs are
then provisioned with the same disk images and configuration that baremetal
would be.
install the opnfv-apex RPMs using the OPNFV artifacts yum repo. This yum
repo is created at release. It will not exist before release day.
- ``sudo yum install http://artifacts.opnfv.org/apex/danube/opnfv-apex-release-danube.noarch.rpm``
+ ``sudo yum install http://artifacts.opnfv.org/apex/euphrates/opnfv-apex-release-euphrates.noarch.rpm``
Once you have installed the repo definitions for Apex, RDO and EPEL then
yum install Apex:
(``/etc/opnfv-apex/``). These files are named with the naming convention
os-sdn_controller-enabled_feature-[no]ha.yaml. These files can be used in
place of the (``/etc/opnfv-apex/deploy_settings.yaml``) file if one suites
- your deployment needs. If a pre-built deploy_settings file is choosen there
+ your deployment needs. If a pre-built deploy_settings file is chosen there
is no need to customize (``/etc/opnfv-apex/deploy_settings.yaml``). The
pre-built file can be used in place of the
(``/etc/opnfv-apex/deploy_settings.yaml``) file.
EOI
fi
- # Configure routing node for odl_l3-fdio
+ # Configure routing node for odl-fdio
if [[ "${deploy_options_array['sdn_l3']}" == 'True' ]]; then
ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
sed -i "/opendaylight::vpp_routing_node:/c\ opendaylight::vpp_routing_node: ${deploy_options_array['odl_vpp_routing_node']}.${domain_name}" ${ENV_FILE}
if [ -n "${deploy_options_array['performance']}" ]; then
ovs_dpdk_perf_flag="False"
for option in "${performance_options[@]}" ; do
+ if [ "${arr[1]}" == "vpp" ]; then
+ if [ "${arr[0]}" == "Compute" ]; then
+ role='NovaCompute'
+ else
+ role=${arr[0]}
+ fi
+ if [ "${arr[2]}" == "main-core" ]; then
+ ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+ sed -i "/${role}ExtraConfig:/ c\ ${role}ExtraConfig:\n fdio::vpp_cpu_main_core: \"'${arr[3]}'\"" ${ENV_FILE}
+EOI
+ elif [ "${arr[2]}" == "corelist-workers" ]; then
+ ssh -T ${SSH_OPTIONS[@]} "stack@$UNDERCLOUD" <<EOI
+ sed -i "/${role}ExtraConfig:/ c\ ${role}ExtraConfig:\n fdio::vpp_cpu_corelist_workers: \"'${arr[3]}'\"" ${ENV_FILE}
+EOI
+ fi
+ fi
arr=($option)
# use compute's kernel settings for all nodes for now.
if [ "${arr[0]}" == "Compute" ] && [ "${arr[1]}" == "kernel" ]; then