salt_control_xenial_image: salt://salt/files/control/images/base_image_opnfv_fuel_vcp.img
# VMs spawned on Foundation Node / Jump Host net ifaces (max 4)
- opnfv_fn_vm_primary_interface: enp1s0
- opnfv_fn_vm_secondary_interface: enp2s0
- opnfv_fn_vm_tertiary_interface: enp3s0
- opnfv_fn_vm_quaternary_interface: enp4s0
-
- # VCP VMs spawned on KVM Hosts net ifaces (max 3)
- # NOTE(armband): Only x86 VCP VMs spawned via salt.control names differ
- opnfv_vcp_vm_primary_interface: ${_param:opnfv_fn_vm_primary_interface}
- opnfv_vcp_vm_secondary_interface: ${_param:opnfv_fn_vm_secondary_interface}
- opnfv_vcp_vm_tertiary_interface: ${_param:opnfv_fn_vm_tertiary_interface}
+ opnfv_vm_primary_interface: enp1s0
+ opnfv_vm_secondary_interface: enp2s0
+ opnfv_vm_tertiary_interface: enp3s0
+ opnfv_vm_quaternary_interface: enp4s0
# MaaS timeouts
opnfv_maas_timeout_comissioning: 10
salt_control_xenial_image: salt://salt/files/control/images/base_image_opnfv_fuel_vcp.img
# VMs spawned on Foundation Node / Jump Host net ifaces (max 4)
- opnfv_fn_vm_primary_interface: ens3
- opnfv_fn_vm_secondary_interface: ens4
- opnfv_fn_vm_tertiary_interface: ens5
- opnfv_fn_vm_quaternary_interface: ens6
-
- # VCP VMs spawned on KVM Hosts net ifaces (max 3)
- # NOTE(armband): Only x86 VCP VMs spawned via salt.control names differ
- opnfv_vcp_vm_primary_interface: ens2
- opnfv_vcp_vm_secondary_interface: ens3
- opnfv_vcp_vm_tertiary_interface: ens4
+ opnfv_vm_primary_interface: ens2
+ opnfv_vm_secondary_interface: ens3
+ opnfv_vm_tertiary_interface: ens4
+ opnfv_vm_quaternary_interface: ens5
# MaaS timeouts
opnfv_maas_timeout_comissioning: 10
salt_master_host: ${_param:infra_config_deploy_address}
# yamllint disable rule:line-length
salt_api_password_hash: "$6$sGnRlxGf$al5jMCetLP.vfI/fTl3Z0N7Za1aeiexL487jAtyRABVfT3NlwZxQGVhO7S1N8OwS/34VHYwZQA8lkXwKMN/GS1"
- dhcp_nic: ${_param:opnfv_fn_vm_primary_interface}
- single_nic: ${_param:opnfv_fn_vm_secondary_interface}
- pxe_nic: ${_param:opnfv_fn_vm_tertiary_interface}
+ dhcp_nic: ${_param:opnfv_vm_primary_interface}
+ single_nic: ${_param:opnfv_vm_secondary_interface}
+ pxe_nic: ${_param:opnfv_vm_tertiary_interface}
linux:
network:
interface:
- cluster.baremetal-mcp-pike-common-ha.infra.lab_proxy_pdf
parameters:
_param:
- dhcp_interface: ${_param:opnfv_fn_vm_primary_interface}
- primary_interface: ${_param:opnfv_fn_vm_secondary_interface}
- pxe_interface: ${_param:opnfv_fn_vm_tertiary_interface}
+ dhcp_interface: ${_param:opnfv_vm_primary_interface}
+ primary_interface: ${_param:opnfv_vm_secondary_interface}
+ pxe_interface: ${_param:opnfv_vm_tertiary_interface}
interface_mtu: 1500
# MaaS has issues using MTU > 1500 for PXE interface
pxe_interface_mtu: 1500
- cluster.baremetal-mcp-pike-common-ha.include.proxy
parameters:
_param:
- dhcp_nic: ${_param:opnfv_vcp_vm_primary_interface}
- single_nic: ${_param:opnfv_vcp_vm_secondary_interface}
+ dhcp_nic: ${_param:opnfv_vm_primary_interface}
+ single_nic: ${_param:opnfv_vm_secondary_interface}
linux:
network:
interface:
- cluster.baremetal-mcp-pike-common-ha.include.proxy
parameters:
_param:
- dhcp_nic: ${_param:opnfv_vcp_vm_primary_interface}
- single_nic: ${_param:opnfv_vcp_vm_secondary_interface}
- control_nic: ${_param:opnfv_vcp_vm_tertiary_interface}
+ dhcp_nic: ${_param:opnfv_vm_primary_interface}
+ single_nic: ${_param:opnfv_vm_secondary_interface}
+ control_nic: ${_param:opnfv_vm_tertiary_interface}
linux:
network:
interface:
interface:
single_int:
enabled: true
- name: ${_param:opnfv_fn_vm_secondary_interface}
+ name: ${_param:opnfv_vm_secondary_interface}
type: eth
proto: static
address: ${_param:single_address}
- system.ceilometer.client.cinder_volume
parameters:
_param:
- primary_interface: ${_param:opnfv_fn_vm_secondary_interface}
- tenant_interface: ${_param:opnfv_fn_vm_tertiary_interface}
- external_interface: ${_param:opnfv_fn_vm_quaternary_interface}
+ primary_interface: ${_param:opnfv_vm_secondary_interface}
+ tenant_interface: ${_param:opnfv_vm_tertiary_interface}
+ external_interface: ${_param:opnfv_vm_quaternary_interface}
interface_mtu: 9000
linux_system_codename: xenial
loopback_device_size: 20
interface:
dhcp_int:
enabled: true
- name: ${_param:opnfv_fn_vm_primary_interface}
+ name: ${_param:opnfv_vm_primary_interface}
proto: dhcp
type: eth
mtu: ${_param:interface_mtu}
interface:
single_int:
enabled: true
- name: ${_param:opnfv_fn_vm_secondary_interface}
+ name: ${_param:opnfv_vm_secondary_interface}
type: eth
proto: static
address: ${_param:single_address}
netmask: 255.255.255.0
public_int:
enabled: true
- name: ${_param:opnfv_fn_vm_quaternary_interface}
+ name: ${_param:opnfv_vm_quaternary_interface}
type: eth
proto: static
address: ${_param:cluster_public_host}
---
parameters:
_param:
- primary_interface: ${_param:opnfv_fn_vm_secondary_interface}
- tenant_interface: ${_param:opnfv_fn_vm_tertiary_interface}
- external_interface: ${_param:opnfv_fn_vm_quaternary_interface}
+ primary_interface: ${_param:opnfv_vm_secondary_interface}
+ tenant_interface: ${_param:opnfv_vm_tertiary_interface}
+ external_interface: ${_param:opnfv_vm_quaternary_interface}
interface_mtu: 9000
linux_system_codename: xenial
neutron:
interface:
dhcp_int:
enabled: true
- name: ${_param:opnfv_fn_vm_primary_interface}
+ name: ${_param:opnfv_vm_primary_interface}
proto: dhcp
type: eth
mtu: ${_param:interface_mtu}
interface:
single_int:
enabled: true
- name: ${_param:opnfv_fn_vm_secondary_interface}
+ name: ${_param:opnfv_vm_secondary_interface}
type: eth
proto: static
address: ${_param:single_address}
node:
openstack_compute_node01:
params:
- dpdk0_name: ${_param:opnfv_fn_vm_tertiary_interface}
+ dpdk0_name: ${_param:opnfv_vm_tertiary_interface}
dpdk0_pci: '"0000:00:05.0"'
openstack_compute_node02:
params:
- dpdk0_name: ${_param:opnfv_fn_vm_tertiary_interface}
+ dpdk0_name: ${_param:opnfv_vm_tertiary_interface}
dpdk0_pci: '"0000:00:05.0"'
openstack_gateway_node01:
params:
IFS='|' read -r -a vnodes <<< "$1"; shift
local vnode_networks=("$@")
- # AArch64: prepare arch specific arguments
- local virt_extra_args=""
- if [ "$(uname -i)" = "aarch64" ]; then
- # No Cirrus VGA on AArch64, use virtio instead
- virt_extra_args="$virt_extra_args --video=virtio"
- fi
-
# create vms with specified options
for serialized_vnode_data in "${vnodes[@]}"; do
IFS=',' read -r -a vnode_data <<< "${serialized_vnode_data}"
--cpu host-passthrough --accelerate ${net_args} \
--disk path="${image_dir}/mcp_${vnode_data[0]}.qcow2",format=qcow2,bus=virtio,cache=none,io=native \
--os-type linux --os-variant none \
- --boot hd --vnc --console pty --autostart --noreboot \
+ --boot hd --nographics --console pty --autostart --noreboot \
--disk path="${image_dir}/mcp_${vnode_data[0]}.iso",device=cdrom \
- --noautoconsole \
- ${virt_extra_args}
+ --noautoconsole
done
}