"${virtual_nodes[@]}"
create_networks "${OPNFV_BRIDGES[@]}"
do_sysctl_cfg
+ do_udev_cfg
create_vms "${STORAGE_DIR}" "${virtual_nodes_data}" "${OPNFV_BRIDGES[@]}"
update_mcpcontrol_network
start_vms "${virtual_nodes[@]}"
salt -I 'ceilometer:server' state.sls ceilometer
salt -I 'ceilometer:agent' state.sls ceilometer
-salt -I 'horizon:server' state.sls horizon
+salt -I 'horizon:server' state.sls apache,horizon
salt -I 'nginx:server' state.sls nginx
cluster_public_host=$(salt -C 'I@nginx:server and *01*' --out=yaml \
salt -I 'ceilometer:server' state.sls ceilometer
salt -I 'ceilometer:agent' state.sls ceilometer
-salt -I 'horizon:server' state.sls horizon
+salt -I 'horizon:server' state.sls apache,horizon
salt -C 'prx*' system.reboot
wait_for 30.0 "salt -C 'prx*' test.ping"
-salt -C 'E@^(?!cfg01|mas01|kvm|cmp00).*' pkg.upgrade refresh=False
+salt -C 'E@^(?!cfg01|mas01|kvm|cmp00).*' pkg.upgrade refresh=False dist_upgrade=True
- system.barbican.server.cluster
- system.apache.server.site.barbican
- service.barbican.server.plugin.simple_crypto
+ - system.apache.server.single
- system.bind.server.single
- system.haproxy.proxy.listen.openstack.placement
- system.glusterfs.client.cluster
apache:
server:
bind:
- ~ports: ~
+ listen_default_ports: false
# sync from common-ha kvm role
glusterfs:
server:
gateway: {{ nm.net_admin_gw }}
name_servers:
- {{ nm.net_admin_gw }}
+ noifupdown: true
single:
enabled: true
type: eth
name: ${_param:pxe_admin_interface}
address: ${_param:pxe_admin_address}
netmask: ${_param:opnfv_net_admin_mask}
+ noifupdown: true
single_int:
enabled: true
type: eth
- system.nginx.server.proxy.openstack_web
- system.nginx.server.proxy.openstack.aodh
- system.nginx.server.proxy.openstack.ceilometer
+ - system.apache.server.single
- system.horizon.server.single
- system.salt.minion.cert.proxy
- system.sphinx.server.doc.reclass
vrrp_scripts:
check_pidof:
args: 'nginx'
+ apache:
+ server:
+ bind:
+ listen_default_ports: false
- python-memcache
apache:
server:
+ bind:
+ listen_default_ports: false
~modules:
- rewrite
{%- if conf.MCP_VCP %} {#- wsgi module will be enabled by a different class inherited later #}
address: ${_param:pxe_admin_address}
netmask: ${_param:opnfv_net_admin_mask}
mtu: ${_param:interface_mtu}
+ noifupdown: true
primary_interface:
enabled: true
name: ${_param:primary_interface}
- system.apache.server.site.gnocchi
- system.apache.server.site.panko
- system.apache.server.site.barbican
+ - system.apache.server.single
- system.horizon.server.single
- service.haproxy.proxy.single
- cluster.mcp-common-noha.haproxy_openstack_api
root_helper_daemon: false
apache:
server:
+ bind:
+ listen_default_ports: false
site:
gnocchi: &wsgi_threads
wsgi:
type: eth
address: ${_param:pxe_admin_address}
netmask: ${_param:opnfv_net_admin_mask}
+ noifupdown: true
single_int:
enabled: true
name: {{ nm.ctl01.nic_mgmt }}
address: ${_param:pxe_admin_address}
netmask: ${_param:opnfv_net_admin_mask}
mtu: ${_param:interface_mtu}
+ noifupdown: true
primary_interface:
enabled: true
name: ${_param:primary_interface}
gateway: {{ nm.net_admin_gw }}
name_servers:
- {{ nm.net_admin_gw }}
+ noifupdown: true
single_int:
enabled: true
name: {{ nm.ctl01.nic_mgmt }}
)
}
+function do_udev_cfg {
+ local _conf='/etc/udev/rules.d/99-opnfv-fuel-vnet-mtu.rules'
+ # http://linuxaleph.blogspot.com/2013/01/how-to-network-jumbo-frames-to-kvm-guest.html
+ echo 'SUBSYSTEM=="net", ACTION=="add", KERNEL=="vnet*", RUN+="/sbin/ip link set mtu 9000 dev '"'"%k"'"'"' |& sudo tee "${_conf}"
+ sudo udevadm control --reload || true
+}
+
function do_sysctl_cfg {
local _conf='/etc/sysctl.d/99-opnfv-fuel-bridge.conf'
# https://wiki.libvirt.org/page/Net.bridge.bridge-nf-call_and_sysctl.conf