Update system reclass
[fuel.git] / mcp / reclass / classes / cluster / mcp-common-ha / openstack_control.yml.j2
index ceb3622..227205e 100644 (file)
@@ -7,8 +7,7 @@
 ##############################################################################
 ---
 classes:
-  - system.linux.system.repo.mcp.mirror.v1.openstack
-  - system.linux.system.repo.glusterfs
+  - system.linux.system.repo.mcp.apt_mirantis.openstack
   - system.ceilometer.client
   - system.memcached.server.single
   - system.keystone.server.cluster
@@ -20,11 +19,16 @@ classes:
   - system.heat.server.cluster
   - system.designate.server.cluster
   - system.designate.server.backend.bind
+  - system.barbican.server.cluster
+  - system.apache.server.site.barbican
+  - service.barbican.server.plugin.simple_crypto
+  - system.apache.server.single
   - system.bind.server.single
   - system.haproxy.proxy.listen.openstack.placement
   - system.glusterfs.client.cluster
   - system.glusterfs.client.volume.glance
   - system.glusterfs.client.volume.keystone
+  - cluster.mcp-common-ha.glusterfs_repo
 {%- if not conf.MCP_VCP %}
   # sync from kvm
   - service.keepalived.cluster.single
@@ -42,8 +46,8 @@ classes:
   # - system.salt.control.cluster.stacklight_log_cluster
   # - system.salt.control.cluster.stacklight_telemetry_cluster
   - cluster.mcp-common-ha.infra.kvm_pdf
-  - cluster.mcp-common-ha.include.maas_proxy
-  - cluster.mcp-common-ha.include.lab_proxy_pdf
+  - cluster.all-mcp-arch-common.opnfv.maas_proxy
+  - cluster.all-mcp-arch-common.opnfv.lab_proxy_pdf
 {%- endif %}
 parameters:
   _param:
@@ -66,10 +70,42 @@ parameters:
     cluster_node03_hostname: ${_param:openstack_control_node03_hostname}
     cluster_node03_address: ${_param:openstack_control_node03_address}
     nova_vncproxy_url: https://${_param:cluster_public_host}:6080
-    glusterfs_version: '3.13'
+    barbican_integration_enabled: 'false'
+    fernet_rotation_driver: 'shared_filesystem'
+    credential_rotation_driver: 'shared_filesystem'
+  nova:
+    controller: &db_conn_recycle_time
+      database:
+        connection_recycle_time: ${_param:db_connection_recycle_time}
+      barbican:
+        enabled: ${_param:barbican_integration_enabled}
+  cinder:
+    controller:
+      <<: *db_conn_recycle_time
+  neutron:
+    server:
+      <<: *db_conn_recycle_time
+      vlan_aware_vms: true
+      root_helper_daemon: false
+      agent_down_time: 300
+      global_physnet_mtu: ${_param:interface_mtu}
+      backend:
+        external_mtu: ${_param:interface_mtu}
+  keystone:
+    server:
+      <<: *db_conn_recycle_time
+      cacert: /etc/ssl/certs/mcp_os_cacert
+      openrc_extra:
+        volume_device_name: sdc
+  glance:
+    server:
+      <<: *db_conn_recycle_time
+      identity:
+        barbican_endpoint: ${barbican:server:host_href}
 {%- if conf.MCP_VCP %}
   heat:
     server:
+      <<: *db_conn_recycle_time
       metadata:
         host: ${_param:openstack_proxy_control_address}
         port: 8000
@@ -98,7 +134,7 @@ parameters:
   apache:
     server:
       bind:
-        ~ports: ~
+        listen_default_ports: false
   # sync from common-ha kvm role
   glusterfs:
     server:
@@ -119,20 +155,22 @@ parameters:
             diagnostics.client-log-level: WARNING
             diagnostics.brick-log-level: WARNING
 {%- endif %}
-  neutron:
-    server:
-      vlan_aware_vms: true
-      root_helper_daemon: false
-  keystone:
-    server:
-      cacert: /etc/ssl/certs/mcp_os_cacert
-      openrc_extra:
-        volume_device_name: vdc
   haproxy:
     proxy:
       listen:
         heat_cloudwatch_api:
           enabled: false
+  barbican:
+    server:
+      ks_notifications_enable: true
+      store:
+        software:
+          crypto_plugin: simple_crypto
+          store_plugin: store_crypto
+          global_default: true
+      database:
+        connection_recycle_time: ${_param:db_connection_recycle_time}
+        host: ${_param:openstack_database_address}
   bind:
     server:
       control:
@@ -148,6 +186,9 @@ parameters:
           keys:
             - designate
   designate:
+    _support:
+      sphinx:
+        enabled: False  # Workaround broken meta/sphinx.yml in salt-formula-designate
     server:
       pools:
         default: