From 7065efb23b90af91f171476360e75602a1205c13 Mon Sep 17 00:00:00 2001
From: Alexandru Avadanii <Alexandru.Avadanii@enea.com>
Date: Tue, 21 Aug 2018 22:12:48 +0200
Subject: .gitignore refresh, fold common expressions

Change-Id: I5346ee523b40f1a249394d59dbbe4d3d85c692cb
Signed-off-by: Alexandru Avadanii <Alexandru.Avadanii@enea.com>
---
 mcp/reclass/classes/cluster/.gitignore             |  10 +-
 .../classes/cluster/mcp-common-ha/infra/kvm.yml    | 160 ---------------------
 .../classes/cluster/mcp-common-ha/infra/kvm.yml.j2 | 160 +++++++++++++++++++++
 3 files changed, 162 insertions(+), 168 deletions(-)
 delete mode 100644 mcp/reclass/classes/cluster/mcp-common-ha/infra/kvm.yml
 create mode 100644 mcp/reclass/classes/cluster/mcp-common-ha/infra/kvm.yml.j2

(limited to 'mcp')

diff --git a/mcp/reclass/classes/cluster/.gitignore b/mcp/reclass/classes/cluster/.gitignore
index 733689355..9266b7cbc 100644
--- a/mcp/reclass/classes/cluster/.gitignore
+++ b/mcp/reclass/classes/cluster/.gitignore
@@ -12,15 +12,9 @@ mcp*common-*/openstack_init.yml
 mcp-common-ha/openstack_interface_vcp_biport.yml
 mcp-common-ha/openstack_interface_vcp_triport.yml
 mcp-common-ha/openstack_proxy.yml
-mcp-ovs-ha/infra/init_vcp.yml
-mcp-ovs-ha/infra/kvm.yml
-mcp-ovs-dpdk-ha/infra/init_vcp.yml
-mcp-ovs-dpdk-ha/infra/kvm.yml
-mcp-odl-ha/infra/init_vcp.yml
-mcp-odl-ha/infra/kvm.yml
+mcp-*-ha/infra/kvm.yml
+mcp-*-ha/infra/init_vcp.yml
 mcp-odl-ha/infra/maas.yml
-mcp-ovn-ha/infra/init_vcp.yml
-mcp-ovn-ha/infra/kvm.yml
 mcp-odl-ha/opendaylight/control.yml
 mcp-odl-ha/openstack/init.yml
 mcp-odl-noha/infra/config.yml
diff --git a/mcp/reclass/classes/cluster/mcp-common-ha/infra/kvm.yml b/mcp/reclass/classes/cluster/mcp-common-ha/infra/kvm.yml
deleted file mode 100644
index 62af0893a..000000000
--- a/mcp/reclass/classes/cluster/mcp-common-ha/infra/kvm.yml
+++ /dev/null
@@ -1,160 +0,0 @@
-##############################################################################
-# Copyright (c) 2018 Mirantis Inc., Enea AB and others.
-# All rights reserved. This program and the accompanying materials
-# are made available under the terms of the Apache License, Version 2.0
-# which accompanies this distribution, and is available at
-# http://www.apache.org/licenses/LICENSE-2.0
-##############################################################################
----
-classes:
-  - system.linux.system.repo.glusterfs
-  - service.keepalived.cluster.single
-  - system.glusterfs.server.volume.glance
-  - system.glusterfs.server.volume.keystone
-  - system.glusterfs.server.cluster
-  - system.salt.control.virt
-  - system.salt.control.cluster.openstack_control_cluster
-  - system.salt.control.cluster.openstack_proxy_cluster
-  - system.salt.control.cluster.openstack_database_cluster
-  - system.salt.control.cluster.openstack_message_queue_cluster
-  - system.salt.control.cluster.openstack_telemetry_cluster
-  # - system.salt.control.cluster.stacklight_server_cluster
-  # - system.salt.control.cluster.stacklight_log_cluster
-  # - system.salt.control.cluster.stacklight_telemetry_cluster
-  - cluster.mcp-common-ha.infra.kvm_pdf
-  - cluster.all-mcp-arch-common.opnfv.maas_proxy
-  - cluster.all-mcp-arch-common.opnfv.lab_proxy_pdf
-parameters:
-  _param:
-    linux_system_codename: xenial
-    glusterfs_version: '3.13'
-    cluster_vip_address: ${_param:infra_kvm_address}
-    cluster_node01_address: ${_param:infra_kvm_node01_address}
-    cluster_node02_address: ${_param:infra_kvm_node02_address}
-    cluster_node03_address: ${_param:infra_kvm_node03_address}
-    keepalived_vip_interface: br-ctl
-    keepalived_vip_virtual_router_id: 69
-  linux:
-    system:
-      kernel:
-        boot_options:
-          - spectre_v2=off
-          - nopti
-  libvirt:
-    server:
-      service: libvirtd
-      config_sys: /etc/default/libvirtd
-      unix_sock_group: libvirt
-  salt:
-    control:
-      size:  # RAM 4096,8192,16384,32768,65536
-        # Default production sizing
-        openstack.control:
-          cpu: 4
-          ram: 12288
-          disk_profile: small
-          net_profile: default
-        openstack.database:
-          cpu: 4
-          ram: 6144
-          disk_profile: large
-          net_profile: default
-        openstack.message_queue:
-          cpu: 4
-          ram: 2048
-          disk_profile: small
-          net_profile: default
-        openstack.telemetry:
-          cpu: 2
-          ram: 3072
-          disk_profile: xxlarge
-          net_profile: default
-        # stacklight.log:
-        #   cpu: 2
-        #   ram: 4096
-        #   disk_profile: xxlarge
-        #   net_profile: default
-        # stacklight.server:
-        #   cpu: 2
-        #   ram: 4096
-        #   disk_profile: small
-        #   net_profile: default
-        # stacklight.telemetry:
-        #   cpu: 2
-        #   ram: 4096
-        #   disk_profile: xxlarge
-        #   net_profile: default
-        openstack.proxy:
-          cpu: 2
-          ram: 2048
-          disk_profile: small
-          net_profile: default_ext
-      cluster:
-        internal:
-          node:
-            mdb01:
-              image: ${_param:salt_control_xenial_image}
-            mdb02:
-              image: ${_param:salt_control_xenial_image}
-            mdb03:
-              image: ${_param:salt_control_xenial_image}
-            ctl01:
-              image: ${_param:salt_control_xenial_image}
-            ctl02:
-              image: ${_param:salt_control_xenial_image}
-            ctl03:
-              image: ${_param:salt_control_xenial_image}
-            dbs01:
-              image: ${_param:salt_control_xenial_image}
-            dbs02:
-              image: ${_param:salt_control_xenial_image}
-            dbs03:
-              image: ${_param:salt_control_xenial_image}
-            msg01:
-              image: ${_param:salt_control_xenial_image}
-            msg02:
-              image: ${_param:salt_control_xenial_image}
-            msg03:
-              image: ${_param:salt_control_xenial_image}
-            prx01:
-              image: ${_param:salt_control_xenial_image}
-            prx02:
-              image: ${_param:salt_control_xenial_image}
-              provider: kvm03.${_param:cluster_domain}
-  virt:
-    nic:
-      default:
-        eth1:
-          bridge: br-mgmt
-          model: virtio
-        eth0:
-          bridge: br-ctl
-          model: virtio
-      default_ext:
-        eth2:
-          bridge: br-mgmt
-          model: virtio
-        eth1:
-          bridge: br-ex
-          model: virtio
-        eth0:
-          bridge: br-ctl
-          model: virtio
-  glusterfs:
-    server:
-      service: glusterd
-      volumes:
-        nova_instances:
-          storage: /srv/glusterfs/nova_instances
-          replica: 3
-          bricks:
-            - ${_param:cluster_node01_address}:/srv/glusterfs/nova_instances
-            - ${_param:cluster_node02_address}:/srv/glusterfs/nova_instances
-            - ${_param:cluster_node03_address}:/srv/glusterfs/nova_instances
-          options:
-            cluster.readdir-optimize: 'True'
-            nfs.disable: 'True'
-            network.remote-dio: 'True'
-            cluster.favorite-child-policy: mtime
-            diagnostics.client-log-level: WARNING
-            diagnostics.brick-log-level: WARNING
diff --git a/mcp/reclass/classes/cluster/mcp-common-ha/infra/kvm.yml.j2 b/mcp/reclass/classes/cluster/mcp-common-ha/infra/kvm.yml.j2
new file mode 100644
index 000000000..62af0893a
--- /dev/null
+++ b/mcp/reclass/classes/cluster/mcp-common-ha/infra/kvm.yml.j2
@@ -0,0 +1,160 @@
+##############################################################################
+# Copyright (c) 2018 Mirantis Inc., Enea AB and others.
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
+---
+classes:
+  - system.linux.system.repo.glusterfs
+  - service.keepalived.cluster.single
+  - system.glusterfs.server.volume.glance
+  - system.glusterfs.server.volume.keystone
+  - system.glusterfs.server.cluster
+  - system.salt.control.virt
+  - system.salt.control.cluster.openstack_control_cluster
+  - system.salt.control.cluster.openstack_proxy_cluster
+  - system.salt.control.cluster.openstack_database_cluster
+  - system.salt.control.cluster.openstack_message_queue_cluster
+  - system.salt.control.cluster.openstack_telemetry_cluster
+  # - system.salt.control.cluster.stacklight_server_cluster
+  # - system.salt.control.cluster.stacklight_log_cluster
+  # - system.salt.control.cluster.stacklight_telemetry_cluster
+  - cluster.mcp-common-ha.infra.kvm_pdf
+  - cluster.all-mcp-arch-common.opnfv.maas_proxy
+  - cluster.all-mcp-arch-common.opnfv.lab_proxy_pdf
+parameters:
+  _param:
+    linux_system_codename: xenial
+    glusterfs_version: '3.13'
+    cluster_vip_address: ${_param:infra_kvm_address}
+    cluster_node01_address: ${_param:infra_kvm_node01_address}
+    cluster_node02_address: ${_param:infra_kvm_node02_address}
+    cluster_node03_address: ${_param:infra_kvm_node03_address}
+    keepalived_vip_interface: br-ctl
+    keepalived_vip_virtual_router_id: 69
+  linux:
+    system:
+      kernel:
+        boot_options:
+          - spectre_v2=off
+          - nopti
+  libvirt:
+    server:
+      service: libvirtd
+      config_sys: /etc/default/libvirtd
+      unix_sock_group: libvirt
+  salt:
+    control:
+      size:  # RAM 4096,8192,16384,32768,65536
+        # Default production sizing
+        openstack.control:
+          cpu: 4
+          ram: 12288
+          disk_profile: small
+          net_profile: default
+        openstack.database:
+          cpu: 4
+          ram: 6144
+          disk_profile: large
+          net_profile: default
+        openstack.message_queue:
+          cpu: 4
+          ram: 2048
+          disk_profile: small
+          net_profile: default
+        openstack.telemetry:
+          cpu: 2
+          ram: 3072
+          disk_profile: xxlarge
+          net_profile: default
+        # stacklight.log:
+        #   cpu: 2
+        #   ram: 4096
+        #   disk_profile: xxlarge
+        #   net_profile: default
+        # stacklight.server:
+        #   cpu: 2
+        #   ram: 4096
+        #   disk_profile: small
+        #   net_profile: default
+        # stacklight.telemetry:
+        #   cpu: 2
+        #   ram: 4096
+        #   disk_profile: xxlarge
+        #   net_profile: default
+        openstack.proxy:
+          cpu: 2
+          ram: 2048
+          disk_profile: small
+          net_profile: default_ext
+      cluster:
+        internal:
+          node:
+            mdb01:
+              image: ${_param:salt_control_xenial_image}
+            mdb02:
+              image: ${_param:salt_control_xenial_image}
+            mdb03:
+              image: ${_param:salt_control_xenial_image}
+            ctl01:
+              image: ${_param:salt_control_xenial_image}
+            ctl02:
+              image: ${_param:salt_control_xenial_image}
+            ctl03:
+              image: ${_param:salt_control_xenial_image}
+            dbs01:
+              image: ${_param:salt_control_xenial_image}
+            dbs02:
+              image: ${_param:salt_control_xenial_image}
+            dbs03:
+              image: ${_param:salt_control_xenial_image}
+            msg01:
+              image: ${_param:salt_control_xenial_image}
+            msg02:
+              image: ${_param:salt_control_xenial_image}
+            msg03:
+              image: ${_param:salt_control_xenial_image}
+            prx01:
+              image: ${_param:salt_control_xenial_image}
+            prx02:
+              image: ${_param:salt_control_xenial_image}
+              provider: kvm03.${_param:cluster_domain}
+  virt:
+    nic:
+      default:
+        eth1:
+          bridge: br-mgmt
+          model: virtio
+        eth0:
+          bridge: br-ctl
+          model: virtio
+      default_ext:
+        eth2:
+          bridge: br-mgmt
+          model: virtio
+        eth1:
+          bridge: br-ex
+          model: virtio
+        eth0:
+          bridge: br-ctl
+          model: virtio
+  glusterfs:
+    server:
+      service: glusterd
+      volumes:
+        nova_instances:
+          storage: /srv/glusterfs/nova_instances
+          replica: 3
+          bricks:
+            - ${_param:cluster_node01_address}:/srv/glusterfs/nova_instances
+            - ${_param:cluster_node02_address}:/srv/glusterfs/nova_instances
+            - ${_param:cluster_node03_address}:/srv/glusterfs/nova_instances
+          options:
+            cluster.readdir-optimize: 'True'
+            nfs.disable: 'True'
+            network.remote-dio: 'True'
+            cluster.favorite-child-policy: mtime
+            diagnostics.client-log-level: WARNING
+            diagnostics.brick-log-level: WARNING
-- 
cgit