aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rwxr-xr-xci/deploy.sh10
-rw-r--r--docs/release/installation/installation.instruction.rst136
-rwxr-xr-xmcp/config/states/baremetal_init1
-rwxr-xr-xmcp/config/states/maas9
-rwxr-xr-xmcp/config/states/networking_gw2
-rwxr-xr-xmcp/config/states/openstack_ha2
-rwxr-xr-xmcp/config/states/openstack_noha5
-rwxr-xr-xmcp/config/states/virtual_control_plane3
-rw-r--r--mcp/patches/0009-maas.py-Add-missing-status-ids.patch38
-rw-r--r--mcp/patches/0011-system.repo-Debian-Add-keyserver-proxy-support.patch118
-rw-r--r--mcp/patches/patches.list2
-rw-r--r--mcp/patches/scripts/0002-salt-master-setup-Group-APT-install-formulas.patch68
-rw-r--r--mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/include/proxy.yml9
-rw-r--r--mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/config.yml1
-rw-r--r--mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/init.yml2
-rw-r--r--mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/lab_proxy_pdf.yml.j224
-rw-r--r--mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/maas.yml1
-rw-r--r--mcp/reclass/classes/cluster/baremetal-mcp-ocata-odl-ha/openstack/compute_pdf.yml.j26
-rw-r--r--mcp/reclass/classes/cluster/baremetal-mcp-ocata-ovs-ha/openstack/compute_pdf.yml.j26
-rw-r--r--mcp/reclass/classes/cluster/virtual-mcp-ocata-common-noha/openstack_init.yml2
-rw-r--r--mcp/salt-formulas/maas/machines/delete.sls1
-rw-r--r--mcp/salt-formulas/maas/machines/mark_broken_fixed.sls1
-rw-r--r--mcp/salt-formulas/maas/machines/override_failed_testing.sls20
-rw-r--r--mcp/salt-formulas/opendaylight/server.sls16
-rw-r--r--mcp/scripts/lib.sh14
-rw-r--r--mcp/scripts/requirements_rpm.yaml2
-rwxr-xr-xmcp/scripts/salt.sh8
27 files changed, 442 insertions, 65 deletions
diff --git a/ci/deploy.sh b/ci/deploy.sh
index 04747ba..f1a4144 100755
--- a/ci/deploy.sh
+++ b/ci/deploy.sh
@@ -321,13 +321,8 @@ fi
# Check scenario file existence
SCENARIO_DIR="../config/scenario"
if [ ! -f "${SCENARIO_DIR}/${DEPLOY_TYPE}/${DEPLOY_SCENARIO}.yaml" ]; then
- notify "[WARN] ${DEPLOY_SCENARIO}.yaml not found!\n" 3
- notify "[WARN] Setting simplest scenario (os-nosdn-nofeature-noha)\n" 3
- DEPLOY_SCENARIO='os-nosdn-nofeature-noha'
- if [ ! -f "${SCENARIO_DIR}/${DEPLOY_TYPE}/${DEPLOY_SCENARIO}.yaml" ]; then
- notify "[ERROR] Scenario definition file is missing!\n" 1>&2
- exit 1
- fi
+ notify "[ERROR] Scenario definition file is missing!\n" 1>&2
+ exit 1
fi
# Check defaults file existence
@@ -428,6 +423,7 @@ else
prepare_vms "${base_image}" "${STORAGE_DIR}" "${virtual_repos_pkgs}" \
"${virtual_nodes[@]}"
create_networks "${OPNFV_BRIDGES[@]}"
+ do_sysctl_cfg
create_vms "${STORAGE_DIR}" "${virtual_nodes_data}" "${OPNFV_BRIDGES[@]}"
update_mcpcontrol_network
start_vms "${virtual_nodes[@]}"
diff --git a/docs/release/installation/installation.instruction.rst b/docs/release/installation/installation.instruction.rst
index 53ef809..355a2e0 100644
--- a/docs/release/installation/installation.instruction.rst
+++ b/docs/release/installation/installation.instruction.rst
@@ -132,14 +132,14 @@ installation of Euphrates using Fuel:
**NOTE:** For aarch64 deployments an UEFI compatible firmware with PXE support is needed (e.g. EDK2).
-
===============================
Help with Hardware Requirements
===============================
Calculate hardware requirements:
-For information on compatible hardware types available for use, please see `Fuel OpenStack Hardware Compatibility List <https://www.mirantis.com/software/hardware-compatibility/>`_.
+For information on compatible hardware types available for use,
+please see `Fuel OpenStack Hardware Compatibility List <https://www.mirantis.com/software/hardware-compatibility/>`_
When choosing the hardware on which you will deploy your OpenStack
environment, you should think about:
@@ -183,7 +183,49 @@ OPNFV Software Prerequisites
The Jumpserver node should be pre-provisioned with an operating system,
according to the Pharos specification. Relevant network bridges should
-also be pre-configured (e.g. admin, management, public).
+also be pre-configured (e.g. admin_br, mgmt_br, public_br).
+
+ - The admin bridge (admin_br) is mandatory for the baremetal nodes PXE booting during fuel installation.
+ - The management bridge (mgmt_br) is required for testing suites (e.g. functest/yardstick), it is
+ suggested to pre-configure it for debugging purposes.
+ - The public bridge (public_br) is also nice to have for debugging purposes, but not mandatory.
+
+The user running the deploy script on the Jumpserver should belong to "sudo" and "libvirt" groups,
+and have passwordless sudo access.
+
+The following example adds the groups to the user "jenkins"
+
+.. code-block:: bash
+
+ $ sudo usermod -aG sudo jenkins
+ $ sudo usermod -aG libvirt jenkins
+ $ reboot
+ $ groups
+ jenkins sudo libvirt
+
+ $ sudo visudo
+ ...
+ %jenkins ALL=(ALL) NOPASSWD:ALL
+
+For an AArch64 Jumpserver, the "libvirt" minimum required version is 3.x, 3.5 or newer highly recommended.
+While not mandatory, upgrading the kernel and QEMU on the Jumpserver is also highly recommended
+(especially on AArch64 Jumpservers).
+
+For CentOS 7.4 (AArch64), distro provided packages are already new enough.
+For Ubuntu 16.04 (arm64), distro packages are too old and 3rd party repositories should be used.
+For convenience, Armband provides a DEB repository holding all the required packages.
+
+To add and enable the Armband repository on an Ubuntu 16.04 system,
+create a new sources list file `/apt/sources.list.d/armband.list` with the following contents:
+
+.. code-block:: bash
+
+ $ cat /etc/apt/sources.list.d/armband.list
+ //for OpenStack ocata release
+ deb http://linux.enea.com/mcp-repos/ocata/xenial ocata main
+ deb http://linux.enea.com/apt-mk/xenial nightly ocata
+
+ $ apt-get update
Fuel@OPNFV has been validated by CI using the following distributions
installed on the Jumpserver:
@@ -191,14 +233,23 @@ installed on the Jumpserver:
- CentOS 7 (recommended by Pharos specification);
- Ubuntu Xenial;
-**NOTE:** The install script will automatically install all required distro
-package dependencies on the Jumpserver, unless explicitly asked not to
-(via `-P` deploy arg). This includes Python, QEMU, libvirt etc.
+**NOTE**: The install script expects 'libvirt' to be already running on the Jumpserver.In case libvirt
+packages are missing, the script will install them; but depending on the OS distribution, the user
+might have to start the 'libvirtd' service manually, then run the deploy script again. Therefore, it
+is recommened to install libvirt-bin explicitly on the Jumpserver before the deployment.
+
+**NOTE**: It is also recommened to install the newer kernel on the Jumpserver before the deployment.
+
+**NOTE**: The install script will automatically install the rest of required distro package
+dependencies on the Jumpserver, unless explicitly asked not to (via -P deploy arg). This includes
+Python, QEMU, libvirt etc.
+
+**NOTE**: The install script will alter Jumpserver sysconf and disable `net.bridge.bridge-nf-call`.
+
+.. code-block:: bash
+
+ $ apt-get install linux-image-generic-hwe-16.04-edge libvirt-bin
-**NOTE:** The install script expects 'libvirt' to be already running on the
-Jumpserver. In case libvirt packages are missing, the script will install
-them; but depending on the OS distribution, the user might have to start the
-'libvirtd' service manually, then run the deploy script again.
==========================================
OPNFV Software Installation and Deployment
@@ -213,9 +264,9 @@ automatic based on deployment scenario.
The reclass model covers:
- Infrastucture node definition: Salt Master node (cfg01) and MaaS node (mas01)
- - Openstack node defition: Controler nodes (ctl01, ctl02, ctl03) and Compute nodes (cmp001, cmp002)
+ - OpenStack node definition: Controller nodes (ctl01, ctl02, ctl03) and Compute nodes (cmp001, cmp002)
- Infrastructure components to install (software packages, services etc.)
- - Openstack components and services (rabbitmq, galera etc.), as well as all configuration for them
+ - OpenStack components and services (rabbitmq, galera etc.), as well as all configuration for them
Automatic Installation of a Virtual POD
@@ -224,9 +275,9 @@ Automatic Installation of a Virtual POD
For virtual deploys all the targets are VMs on the Jumpserver. The deploy script will:
- Create a Salt Master VM on the Jumpserver which will drive the installation
- - Create the bridges for networking with virsh (only if a real bridge does not already exists for a given network)
- - Install Openstack on the targets
- - Leverage Salt to install & configure Openstack services
+ - Create the bridges for networking with virsh (only if a real bridge does not already exist for a given network)
+ - Install OpenStack on the targets
+ - Leverage Salt to install & configure OpenStack services
.. figure:: img/fuel_virtual.png
:align: center
@@ -249,18 +300,18 @@ For virtual deploys all the targets are VMs on the Jumpserver. The deploy script
In this figure there are examples of two virtual deploys:
- Jumphost 1 has only virsh bridges, created by the deploy script
- - Jumphost 2 has a mix of linux and virsh briges; when linux bridge exist for a specified network,
+ - Jumphost 2 has a mix of Linux and virsh bridges; When Linux bridge exists for a specified network,
the deploy script will skip creating a virsh bridge for it
-**Note**: A virtual network "mcpcontrol" is always created. For virtual deploys, "mcpcontrol" is also used
-for Admin, leaving the PXE/Admin bridge unused.
+**Note**: A virtual network "mcpcontrol" is always created. For virtual deploys, "mcpcontrol" is also
+ used for Admin, leaving the PXE/Admin bridge unused.
Automatic Installation of a Baremetal POD
=========================================
The baremetal installation process can be done by editing the information about
-hardware and enviroment in the reclass files, or by using a Pod Descriptor File (PDF).
+hardware and environment in the reclass files, or by using a Pod Descriptor File (PDF).
This file contains all the information about the hardware and network of the deployment
the will be fed to the reclass model during deployment.
@@ -268,10 +319,10 @@ The installation is done automatically with the deploy script, which will:
- Create a Salt Master VM on the Jumpserver which will drive the installation
- Create a MaaS Node VM on the Jumpserver which will provision the targets
- - Install Openstack on the targets
+ - Install OpenStack on the targets
- Leverage MaaS to provision baremetal nodes with the operating system
- - Leverage Salt to configure the operatign system on the baremetal nodes
- - Leverage Salt to install & configure Openstack services
+ - Leverage Salt to configure the operating system on the baremetal nodes
+ - Leverage Salt to install & configure OpenStack services
.. figure:: img/fuel_baremetal.png
:align: center
@@ -301,11 +352,12 @@ The installation is done automatically with the deploy script, which will:
| Tenant VM | VM running in the cloud |
+-----------------------+---------------------------------------------------------+
-In the baremetal deploy all bridges but "mcpcontrol" are linux bridges. For the Jumpserver, if they are already created
-they will be used; otherwise they will be created. For the targets, the bridges are created by the deploy script.
+In the baremetal deploy all bridges but "mcpcontrol" are Linux bridges. For the Jumpserver, it is
+required to pre-configure at least the admin_br bridge for the PXE/Admin.
+For the targets, the bridges are created by the deploy script.
-**Note**: A virtual network "mcpcontrol" is always created. For baremetal deploys, PXE bridge is used for
-baremetal node provisioning, while "mcpcontrol" is used to provision the infrastructure VMs only.
+**Note**: A virtual network "mcpcontrol" is always created. For baremetal deploys, PXE bridge is used
+for baremetal node provisioning, while "mcpcontrol" is used to provision the infrastructure VMs only.
Steps to Start the Automatic Deploy
@@ -337,13 +389,22 @@ These steps are common both for virtual and baremetal deploys.
#. Start the deploy script
+ Besides the basic options, there are other recommended deploy arguments:
+
+ - use **-D** option to enable the debug info
+ - use **-S** option to point to a tmp dir where the disk images are saved. The images will be
+ re-used between deploys
+ - use **|& tee** to save the deploy log to a file
+
.. code-block:: bash
$ ci/deploy.sh -l <lab_name> \
-p <pod_name> \
-b <URI to configuration repo containing the PDF file> \
-s <scenario> \
- -B <list of admin, management, private and public bridges>
+ -B <list of admin, management, private and public bridges> \
+ -D \
+ -S <Storage directory for disk images> |& tee deploy.log
Examples
--------
@@ -360,7 +421,9 @@ Examples
$ ci/deploy.sh -b file:///home/jenkins/tmpdir/securedlab \
-l ericsson \
-p virtual_kvm \
- -s os-nosdn-nofeature-noha
+ -s os-nosdn-nofeature-noha \
+ -D \
+ -S /home/jenkins/tmpdir |& tee deploy.log
Once the deployment is complete, the OpenStack Dashboard, Horizon is
available at http://<controller VIP>:8078, e.g. http://10.16.0.101:8078.
@@ -377,6 +440,8 @@ Examples
-p pod2 \
-s os-nosdn-nofeature-ha \
-B pxebr,br-ctl
+ -D \
+ -S /home/jenkins/tmpdir |& tee deploy.log
.. figure:: img/lf_pod2.png
:align: center
@@ -391,11 +456,12 @@ Examples
.. code-block:: bash
- $ ci/deploy.sh -b file:///home/jenkins/tmpdir/securedlab \
- -l arm \
- -p pod5 \
- -s os-nosdn-nofeature-ha \
- -B admin7_br0,mgmt7_br0,,public7_br0
+ $ ci/deploy.sh -b file:///home/jenkins/tmpdir/securedlab \
+ -l arm \
+ -p pod5 \
+ -s os-nosdn-nofeature-ha \
+ -D \
+ -S /home/jenkins/tmpdir |& tee deploy.log
.. figure:: img/arm_pod5.png
:align: center
@@ -403,10 +469,6 @@ Examples
Fuel@OPNFV ARM POD5 Network Layout
- Once the deployment is complete, the SaltStack Deployment Documentation is
- available at http://<Proxy VIP>:8090, e.g. http://10.0.8.103:8090.
-
-
Pod Descriptor Files
====================
diff --git a/mcp/config/states/baremetal_init b/mcp/config/states/baremetal_init
index 1d38201..5a8b6c7 100755
--- a/mcp/config/states/baremetal_init
+++ b/mcp/config/states/baremetal_init
@@ -25,7 +25,6 @@ salt -C 'kvm* or cmp*' file.replace $debian_ip_source \
salt -C 'kvm*' pkg.install bridge-utils
salt -C 'kvm*' state.apply linux.network
salt -C 'kvm* or cmp*' state.apply salt.minion
-salt -C 'kvm* or cmp*' service.force_reload salt-minion
salt -C 'cmp*' state.apply linux.system
salt -C 'cmp*' state.apply linux.network || true
diff --git a/mcp/config/states/maas b/mcp/config/states/maas
index 39f6bad..24249d4 100755
--- a/mcp/config/states/maas
+++ b/mcp/config/states/maas
@@ -27,12 +27,19 @@ function maas_fixup() {
local fcnodes=$(echo "${statusout}" | \
grep -Pzo 'status: Failed commissioning\n\s+system_id: \K.+\n')
+ local ftnodes=$(echo "${statusout}" | \
+ grep -Pzo 'status: Failed testing\n\s+system_id: \K.+\n')
for node_system_id in ${fcnodes}; do
salt -C 'mas01*' state.apply maas.machines.delete \
pillar="{'system_id': '${node_system_id}'}"
sleep 10
done
- if [ -n "${fcnodes}" ]; then
+ for node_system_id in ${ftnodes}; do
+ salt -C 'mas01*' state.apply maas.machines.override_failed_testing \
+ pillar="{'system_id': '${node_system_id}'}"
+ sleep 10
+ done
+ if [ -n "${fcnodes}" ] || [ -n "${ftnodes}" ]; then
salt -C 'mas01*' state.apply maas.machines
return 1
fi
diff --git a/mcp/config/states/networking_gw b/mcp/config/states/networking_gw
index 70fb67d..6fa7111 100755
--- a/mcp/config/states/networking_gw
+++ b/mcp/config/states/networking_gw
@@ -9,4 +9,4 @@
CI_DEBUG=${CI_DEBUG:-0}; [[ "${CI_DEBUG}" =~ (false|0) ]] || set -x
-salt -I 'neutron:gateway' service.restart networking
+salt -I 'neutron:gateway' cmd.run 'ifup --force --ignore-errors float-to-ex'
diff --git a/mcp/config/states/openstack_ha b/mcp/config/states/openstack_ha
index fbddc6e..02b7009 100755
--- a/mcp/config/states/openstack_ha
+++ b/mcp/config/states/openstack_ha
@@ -46,7 +46,7 @@ salt -I 'nova:controller' state.sls nova -b 1
salt -I 'heat:server' state.sls heat -b 1
wait_for 5 "salt -I 'cinder:controller' state.sls cinder -b 1"
-salt -I 'cinder:volume' state.sls cinder
+wait_for 3 "salt -I 'cinder:volume' state.sls cinder"
salt -I 'neutron:server' state.sls neutron -b 1
salt -I 'neutron:gateway' state.sls neutron.gateway
diff --git a/mcp/config/states/openstack_noha b/mcp/config/states/openstack_noha
index 369e165..cc46ac1 100755
--- a/mcp/config/states/openstack_noha
+++ b/mcp/config/states/openstack_noha
@@ -9,6 +9,9 @@
CI_DEBUG=${CI_DEBUG:-0}; [[ "${CI_DEBUG}" =~ (false|0) ]] || set -x
+# shellcheck disable=SC1090
+source "$(dirname "${BASH_SOURCE[0]}")/../../scripts/lib.sh"
+
salt -I 'nfs:server' file.mkdir /srv/nova/instances
salt -I 'nfs:server' state.sls nfs
salt -I 'nfs:client' state.sls nfs
@@ -34,7 +37,7 @@ salt -I 'nova:controller' state.sls nova
salt -I 'heat:server' state.sls heat
salt -I 'cinder:controller' state.sls cinder
-salt -I 'cinder:volume' state.sls cinder
+wait_for 3 "salt -I 'cinder:volume' state.sls cinder"
salt -I 'neutron:server' state.sls neutron
salt -I 'neutron:gateway' state.sls neutron
diff --git a/mcp/config/states/virtual_control_plane b/mcp/config/states/virtual_control_plane
index 33cc9dc..f97e9b7 100755
--- a/mcp/config/states/virtual_control_plane
+++ b/mcp/config/states/virtual_control_plane
@@ -59,10 +59,9 @@ salt -C 'E@^(?!cfg01|mas01|kvm|cmp00).*' cp.get_file \
"salt://maas/files/$(basename "${APT_CONF_D_CURTIN}")" "${APT_CONF_D_CURTIN}"
wait_for 10 "salt -C 'E@^(?!cfg01|mas01|kvm|cmp00).*' state.apply salt"
-wait_for 10 "salt -C 'E@^(?!cfg01|mas01|kvm|cmp00).*' service.force_reload salt-minion"
wait_for 10.0 "salt -C 'E@^(?!cfg01|mas01|kvm|cmp00).*' state.apply linux,ntp"
-wait_for 10 "salt -C 'E@^(?!cfg01|mas01|kvm|cmp00).*' ssh.set_auth_key ${SUDO_USER} \
+wait_for 10.0 "salt -C 'E@^(?!cfg01|mas01|kvm|cmp00).*' ssh.set_auth_key ${SUDO_USER} \
$(awk 'NR==1{print $2}' "$(eval echo "~${SUDO_USER}/.ssh/authorized_keys")")"
# Disable proxy dhcp routes after installation
diff --git a/mcp/patches/0009-maas.py-Add-missing-status-ids.patch b/mcp/patches/0009-maas.py-Add-missing-status-ids.patch
new file mode 100644
index 0000000..648e62e
--- /dev/null
+++ b/mcp/patches/0009-maas.py-Add-missing-status-ids.patch
@@ -0,0 +1,38 @@
+::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::
+: Copyright (c) 2017 Mirantis Inc., Enea AB and others.
+:
+: All rights reserved. This program and the accompanying materials
+: are made available under the terms of the Apache License, Version 2.0
+: which accompanies this distribution, and is available at
+: http://www.apache.org/licenses/LICENSE-2.0
+::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::
+From: Alexandru Avadanii <Alexandru.Avadanii@enea.com>
+Date: Sun, 17 Dec 2017 06:30:27 +0100
+Subject: [PATCH] maas.py: Add missing status ids
+
+Status ID list synced from MaaS 2.3 [1].
+
+[1] https://github.com/maas/maas/blob/2.3/src/maasserver/\
+ migrations/builtin/maasserver/0106_testing_status.py#L25
+
+Signed-off-by: Alexandru Avadanii <Alexandru.Avadanii@enea.com>
+---
+ _modules/maas.py | 5 ++++-
+ 1 file changed, 4 insertions(+), 1 deletion(-)
+
+diff --git a/_modules/maas.py b/_modules/maas.py
+index e279b70..1df4698 100644
+--- a/_modules/maas.py
++++ b/_modules/maas.py
+@@ -52,7 +52,10 @@ def __virtual__():
+ (9, 'Deploying'), (6, 'Deployed'), (7, 'Retired'), (8, 'Broken'),
+ (11, 'Failed deployment'), (12, 'Releasing'),
+ (13, 'Releasing failed'), (14, 'Disk erasing'),
+- (15, 'Failed disk erasing')])
++ (15, 'Failed disk erasing'), (16, 'Rescue mode'),
++ (17, 'Entering rescue mode'), (18, 'Failed to enter rescue mode'),
++ (19, 'Exiting rescue mode'), (20, 'Failed to exit rescue mode'),
++ (21, 'Testing'), (22, 'Failed testing')])
+
+
+ def _format_data(data):
diff --git a/mcp/patches/0011-system.repo-Debian-Add-keyserver-proxy-support.patch b/mcp/patches/0011-system.repo-Debian-Add-keyserver-proxy-support.patch
new file mode 100644
index 0000000..fdf84dc
--- /dev/null
+++ b/mcp/patches/0011-system.repo-Debian-Add-keyserver-proxy-support.patch
@@ -0,0 +1,118 @@
+::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::
+: Copyright (c) 2017 Mirantis Inc., Enea AB and others.
+:
+: All rights reserved. This program and the accompanying materials
+: are made available under the terms of the Apache License, Version 2.0
+: which accompanies this distribution, and is available at
+: http://www.apache.org/licenses/LICENSE-2.0
+::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::
+From: Alexandru Avadanii <Alexandru.Avadanii@enea.com>
+Date: Mon, 22 Jan 2018 00:28:09 +0100
+Subject: [PATCH] system.repo: Debian: Add keyserver proxy support
+
+Introduce a new, optional set of parameters to configure the proxy
+used for key fetching / keyserver access under:
+linux:system:proxy:keyserver:http(s).
+
+Previously, when fetching GPG keys for APT keyring, either using
+public key download & import (as for default repos) or via keyserver,
+we relied on simple `curl` calls or passed it down to Salt aptpkg
+module.
+To be able to retrieve APT keys behind a proxy, one used to have to
+configure the proxy for the Salt minion, which does not yet have
+`no_proxy` support (either *all* or *no* traffic hits the proxy).
+
+When the new http(s) proxy param is set:
+- no longer pass key configuration to Salt aptpkg (until it properly
+ supports `no_proxy`);
+- handle all keys explicitly with `curl` and `apt-key`;
+- set 'http(s)_proxy' env vars for `cmd.run`/`cmd.wait` calls;
+
+If linux:system:proxy:keyserver is not defined, the behavior is
+unchanged for backwards compatibility.
+
+Signed-off-by: Alexandru Avadanii <Alexandru.Avadanii@enea.com>
+---
+ README.rst | 16 ++++++++++++++++
+ linux/system/repo.sls | 40 ++++++++++++++++++++++++++++++++++++++++
+ 2 files changed, 56 insertions(+)
+
+diff --git a/linux/system/repo.sls b/linux/system/repo.sls
+index 5d4d059..724db5a 100644
+--- a/linux/system/repo.sls
++++ b/linux/system/repo.sls
+@@ -96,13 +96,50 @@ linux_repo_{{ name }}_key:
+ - name: "curl -s {{ repo.key_url }} | apt-key add -"
+ - watch:
+ - file: default_repo_list
++{%- if system.proxy.keyserver is defined %}
++ - env:
++ - http_proxy: {{ system.proxy.get('keyserver', {}).get('http', '') }}
++ - https_proxy: {{ system.proxy.get('keyserver', {}).get('https', '') }}
++{%- endif %}
+
+ {%- endif %}
+
++{#- repo.default is false #}
+ {%- else %}
+
+ {%- if repo.get('enabled', True) %}
+
++{%- if system.proxy.keyserver is defined %}
++
++{%- if repo.get('key') %}
++
++linux_repo_{{ name }}_key:
++ cmd.run:
++ - name: "echo '{{ repo.key }}' | apt-key add -"
++
++{%- elif repo.key_url|default(False) %}
++
++linux_repo_{{ name }}_key:
++ cmd.run:
++ - name: "curl -s {{ repo.key_url }} | apt-key add -"
++ - env:
++ - http_proxy: {{ system.proxy.get('keyserver', {}).get('http', '') }}
++ - https_proxy: {{ system.proxy.get('keyserver', {}).get('https', '') }}
++
++{%- elif repo.key_id is defined and repo.key_server is defined %}
++
++linux_repo_{{ name }}_key:
++ cmd.run:
++ - name: "apt-key adv --keyserver {{ repo.key_server }} --recv {{ repo.key_id }}"
++ - env:
++ - http_proxy: {{ system.proxy.get('keyserver', {}).get('http', '') }}
++ - https_proxy: {{ system.proxy.get('keyserver', {}).get('https', '') }}
++
++{%- endif %}
++
++{#- system.proxy.keyserver #}
++{%- endif %}
++
+ linux_repo_{{ name }}:
+ pkgrepo.managed:
+ {%- if repo.ppa is defined %}
+@@ -115,6 +152,7 @@ linux_repo_{{ name }}:
+ {%- endif %}
+ - file: /etc/apt/sources.list.d/{{ name }}.list
+ - clean_file: {{ repo.clean|default(True) }}
++ {%- if system.proxy.keyserver is not defined %}
+ {%- if repo.key_id is defined %}
+ - keyid: {{ repo.key_id }}
+ {%- endif %}
+@@ -124,6 +162,7 @@ linux_repo_{{ name }}:
+ {%- if repo.key_url is defined %}
+ - key_url: {{ repo.key_url }}
+ {%- endif %}
++ {%- endif %}
+ - consolidate: {{ repo.get('consolidate', False) }}
+ - clean_file: {{ repo.get('clean_file', False) }}
+ - refresh_db: {{ repo.get('refresh_db', True) }}
+@@ -140,6 +179,7 @@ linux_repo_{{ name }}:
+ {%- endif %}
+ {%- endif %}
+
++{#- repo.enabled is false #}
+ {%- else %}
+
+ linux_repo_{{ name }}_absent:
diff --git a/mcp/patches/patches.list b/mcp/patches/patches.list
index 4c55223..62b0f60 100644
--- a/mcp/patches/patches.list
+++ b/mcp/patches/patches.list
@@ -13,6 +13,8 @@
/usr/share/salt-formulas/env: 0006-maas-module-Add-VLAN-DHCP-enable-support.patch
/usr/share/salt-formulas/env: 0007-network.interface-Fix-ifup-OVS-port-with-route.patch
/usr/share/salt-formulas/env: 0008-Handle-file_recv-option.patch
+/usr/share/salt-formulas/env: 0009-maas.py-Add-missing-status-ids.patch
/usr/share/salt-formulas/env: 0010-maas-region-allow-timeout-override.patch
/usr/share/salt-formulas/reclass: 0011-service.horizon.server.cluster-Default-to-v2-API.patch
+/usr/share/salt-formulas/env: 0011-system.repo-Debian-Add-keyserver-proxy-support.patch
/usr/share/salt-formulas/env: 0012-linux.storage.lvm-Disable-filter.patch
diff --git a/mcp/patches/scripts/0002-salt-master-setup-Group-APT-install-formulas.patch b/mcp/patches/scripts/0002-salt-master-setup-Group-APT-install-formulas.patch
new file mode 100644
index 0000000..7163268
--- /dev/null
+++ b/mcp/patches/scripts/0002-salt-master-setup-Group-APT-install-formulas.patch
@@ -0,0 +1,68 @@
+::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::
+: Copyright (c) 2017 Mirantis Inc., Enea AB and others.
+:
+: All rights reserved. This program and the accompanying materials
+: are made available under the terms of the Apache License, Version 2.0
+: which accompanies this distribution, and is available at
+: http://www.apache.org/licenses/LICENSE-2.0
+::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::
+From: Alexandru Avadanii <Alexandru.Avadanii@enea.com>
+Date: Tue, 30 Jan 2018 01:23:54 +0100
+Subject: [PATCH] salt-master-setup: Group APT install formulas
+
+Instead of calling `apt install` for each salt formula package that
+we miss, construct a list and install them all at once.
+
+While at it, disable colored output on terminals that don't support
+it, like vt220 (used in OPNFV CI).
+
+Signed-off-by: Alexandru Avadanii <Alexandru.Avadanii@enea.com>
+---
+ salt-master-init.sh | 6 ++++++
+ salt-master-setup.sh | 12 +++++++-----
+ 2 files changed, 13 insertions(+), 5 deletions(-)
+
+diff --git a/salt-master-init.sh b/salt-master-init.sh
+index 343324c..48a05b9 100755
+--- a/salt-master-init.sh
++++ b/salt-master-init.sh
+@@ -24,6 +24,12 @@ options() {
+ export $(find $path -maxdepth 1 -name '*.env' 2> /dev/null | xargs --no-run-if-empty cat ) > /dev/null
+ done;
+
++ # If terminal does not support color output, stop using it
++ if ! tput setaf 1 && tput sgr0; then
++ SALT_OPTS="${SALT_OPTS/--force-color/--no-color}"
++ return 0
++ fi
++
+ export MAGENTA='\033[0;95m'
+ export YELLOW='\033[1;33m'
+ export BLUE='\033[0;35m'
+diff --git a/salt-master-setup.sh b/salt-master-setup.sh
+index 0dd3036..a0c6311 100755
+--- a/salt-master-setup.sh
++++ b/salt-master-setup.sh
+@@ -347,15 +347,17 @@ install_salt_formula_pkg()
+ # Set essentials if FORMULAS_SALT_MASTER is not defined at all
+ [ -z ${FORMULAS_SALT_MASTER+x} ] && declare -a FORMULAS_SALT_MASTER=("linux" "reclass" "salt" "memcached")
+ for formula_service in "${FORMULAS_SALT_MASTER[@]}"; do
+- echo -e "\nConfiguring salt formula ${formula_service} ...\n"
++ echo -e "Configuring salt formula ${formula_service} ..."
+ [ ! -d "${FORMULAS_PATH}/env/${formula_service}" ] && \
+- if ! $SUDO apt-get install -y salt-formula-${formula_service}; then
+- echo -e "\nInstall salt-formula-${formula_service} failed.\n"
+- exit 1
+- fi
++ _FORMULAS_SALT_MASTER="${_FORMULAS_SALT_MASTER} salt-formula-${formula_service}"
++ # Create links first, install pkgs later
+ [ ! -L "/srv/salt/reclass/classes/service/${formula_service}" ] && \
+ ln -sf ${FORMULAS_PATH}/reclass/service/${formula_service} /srv/salt/reclass/classes/service/${formula_service}
+ done
++ if ! $SUDO apt-get install -qqq -y ${_FORMULAS_SALT_MASTER}; then
++ echo -e "\nInstall ${_FORMULAS_SALT_MASTER} failed.\n"
++ exit 1
++ fi
+ ;;
+ rhel)
+ # TODO
diff --git a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/include/proxy.yml b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/include/proxy.yml
index 5d74758..3c324db 100644
--- a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/include/proxy.yml
+++ b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/include/proxy.yml
@@ -16,8 +16,9 @@ parameters:
# https_proxy: http://${_param:infra_maas_node01_deploy_address}:8000
# no_proxy:
# - .local
- salt:
- minion:
+ linux:
+ system:
proxy:
- host: ${_param:infra_maas_node01_deploy_address}
- port: 8000
+ keyserver:
+ http: http://${_param:infra_maas_node01_deploy_address}:8000
+ https: http://${_param:infra_maas_node01_deploy_address}:8000
diff --git a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/config.yml b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/config.yml
index 0b16986..5fbc22f 100644
--- a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/config.yml
+++ b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/config.yml
@@ -25,6 +25,7 @@ classes:
# - system.reclass.storage.system.stacklight_monitor_cluster
# - system.reclass.storage.system.stacklight_telemetry_cluster
- system.reclass.storage.system.infra_maas_single
+ - cluster.baremetal-mcp-ocata-common-ha.infra.lab_proxy_pdf
parameters:
_param:
salt_master_base_environment: prd
diff --git a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/init.yml b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/init.yml
index a4f77a7..875296b 100644
--- a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/init.yml
+++ b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/init.yml
@@ -13,7 +13,7 @@ classes:
# - cluster.baremetal-mcp-ocata-common-ha.stacklight.client
parameters:
_param:
- apt_mk_version: nightly
+ apt_mk_version: 2017.12
mcp_repo_version: 1.1
salt_version: 2016.11
cluster_domain: ${_param:cluster_name}.local
diff --git a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/lab_proxy_pdf.yml.j2 b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/lab_proxy_pdf.yml.j2
new file mode 100644
index 0000000..0e95d2d
--- /dev/null
+++ b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/lab_proxy_pdf.yml.j2
@@ -0,0 +1,24 @@
+##############################################################################
+# Copyright (c) 2018 Mirantis Inc., Enea AB and others.
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
+---
+# We'll craft this class so it can be reused on both cfg01 and mas01
+{%- if conf.idf.fuel.network.upstream_proxy is defined %}
+{%- set upstream_proxy=conf.idf.fuel.network.upstream_proxy %}
+parameters:
+ maas:
+ region:
+ upstream_proxy:
+ address: {{ upstream_proxy.address }}
+ port: {{ upstream_proxy.port }}
+ linux:
+ system:
+ proxy:
+ keyserver:
+ http: http://{{ upstream_proxy.address }}:{{ upstream_proxy.port }}
+ https: http://{{ upstream_proxy.address }}:{{ upstream_proxy.port }}
+{%- endif %}
diff --git a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/maas.yml b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/maas.yml
index 04ffdb4..1136908 100644
--- a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/maas.yml
+++ b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-common-ha/infra/maas.yml
@@ -9,6 +9,7 @@
classes:
- system.maas.region.single
- service.maas.cluster.single
+ - cluster.baremetal-mcp-ocata-common-ha.infra.lab_proxy_pdf
parameters:
_param:
dhcp_interface: ${_param:opnfv_fn_vm_primary_interface}
diff --git a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-odl-ha/openstack/compute_pdf.yml.j2 b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-odl-ha/openstack/compute_pdf.yml.j2
index 7809bd8..519297d 100644
--- a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-odl-ha/openstack/compute_pdf.yml.j2
+++ b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-odl-ha/openstack/compute_pdf.yml.j2
@@ -16,7 +16,7 @@
{%- set idx_public = conf['net_config']['public']['interface'] -%}
{#- Physical interface OS name for each network (e.g. em1, enp1s0f1) -#}
- {%- set node = conf['idf']['fuel']['network']['node'][1] -%}
+ {%- set node = conf['idf']['fuel']['network']['node'][3] -%}
{%- set nic_admin = node['interfaces'][idx_admin] -%}
{%- set nic_mgmt = node['interfaces'][idx_mgmt] -%}
{%- set nic_private = node['interfaces'][idx_private] -%}
@@ -53,7 +53,11 @@ parameters:
{{ nic }}:
enabled: true
type: eth
+ {%- if nic == nic_admin %}
+ proto: dhcp
+ {%- else %}
proto: manual
+ {%- endif %}
name: {{ nic }}
{%- endfor %}
diff --git a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-ovs-ha/openstack/compute_pdf.yml.j2 b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-ovs-ha/openstack/compute_pdf.yml.j2
index 7acb4b4..519297d 100644
--- a/mcp/reclass/classes/cluster/baremetal-mcp-ocata-ovs-ha/openstack/compute_pdf.yml.j2
+++ b/mcp/reclass/classes/cluster/baremetal-mcp-ocata-ovs-ha/openstack/compute_pdf.yml.j2
@@ -16,7 +16,7 @@
{%- set idx_public = conf['net_config']['public']['interface'] -%}
{#- Physical interface OS name for each network (e.g. em1, enp1s0f1) -#}
- {%- set node = conf['idf']['fuel']['network']['node'][0] -%}
+ {%- set node = conf['idf']['fuel']['network']['node'][3] -%}
{%- set nic_admin = node['interfaces'][idx_admin] -%}
{%- set nic_mgmt = node['interfaces'][idx_mgmt] -%}
{%- set nic_private = node['interfaces'][idx_private] -%}
@@ -53,7 +53,11 @@ parameters:
{{ nic }}:
enabled: true
type: eth
+ {%- if nic == nic_admin %}
+ proto: dhcp
+ {%- else %}
proto: manual
+ {%- endif %}
name: {{ nic }}
{%- endfor %}
diff --git a/mcp/reclass/classes/cluster/virtual-mcp-ocata-common-noha/openstack_init.yml b/mcp/reclass/classes/cluster/virtual-mcp-ocata-common-noha/openstack_init.yml
index 81f050f..7037d64 100644
--- a/mcp/reclass/classes/cluster/virtual-mcp-ocata-common-noha/openstack_init.yml
+++ b/mcp/reclass/classes/cluster/virtual-mcp-ocata-common-noha/openstack_init.yml
@@ -9,7 +9,7 @@
parameters:
_param:
openstack_version: ocata
- apt_mk_version: nightly
+ apt_mk_version: 2017.12
mcp_repo_version: 1.1
openstack_region: RegionOne
admin_email: root@localhost
diff --git a/mcp/salt-formulas/maas/machines/delete.sls b/mcp/salt-formulas/maas/machines/delete.sls
index dbb625a..2903f92 100644
--- a/mcp/salt-formulas/maas/machines/delete.sls
+++ b/mcp/salt-formulas/maas/machines/delete.sls
@@ -10,6 +10,7 @@
maas_login_admin:
cmd.run:
- name: "maas-region apikey --username {{ region.admin.username }} > /var/lib/maas/.maas_credentials"
+ - unless: 'test -e /var/lib/maas/.maas_credentials'
# TODO: implement delete_machine via _modules/maas.py
delete_machine:
diff --git a/mcp/salt-formulas/maas/machines/mark_broken_fixed.sls b/mcp/salt-formulas/maas/machines/mark_broken_fixed.sls
index 17a7df8..3983729 100644
--- a/mcp/salt-formulas/maas/machines/mark_broken_fixed.sls
+++ b/mcp/salt-formulas/maas/machines/mark_broken_fixed.sls
@@ -10,6 +10,7 @@
maas_login_admin:
cmd.run:
- name: "maas-region apikey --username {{ region.admin.username }} > /var/lib/maas/.maas_credentials"
+ - unless: 'test -e /var/lib/maas/.maas_credentials'
# TODO: implement mark_broken_fixed_machine via _modules/maas.py
mark_broken_fixed_machine:
diff --git a/mcp/salt-formulas/maas/machines/override_failed_testing.sls b/mcp/salt-formulas/maas/machines/override_failed_testing.sls
new file mode 100644
index 0000000..e7fe1d2
--- /dev/null
+++ b/mcp/salt-formulas/maas/machines/override_failed_testing.sls
@@ -0,0 +1,20 @@
+##############################################################################
+# Copyright (c) 2018 Mirantis Inc., Enea AB and others.
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
+{%- from "maas/map.jinja" import region with context %}
+
+maas_login_admin:
+ cmd.run:
+ - name: "maas-region apikey --username {{ region.admin.username }} > /var/lib/maas/.maas_credentials"
+ - unless: 'test -e /var/lib/maas/.maas_credentials'
+
+# TODO: implement override_failed_testing via _modules/maas.py
+mark_broken_fixed_machine:
+ cmd.run:
+ - name: "maas login {{ region.admin.username }} http://{{ region.bind.host }}:5240/MAAS/api/2.0 - < /var/lib/maas/.maas_credentials && maas opnfv machine override-failed-testing {{ pillar['system_id'] }}"
+ - require:
+ - cmd: maas_login_admin
diff --git a/mcp/salt-formulas/opendaylight/server.sls b/mcp/salt-formulas/opendaylight/server.sls
index 30c337c..f953cdd 100644
--- a/mcp/salt-formulas/opendaylight/server.sls
+++ b/mcp/salt-formulas/opendaylight/server.sls
@@ -6,9 +6,23 @@
# http://www.apache.org/licenses/LICENSE-2.0
##############################################################################
{% from "opendaylight/map.jinja" import server with context %}
+{% from "linux/map.jinja" import system with context %}
{%- if server.enabled %}
+opendaylight_repo_key:
+ cmd.run:
+ - name: "apt-key adv --keyserver keyserver.ubuntu.com --recv 44C05248"
+{%- if system.proxy is defined and system.proxy.keyserver is defined %}
+ - env:
+{%- if system.proxy.keyserver.http is defined %}
+ - http_proxy: {{ system.proxy.keyserver.http }}
+{%- endif %}
+{%- if system.proxy.keyserver.https is defined %}
+ - https_proxy: {{ system.proxy.keyserver.https }}
+{%- endif %}
+{%- endif %}
+
opendaylight_repo:
pkgrepo.managed:
# NOTE(armband): PPA handling behind proxy broken, define it explicitly
@@ -17,8 +31,6 @@ opendaylight_repo:
- human_name: opendaylight-ppa
- name: deb http://ppa.launchpad.net/odl-team/nitrogen/ubuntu xenial main
- file: /etc/apt/sources.list.d/odl-team-ubuntu-nitrogen-xenial.list
- - keyid: 49B07274951063870A8B7EAE7B8AA1A344C05248
- - keyserver: keyserver.ubuntu.com
opendaylight:
pkg.installed:
diff --git a/mcp/scripts/lib.sh b/mcp/scripts/lib.sh
index 29c27d9..6bbcb44 100644
--- a/mcp/scripts/lib.sh
+++ b/mcp/scripts/lib.sh
@@ -32,7 +32,7 @@ function get_base_image {
local image_dir=$2
mkdir -p "${image_dir}"
- wget -P "${image_dir}" -N "${base_image}"
+ wget --progress=dot:giga -P "${image_dir}" -N "${base_image}"
}
function __kernel_modules {
@@ -456,3 +456,15 @@ function wait_for {
return 1
)
}
+
+function do_sysctl_cfg {
+ local _conf='/etc/sysctl.d/99-opnfv-fuel-bridge.conf'
+ # https://wiki.libvirt.org/page/Net.bridge.bridge-nf-call_and_sysctl.conf
+ if modprobe br_netfilter bridge; then
+ echo 'net.bridge.bridge-nf-call-arptables = 0' |& sudo tee "${_conf}"
+ echo 'net.bridge.bridge-nf-call-iptables = 0' |& sudo tee -a "${_conf}"
+ echo 'net.bridge.bridge-nf-call-ip6tables = 0' |& sudo tee -a "${_conf}"
+ # Some distros / sysadmins explicitly blacklist br_netfilter
+ sudo sysctl -q -p "${_conf}" || true
+ fi
+}
diff --git a/mcp/scripts/requirements_rpm.yaml b/mcp/scripts/requirements_rpm.yaml
index ebd7e85..a9c624f 100644
--- a/mcp/scripts/requirements_rpm.yaml
+++ b/mcp/scripts/requirements_rpm.yaml
@@ -16,6 +16,8 @@ requirements_pkg:
- elfutils-libelf-devel
- e2fsprogs
- genisoimage
+ - gcc
+ - gdisk
- git
- kpartx
- libvirt
diff --git a/mcp/scripts/salt.sh b/mcp/scripts/salt.sh
index ac9d74f..ef7d383 100755
--- a/mcp/scripts/salt.sh
+++ b/mcp/scripts/salt.sh
@@ -47,6 +47,7 @@ fi
ssh ${SSH_OPTS} "${SSH_SALT}" bash -s -e << SALT_INSTALL_END
sudo -i
set -e
+ export TERM=${TERM}
export CI_DEBUG=${CI_DEBUG}; [[ "${CI_DEBUG}" =~ (false|0) ]] || set -x
echo -n 'Checking out cloud-init has finished running ...'
@@ -54,7 +55,7 @@ ssh ${SSH_OPTS} "${SSH_SALT}" bash -s -e << SALT_INSTALL_END
echo ' done'
mkdir -p /srv/salt /usr/share/salt-formulas/reclass
- rm -rf ${OPNFV_GIT_DIR}
+ rm -rf ${OPNFV_GIT_DIR} ${OPNFV_FUEL_DIR}
mv ${OPNFV_TMP_DIR} ${OPNFV_GIT_DIR} && chown -R root.root ${OPNFV_GIT_DIR}
find ${OPNFV_GIT_DIR} -name '.git' -type f | while read f_git; do
sed -i 's@${LOCAL_GIT_DIR}@${OPNFV_GIT_DIR}@g' \$f_git
@@ -70,7 +71,8 @@ ssh ${SSH_OPTS} "${SSH_SALT}" bash -s -e << SALT_INSTALL_END
cd /srv/salt/scripts
export DEBIAN_FRONTEND=noninteractive
- OLD_DOMAIN=\$(grep -Pzo "id: cfg01\.\K(\S*)" /etc/salt/minion.d/minion.conf) || true
+ echo 'Dpkg::Use-Pty "0";' > /etc/apt/apt.conf.d/90silence-dpkg
+ OLD_DOMAIN=\$(grep -sPzo "id: cfg01\.\K(\S*)" /etc/salt/minion.d/minion.conf) || true
BOOTSTRAP_SALTSTACK_OPTS=" -r -dX stable 2016.11 " \
MASTER_HOSTNAME=cfg01.${CLUSTER_DOMAIN} DISTRIB_REVISION=2017.12 \
EXTRA_FORMULAS="nfs" \
@@ -86,7 +88,7 @@ ssh ${SSH_OPTS} "${SSH_SALT}" bash -s -e << SALT_INSTALL_END
# In case scenario changed (and implicitly domain name), re-register minions
if [ -n "\${OLD_DOMAIN}" ] && [ "\${OLD_DOMAIN}" != "${CLUSTER_DOMAIN}" ]; then
- salt "*.\${OLD_DOMAIN}" cmd.run "grep \${OLD_DOMAIN} -Rl /etc/salt | \
+ salt "*.\${OLD_DOMAIN}" cmd.run "grep \${OLD_DOMAIN} -sRl /etc/salt | \
xargs --no-run-if-empty sed -i 's/\${OLD_DOMAIN}/${CLUSTER_DOMAIN}/g'; \
service salt-minion restart" || true
salt-key -yd "*.\${OLD_DOMAIN}"