summaryrefslogtreecommitdiffstats
path: root/lib/ansible/playbooks/post_deploy_overcloud.yml
AgeCommit message (Collapse)AuthorFilesLines
2019-01-11Fixes undercloud install failure with setting hostnameTim Rozet1-3/+3
There is a new bug when deploying master/rocky where the OS of the undercloud/overcloud is now upgraded to CentOS 7.6. When the undercloud install runs it fails to configure the hostname using hostnamectl. This is because systemd-hostnamed is not running and fails to start. Simply reloading dbus seems to fix the issue. In the dbus logs there are odd error messages like: dbus-daemon[3230]: Unknown username "root" in message bus configuration file Disabling selinux seems to fix this. This patch also moves to use podman instead of docker for container management and invokes a script in Ansible which updates NIC templates as new variables are added upstream. Furthermore, with the new patches for routed networks in OOO, it is now required that the MTU is set in network-data, as well as adding the External network to the Compute role. Now the External network is removed by default from the Compute role. Change-Id: Ie8b86f6f28d69bda11b1f7a430df882970ac3cb9 Signed-off-by: Tim Rozet <trozet@redhat.com>
2018-12-14Fix NFS issues with NovaTim Rozet1-5/+3
There are problems with Nova launching instantces due to permissions with nova being able to read/write certain directories on the NFS. The permissions are right on the NFS and the folders the NFS mounts to, but there still seems to be issues. The cause may be using a directory under /root as the NFS mount. This patch moves the NFS mounts to be individual folders under /. The patch also restart nova_compute docker container as NFS problems still persist unless this is done. JIRA: APEX-654 Change-Id: I25eee98c1a6516dfa44c686c2e614f6dc7000d98 Signed-off-by: Tim Rozet <trozet@redhat.com>
2018-09-27Enable OVN scenariosTim Rozet1-1/+4
As of Queens only HA OVN deployments are supported. Change-Id: I184c5a096fec9cbc3cf2ec06218700138ea3ed57 Signed-off-by: Tim Rozet <trozet@redhat.com>
2018-08-22Enable SFC scenarios for GambiaRicardo Noriega1-9/+0
- This patch will install OVS 2.9.2 including its kernel module which allows native NSH headers. - Fix Custom OVS due to bug: https://bugzilla.redhat.com/show_bug.cgi?id=1544892 - Tacker is disable for the time being, tacker-conductor needs to be enabled. JIRA: APEX-630 Change-Id: Ia410309fd7053602ce78eae919839d0f57c9742a Signed-off-by: Ricardo Noriega <rnoriega@redhat.com>
2018-07-31Use metadata IP instead of FQDNTim Rozet1-0/+11
There is an issue with loss of external network connectivity that prevents cloud init to instances working. This becomes a big problem with snapshots where there is no external network connectivity. Cloud init fails because each request takes over 30 seconds to get a response. This is because in the background neutron metadata agent is proxying the request to nova metadata agent with an HTTP GET using the FQDN. For whatever reason, a DNS lookup happens even though the entry exists for the FQDN in /etc/hosts and waits 30 seconds until timing out. After this timeout, a 200 OK is sent and metadata works. This patch modifies the config file for metadata to use nova metadata server's internal IP rather than FQDN as there is no option in OOO to use IPs instead of FQDN. Change-Id: I6960181a227d0002c99aeae5112f59807dc41d7a Signed-off-by: Tim Rozet <trozet@redhat.com>
2018-07-24Open port 8101 on controllers for karafTim Rozet1-0/+15
By default 8101 (karaf shell) is blocked on controllers. In Apex we advertise in our user guide (and tools scripts) the ability to connect to karaf shell. It is also required to run CSIT. This patch opens the port when ODL is deployed. Change-Id: Ib3ece41f19607bafc329d9de390cf774766a26cd Signed-off-by: Tim Rozet <trozet@redhat.com>
2018-07-20Fixes for snapshotsTim Rozet1-0/+6
With deploying snapshots with a new ODL, we currently bring down the docker container and bring up the tar.gz distro of ODL on the Overcloud host itself (not rebuilding/using container). Therefore we need java installed so that ODL can run on the host. In the future this may change, but it works well and keeps things simple for now. Additionally, there was a change upstream to make the opendaylight container docker restart policy "unless-stopped" which means it will no longer restart automatically when docker is stopped/started. Therefore on first snapshot bring up (without the previously mentioned ODL reinstallation) the container does not start, and snapshot deployment fails. This patch includes a change to the restart policy to always restart it. Change-Id: Icc712ba147e578a28e371313154ae3190676f0dc Signed-off-by: Tim Rozet <trozet@redhat.com>
2018-05-21Migrates master to use direct upstreamTim Rozet1-0/+1
We now move master to deploy from upstream. That means we do not need to build undercloud/overcloud images anymore. Changes-Include: - Remove bash build scripts as we do not need to build anything other than the python package anymore - Remove building images or iso from build.py - Remove building of images and iso from Makefile - Rename/refactor deploy settings files for nosdn and odl. The new convention is that the typical scenario names we use will deploy master. We also support n-1 OS, so in that case we use the branch name for the "feature" in the scenario name: os-odl-queens-noha. - Tacker/Congress are disabled in settings files until we fix that with upstream. Containers are now enabled by default. - Disable TLS for undercloud (was changed upstream to default enabled) - Fix environments docker directory for master THT (was changed upstream) - Includes fix for LP#1768901 - Includes workaround for LP#1770692 - Moves to docker.io for container images as it is more stable and should contain the same images - Removes the term 'common' from apex packaging for referencing the Python Apex package Change-Id: If6b433860b3ff882686c78d0f24a2f0c52b9b57a Signed-off-by: Tim Rozet <trozet@redhat.com>
2018-04-09Fix functional issues after nosdn deploymentTim Rozet1-0/+17
After deploying with nosdn, it looks like there is some out of state issue between the services. First guess looks like something is going on with the services and timing of registering to each other through rabbit. Simply restarting the services seems to sync them back up correctly. Change-Id: I417911067c841725ee12eb9354e5759054724e01 Signed-off-by: Tim Rozet <trozet@redhat.com>
2017-10-18Adapt zrpcd to work for Carbon/Nitrogentomsou1-6/+6
It updates the quagga tar.gz file which is used to install zrpcd/quagga. The new rpms are compiled with thrift-version 4 Zrpcd is installed as a process. Additional changes takwqe place adopting the installation and checking of zrpcd process Note that an extra patch is needed to update or remove the building procedure, which is present within the project but not used actually. JIRA: APEX-508 Change-Id: If6da075d2aafdb8c2ecfc599f595a7201490aa23 Signed-off-by: tomsou <soth@intracom-telecom.com>
2017-09-25Applies Tacker upstream patches and fixes fernet dir permsTim Rozet1-1/+3
We need a couple patches to Tacker upstream that we will not get now that our images are frozen in RDO. We should remove them on upgrade to Pike. Also this patch fixes the permissions on the fernet keys dir so that the Tacker group can write. Change-Id: I6468356f1d16d3b8ab58d46f653431e29960e190 Signed-off-by: Tim Rozet <trozet@redhat.com>
2017-09-22Fixes Tacker auth and fernet keys dirTim Rozet1-0/+6
Tacker authtoken was missing password which is now added in the tht patch. Tacker also needs 'etc/tacker/vim/fernet_keys' in order to be able to create VIMs. The real fix to this missing directory is to the rpm itself: https://review.rdoproject.org/r/#/c/9725/ However in this patch we ensure it is created in Ansible because RDO images are frozen in for now in Apex so we would not pick up new RDO packages. apex-tripleo-heat-templates: I2b347cbc4595e6651b0d4be032cb862fde72e15f Change-Id: I76959cbfd581b9fec69749ccbcbcb39dc959a883 Signed-off-by: Tim Rozet <trozet@redhat.com>
2017-08-23Migrates Apex to PythonTim Rozet1-0/+45
Removes all bash libraries and converts almost all of the code to a mixture of Python and Ansible. utils.sh and clean.sh still exist. clean.sh will be migrated fully to clean.py in another patch. The Apex Python package is now built into the opnfv-apex-common RPM. To install locally do 'pip3 install .'. To deploy: opnfv-deploy -d <file> -n <file> --image-dir /root/apex/.build -v --debug Non-python files (THT yaml, settings files, ansible playbooks) are all installed into /usr/share/opnfv-apex/. The RPM will copy settings files into /etc/opnfv-apex/. JIRA: APEX-317 Change-Id: I3232f0329bcd13bce5a28da6a8c9c84d0b048024 Signed-off-by: Tim Rozet <trozet@redhat.com>