Age | Commit message (Collapse) | Author | Files | Lines |
|
As of Queens only HA OVN deployments are supported.
Change-Id: I184c5a096fec9cbc3cf2ec06218700138ea3ed57
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
- This patch will install OVS 2.9.2 including
its kernel module which allows native NSH
headers.
- Fix Custom OVS due to bug:
https://bugzilla.redhat.com/show_bug.cgi?id=1544892
- Tacker is disable for the time being, tacker-conductor
needs to be enabled.
JIRA: APEX-630
Change-Id: Ia410309fd7053602ce78eae919839d0f57c9742a
Signed-off-by: Ricardo Noriega <rnoriega@redhat.com>
|
|
There is an issue with loss of external network connectivity that
prevents cloud init to instances working. This becomes a big problem
with snapshots where there is no external network connectivity. Cloud
init fails because each request takes over 30 seconds to get a response.
This is because in the background neutron metadata agent is proxying the
request to nova metadata agent with an HTTP GET using the FQDN. For
whatever reason, a DNS lookup happens even though the entry exists for
the FQDN in /etc/hosts and waits 30 seconds until timing out. After this
timeout, a 200 OK is sent and metadata works.
This patch modifies the config file for metadata to use nova metadata
server's internal IP rather than FQDN as there is no option in OOO to
use IPs instead of FQDN.
Change-Id: I6960181a227d0002c99aeae5112f59807dc41d7a
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
By default 8101 (karaf shell) is blocked on controllers. In Apex we
advertise in our user guide (and tools scripts) the ability to connect
to karaf shell. It is also required to run CSIT. This patch opens the
port when ODL is deployed.
Change-Id: Ib3ece41f19607bafc329d9de390cf774766a26cd
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
With deploying snapshots with a new ODL, we currently bring down the
docker container and bring up the tar.gz distro of ODL on the Overcloud
host itself (not rebuilding/using container). Therefore we need java
installed so that ODL can run on the host. In the future this may
change, but it works well and keeps things simple for now.
Additionally, there was a change upstream to make the opendaylight
container docker restart policy "unless-stopped" which means it will
no longer restart automatically when docker is stopped/started.
Therefore on first snapshot bring up (without the previously mentioned
ODL reinstallation) the container does not start, and snapshot
deployment fails. This patch includes a change to the restart policy to
always restart it.
Change-Id: Icc712ba147e578a28e371313154ae3190676f0dc
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
We now move master to deploy from upstream. That means we do not need
to build undercloud/overcloud images anymore.
Changes-Include:
- Remove bash build scripts as we do not need to build anything
other than the python package anymore
- Remove building images or iso from build.py
- Remove building of images and iso from Makefile
- Rename/refactor deploy settings files for nosdn and odl. The new
convention is that the typical scenario names we use will deploy
master. We also support n-1 OS, so in that case we use the branch
name for the "feature" in the scenario name: os-odl-queens-noha.
- Tacker/Congress are disabled in settings files until we fix that with
upstream. Containers are now enabled by default.
- Disable TLS for undercloud (was changed upstream to default enabled)
- Fix environments docker directory for master THT (was changed upstream)
- Includes fix for LP#1768901
- Includes workaround for LP#1770692
- Moves to docker.io for container images as it is more stable and
should contain the same images
- Removes the term 'common' from apex packaging for referencing the
Python Apex package
Change-Id: If6b433860b3ff882686c78d0f24a2f0c52b9b57a
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
After deploying with nosdn, it looks like there is some out of state
issue between the services. First guess looks like something is going
on with the services and timing of registering to each other through
rabbit. Simply restarting the services seems to sync them back up
correctly.
Change-Id: I417911067c841725ee12eb9354e5759054724e01
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
It updates the quagga tar.gz file which is used to
install zrpcd/quagga.
The new rpms are compiled with thrift-version 4
Zrpcd is installed as a process. Additional changes takwqe place
adopting the installation and checking of zrpcd process
Note that an extra patch is needed to update or remove the
building procedure, which is present within the project
but not used actually.
JIRA: APEX-508
Change-Id: If6da075d2aafdb8c2ecfc599f595a7201490aa23
Signed-off-by: tomsou <soth@intracom-telecom.com>
|
|
We need a couple patches to Tacker upstream that we will not get now
that our images are frozen in RDO. We should remove them on upgrade to
Pike. Also this patch fixes the permissions on the fernet keys dir so
that the Tacker group can write.
Change-Id: I6468356f1d16d3b8ab58d46f653431e29960e190
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
Tacker authtoken was missing password which is now added in the tht
patch. Tacker also needs 'etc/tacker/vim/fernet_keys' in order to be
able to create VIMs. The real fix to this missing directory is to the
rpm itself:
https://review.rdoproject.org/r/#/c/9725/
However in this patch we ensure it is created in Ansible because
RDO images are frozen in for now in Apex so we would not pick up new
RDO packages.
apex-tripleo-heat-templates: I2b347cbc4595e6651b0d4be032cb862fde72e15f
Change-Id: I76959cbfd581b9fec69749ccbcbcb39dc959a883
Signed-off-by: Tim Rozet <trozet@redhat.com>
|
|
Removes all bash libraries and converts almost all of the code to a
mixture of Python and Ansible. utils.sh and clean.sh still exist.
clean.sh will be migrated fully to clean.py in another patch.
The Apex Python package is now built into the opnfv-apex-common RPM. To
install locally do 'pip3 install .'. To deploy:
opnfv-deploy -d <file> -n <file> --image-dir /root/apex/.build -v --debug
Non-python files (THT yaml, settings files, ansible playbooks) are all
installed into /usr/share/opnfv-apex/. The RPM will copy settings files
into /etc/opnfv-apex/.
JIRA: APEX-317
Change-Id: I3232f0329bcd13bce5a28da6a8c9c84d0b048024
Signed-off-by: Tim Rozet <trozet@redhat.com>
|