Age | Commit message (Collapse) | Author | Files | Lines |
|
|
|
|
|
The redis_vip should come from a Neutron Port as its cidr depends
on the Neutron Network configuration. This change adds 2 new files
and modifies 1 in the network/ports directory:
- noop.yaml - Passes through the ctlplane Controller IP (modified)
- ctlplane_vip.yaml - Creates a new VIP on the control plane
- vip.yaml - Creates a VIP on the named network (for isolated nets)
Also, changes to overcloud-without-mergepy.yaml create the
Redis Virtual IP. The standard resource registry was modified to
use noop.yaml for the new Redis VIP. The Puppet resource registry
was modified to use ctlplane_vip.yaml by default, but can be made
to use vip.yaml when network isolation is used by using an
environment file. vip.yaml will place the VIP according to the
ServiceNetMap, which can also be overridden.
We use this new VIP port definition to assign a VIP to Redis,
but follow-up patches will assign VIPs to the rest of the
services in a similar fashion.
Co-Authored-By: Dan Sneddon <dsneddon@redhat.com>
Change-Id: I2cb44ea7a057c4064d0e1999702623618ee3390c
|
|
The Redis bind host should be set from [1] template.
1. https://github.com/openstack/tripleo-heat-templates/blob/master/puppet/controller-puppet.yaml#L985
Change-Id: I8713db1a7ad739692817921248edcc6b0e819ade
|
|
Adds the horizon (httpd) service as pacemaker resource
Also adds a default for the horizon::django_session_engine [1]
which was previously unconfigured. Also adds a server-status.conf
for httpd/pacemaker [2]
[1] https://docs.djangoproject.com/en/dev/topics/http/sessions/#using-cached-sessions
[2] https://github.com/beekhof/osp-ha-deploy/blob/master/pcmk/horizon.scenario#L72
Change-Id: I320837dfecf3241355e8a3345d0ff271592da491
|
|
|
|
|
|
|
|
Two PR have been merged upstream that let use improve our current
implementation :
* service_manage[1]
* conn string has namevar[2]
[1] https://github.com/puppetlabs/puppetlabs-mongodb/pull/198
[2] https://github.com/puppetlabs/puppetlabs-mongodb/pull/200
Change-Id: Ia2247348a9e0292b5fcbc65ea1e41e6bc7c477fa
|
|
Since t-h-t can now deploy a Redis cluster, we can rely on it as the
tooz backend for high availability.
Change-Id: If045a273388aa2e725b6de624e09aa9c85248cc4
|
|
Change-Id: I731b408f24da01c1bc897bfffe8fd4d5638932ed
|
|
Enables support for configuring Cinder with a NetApp backend.
This change adds all relevant parameters for:
- Clustered Data ONTAP (NFS, iSCSI, FC)
- Data ONTAP 7-Mode (NFS, iSCSI, FC)
- E-Series (iSCSI)
Change-Id: If6c6e511ef2d26c4794e3b37c61e5318485ff4db
|
|
The list of drivers loaded by the ML2 plugin does not have to
match the list of tenant_network_types, this will make ML2 load
the flat, gre, vxlan and vlan drivers so that the provider
networks can be of flat (default) and vlan type as well.
Change-Id: I0b74f86acf5c1ff644deb46c0a1d14129c1882d4
|
|
This patch updates the overcloud pacemaker role manifest so
that it optionally configures VIPs on isolated networks if
they are enabled.
Change-Id: I6123ee622abe4d8d7b5f76cf9bac43acd80c1f64
|
|
This patch refactors the puppet controller role so that it
makes use of per service VIP settings for each service.
Previously the VIP for the ctlplane was hard wired to
many of the controller service. With this patch we have
the ability to isolate traffic for services which
made use of the ctlplane and public VIPs for their
settings.
The implementation includes:
* stops the use of the VirtualIP and PublicVirtualIP within the
controller role. These parameters have now been replaced with
per service heat parameters for the controller nested stack which
are determined via VipMap based on per service settings in the heat
environment.
* All VIP configuration is now moved into puppet/vip-config.yaml.
This made sense so we could deprecate the use of the VirtualIP
and PublicVirtualIP settings above.
* The puppet manifests for the controller were cleaned up for several
to use Hiera directly instead of constructing URLs based on the
static controller and public network VIPs. This improvement
was something we wanted to do anyways and made the implementation
cleaner.
Change-Id: I9b9a15be67f74bec97366408f7047acfd6ea0ec6
|
|
This patch adds a new NetIpListMap abstraction which we can use
to make the all-nodes-config IP list network assignments
configurable. Ip address lists for all overcloud services
which require IPs were added to all-nodes-config so
that puppet manifests can be directly supplied the
correct network list for each service.
Change-Id: I209f2b4f97a4bb78648c54813dad8615770bcf1a
|
|
|
|
|
|
|
|
With current effort of creating isolated networks, the controller_host
hiera variable does not exist anymore. Hence we remove it else the
lookup will fail.
The hiera binding neutron::agents::ml2::ovs::local_ip has been written
in another review[1]
[1] I1dc11987b4ea3c37775b14fbdddb75588499e9bb
Change-Id: I12777c512d379210e5cddb5e683be4d79808fa2c
|
|
|
|
Change-Id: I1c8fc6beacc8352ad2aabe44ff20614ac52c1795
|
|
Change-Id: I1243b68506f37d6b78807c03948874ae100fef65
|
|
Constraints based on vncproxy are commented due to it not starting
with websockify < 0.6, see [1]
1. http://lists.openstack.org/pipermail/openstack-dev/2014-October/048535.html
Co-Authored-By: Jiri Stransky <jistr@redhat.com>
Change-Id: Ie51014bf563920d2e75c5e38942bc42ddc2a3939
|
|
Adds neutron-server, neutron-l3-agent, neutron-dhcp-agent,
neutron-openvswitch-agent and neutron-metadata-agent as
pacemaker resources.
Change-Id: I4dcc6f56db4c27a2a4f627fa8303cbeb2bd563d4
|
|
This change adds parameters to specify which networks the MySQL
service will use. If the internal_api network exists the MySQL
service will bind to the IP address on that network, otherwise
the services will default to the IP on the Undercloud 'ctlplane'
network.
This patch also drop the old 'controller_host' variable from
the puppet controller template since it is no longer in use.
Change-Id: I4fba2c957f7db47e916bc269fb4bd32ccc99bd4c
|
|
|
|
|
|
|
|
|
|
Fixes the colocation order between glance-api and glance-registry to
match the ref-arch[1]
[1]
https://github.com/beekhof/osp-ha-deploy/blob/master/pcmk/glance.scenario#L108
Change-Id: I40f35afedb3333d97c8b689538bb80a90a66afe8
|
|
Make sure the keystone service starts before the glance-registry one.
Change-Id: Ia81df13682bf556a39cc36520def48105ee3e27d
|
|
Make sure the keystone service starts before the cinder-api one.
Change-Id: I21549c066afcf051e52fc4bba4fae2f34ad2ba4b
|
|
The interface for pcmk_resource offers the parameter master_params to set
--master during the resource creation.
Change-Id: I6fa769f14a6248b371810af3ba6819a1f9ed9442
|
|
Depends-On: I7b992450176595a89dba9fe2eccf619af2645d6b
Change-Id: I30cebb6d3a8670f49587bedaf51af18a87a8d24c
|
|
|
|
|
|
|
|
|
|
Previously the Glance Pacemaker resources were mistakenly defined
on all nodes causing intermittent duplication errors.
Change-Id: I839ee49b153aa96ec08ebdb7e44aaeac28785963
|
|
Change-Id: I4631f962415164975143e94ec0b251ee5972c552
|
|
Change-Id: If87cc4d55e8524246d2cd41a62805f84780006b2
|
|
Add Pacemaker resources for Cinder services, also add relevant ordering
and colocation constraints.
Change-Id: Idc2e1b5ec96d882543f7a1a4ec723a010020ab02
|
|
|
|
With change I4b6b77e878017bf92d7c59c868d393e74405a355 we started
using the root user for clustercheck script so we don't need to
create the clustercheck user anymore.
Change-Id: Ic92bd12baeeeaf3f674e766fbc0a8badfb44822f
|
|
|
|
Previously we've been starting non-pacemakerized services in step 3 on
bootstrap node and in step 4 on others. Now that $sync_db in OpenStack
Puppet modules is decoupled from $enabled and $manage_service [1] we can
start the services in step 4 on all nodes.
[1] https://bugs.launchpad.net/puppet-glance/+bug/1452278
Change-Id: I6351d972ab00f4661d98338d95310d33f271de2f
|
|
|
|
|
|
|