summaryrefslogtreecommitdiffstats
path: root/puppet/controller-role.yaml
AgeCommit message (Collapse)AuthorFilesLines
2017-08-02Make RoleParameters and key_name descriptions consistentBen Nemec1-1/+1
The key_name default is ignored because the parameter is used in some mutually exclusive environments where the default doesn't need to be the same. Change-Id: I77c1a1159fae38d03b0e59b80ae6bee491d734d7 Partial-Bug: 1700664
2017-07-13Merge "Remove hardcoded enable_load_balancer from Controller role"Jenkins1-7/+0
2017-07-13Merge "Revert "Revert "Blacklist support for ExtraConfig"""Jenkins1-0/+17
2017-07-13Merge "Add DeployedServerEnvironmentOutput"Jenkins1-0/+39
2017-07-11Remove hardcoded enable_load_balancer from Controller roleSteven Hardy1-7/+0
This is associated with the haproxy service, so set the hieradata there instead. This is needed so we can render the controller role template via j2, and also if anyone ever wants to run haproxy on some role other then the Controller. Change-Id: I82b992afe42f6da7788f6efca2366863c3bf68f7 Partially-Implements: blueprint composable-networks
2017-07-11Remove controller specific bootstack_nodeidSteven Hardy1-4/+0
This has been replaced for some time by bootstrap_nodeid which isn't hard-coded to the Controller role. Change-Id: I2c172de13646e5b88cb9930a93ca71fcc990e522 Depends-On: I0a9fced847caf344e5d26b452f1bd40afab8f029
2017-07-10Revert "Revert "Blacklist support for ExtraConfig""James Slagle1-0/+17
There is a Heat patch posted (via Depends-On) that resolves the issue that caused this to be reverted. This reverts the revert and we need to make sure all the upgrades jobs pass before we merge this patch. This reverts commit 69936229f4def703cd44ab164d8d1989c9fa37cb. Closes-Bug: #1699463 implements blueprint disable-deployments Change-Id: Iedf680fddfbfc020d301bec8837a0cb98d481eb5
2017-07-10Add DeployedServerEnvironmentOutputJames Slagle1-0/+39
Add a new output, DeployedServerEnvionmentOutput, that can be used as the contents of an environment file to input into a services only stack when using split-stack. The parameter simplifies the manual steps needed to deploy split-stack. By default, the resource that generates the output is mapped to OS::Heat::None. implements blueprint split-stack-default Change-Id: I6004cd3f56778f078a69a20e93a0eba0c574b3db
2017-06-27Re-enable default for RoleParametersBen Nemec1-0/+1
I471037de35e7f349d900462ec3ffb16fe2d6ebd9 accidentally removed the default from the RoleParameters parameter. This change just puts it back. Change-Id: I29b472897e07229715fc2fea3b55e90473eb0069
2017-06-27Merge "Enable DPDK on boot using PreNetworkConfig"Jenkins1-0/+1
2017-06-26Merge "Add DeploymentSwiftDataMap parameter"Jenkins1-1/+30
2017-06-26Merge "Add os-collect-config data as an output"Jenkins1-0/+3
2017-06-24Merge "Replace hardcoded 'br-ex' with NeutronPhysicalBridge parameter."Jenkins1-2/+6
2017-06-23Replace hardcoded 'br-ex' with NeutronPhysicalBridge parameter.Dan Sneddon1-2/+6
This change uses the NeutronPhysicalBridge parameter on all roles, rather than hard-coding the "br-ex" name. Previously, there were different parameters for controller and compute roles, but since we use a unified bridge name with OVS, this is unnecessary. Change-Id: I6d9189404fae67bcc33ddc2ba3ce1b0385dd989d Closes-bug: 1669130
2017-06-23Enable DPDK on boot using PreNetworkConfigSaravanan KR1-0/+1
DPDK has to be enabled on openvswitch on the boot before configuring the network as when the network uses DPDK ports OvS should be ready to handle DPDK. Enabled DPDK via PreNetworkConfig by checking if ServiceNames contains DPDK service. Implements: blueprint ovs-2-6-dpdk Closes-Bug: #1654975 Depends-On: I83a540336c01a696780621fb2b39486a6abf0917 Change-Id: I7af4534d91e67c94ba559b78b9ac6a001e639db3
2017-06-22Merge "Revert "Blacklist support for ExtraConfig""Jenkins1-17/+0
2017-06-22Revert "Blacklist support for ExtraConfig"Alex Schultz1-17/+0
This reverts commit d6c0979eb3de79b8c3a79ea5798498f0241eb32d. This seems to be causing issues in Heat in upgrades. Change-Id: I379fb2133358ba9c3c989c98a2dd399ad064f706 Related-Bug: #1699463
2017-06-19Merge "Blacklist support for ExtraConfig"Jenkins1-0/+17
2017-06-16Blacklist support for ExtraConfigJames Slagle1-0/+17
Commit I46941e54a476c7cc8645cd1aff391c9c6c5434de added support for blacklisting servers from triggered Heat deployments. This commit adds that functionality to the remaining Deployments in tripleo-heat-templates for the ExtraConfig interfaces. Since we can not (should not) change the interface to ExtraConfig, Heat conditions are used on the actual <role>ExtraConfigPre and NodeExtraConfig resources instead of using the actions approach on Deployments. Change-Id: I38fdb50d1d966a6c3651980c52298317fa3bece4
2017-06-15Add DeploymentSwiftDataMap parameterJames Slagle1-1/+30
The DeploymentSwiftDataMap parameter is used to set the deployment_swift_data property on the Server resoures. The parameter is a map of role names and node indexes to Swift container and object names to be used for storing deployment data. The parameter allows for using predefined Swift objects for storing deployment data instead of container/object names with generated uuid's from Heat. implements blueprint split-stack-default Depends-On: Ia07e9374a4b95bd0e74fc47fb9df4bf6ad096715 Change-Id: I471037de35e7f349d900462ec3ffb16fe2d6ebd9
2017-06-15Add os-collect-config data as an outputJames Slagle1-0/+3
Adds a new output, ServerOsCollectConfigData, which is the os-collect-config configuration associated with each server resource. This can be used to [pre]configure the os-collect-config agents on deployed-server's. Having the data available as a stack output is more user friendly than having to query several nested levels of stack resources, and then inspect resource metadata. implements blueprint split-stack-default Change-Id: Iaf062f1a72e2a9e4d97f84c67f72408a6b5cebfc Depends-On: I8acfd67cd8138d587cc362184c84a08134bf3157
2017-06-15Update CloudDomain descriptionBen Nemec1-2/+2
First, this parameter must match what is configured on the undercloud, so strengthen that language. There is also now an undercloud.conf parameter that can be used to set the requisite options on the undercloud services, so just point users at that rather than trying to explain how to configure the services manually (which is error-prone and doesn't survive undercloud updates). Change-Id: I002cce176e3430473a29e79efde3464bddb24cc7
2017-06-15Merge "Modify PreNetworkConfig config inline with role-specific parameters"Jenkins1-0/+5
2017-06-13Add fqdn_externalAlex Schultz1-0/+1
In newton, we used to construct the fqdn_$NETWORK in puppet-tripleo for external, internal_api, storage, storage_mgmt, tenant, management, and ctrlplane. When this was moved into THT, we accidently dropped external which leads to deployment failures if a service is moved to the external network and the configuration consumes the fqdn_external hiera key. Specifically this is reproduced if the MysqlNetwork is switch to to exernal, then the deployment fails because the bind address which is set to use fqdn_external is blank. Change-Id: I01ad0c14cb3dc38aad7528345c928b86628433c1 Closes-Bug: #1697722
2017-06-13Modify PreNetworkConfig config inline with role-specific parametersSaravanan KR1-0/+5
Existing host_config_and_reboot.role.j2.yaml is done in ocata to configure kernel args. This can be enhanced with use of role-specific parameters, which is done in the current patch. The earlier method is deprecated and will be removed in Q releae. Implements: blueprint ovs-2-6-dpdk Change-Id: Ib864f065527167a49a0f60812d7ad4ad12c836d1
2017-06-08Merge "Use Deployment actions for blacklist"Jenkins1-5/+20
2017-06-07Use Deployment actions for blacklistJames Slagle1-5/+20
Instead of using the Heat condition directly on the Deployment resources, use it to set the action list to an empty list when the server is blacklisted. This has a couple advantages over the previous approach in that the actual resources are not deleted and recreated when servers are added and removed from the blacklist. Recreating the resources can be problematic, as it would then force the Deployments to re-run when a server is removed from the blacklist. That is likely not always desirable, especially in the case of NetworkDeloyment. Additionally, you will still see the resources for a blacklisted server in the stack, just with an empty set of actions. This has the benefit of preserving the history of the previous time the Deployment was triggered. implements blueprint disable-deployments Change-Id: I3d0263a6319ae4871b1ae11383ae838bd2540d36
2017-06-06Convert puppet and docker steps to ansibleSteven Hardy1-0/+1
Replace the multiple SoftwareDeployment resources with a common playbook that runs on all roles, consuming the configuration data written via the HostPrepAnsible tasks. This hopefully simplifies things, and will enable re-running the deploy steps for minor updates (we'll need some way to detect a container should be replaced, but that will be done via a follow-up patch). Change-Id: I674a4d9d2c77d1f6fbdb0996f6c9321848e32662
2017-05-31Server blacklist supportJames Slagle1-0/+20
Adds the ability to blacklist servers from all SoftwareDeployment resources. The servers are specified in a new list parameter, DeploymentServerBlacklist by the Heat assigned name (overcloud-compute-0, etc). implements blueprint disable-deployments Change-Id: I46941e54a476c7cc8645cd1aff391c9c6c5434de
2017-05-19Update the template_version alias for all the templates to pike.Carlos Camacho1-1/+1
Master is now the development branch for pike changing the release alias name. Change-Id: I938e4a983e361aefcaa0bd9a4226c296c5823127
2017-05-18Merge "Enable splay for os-collect-config"Jenkins1-0/+10
2017-05-01Enable splay for os-collect-configAlex Schultz1-0/+10
At scale, having the os-collect-config instances all check in at the same time can cause performance problems. This change enables splay and sets it to a default maximum random sleep of 30 seconds prior to the os-collect-config polling. Change-Id: Iab8b51f4e5fb4727b8aa7e081f5cbfcbf11f7fcb Depends-On: I88f623c9e8db9ed4a186918206a63faec8f7f673 Closes-Bug: #1677314
2017-04-20Merge "Pluggable server type per Role"Jenkins1-1/+1
2017-04-18Merge "SSH known_hosts config"Jenkins1-0/+37
2017-04-13SSH known_hosts configOliver Walsh1-0/+37
Fetch the host public keys from each node, combine them all and write to the system-wide ssh known hosts. The alternative of disabling host key verification is vulnerable to a MITM attack. Change-Id: Ib572b5910720b1991812256e68c975f7fbe2239c
2017-04-13Pluggable server type per RoleJames Slagle1-1/+1
The server resource type, OS::TripleO::Server can now be mapped per role instead of globally. This allows users to mix baremetal (OS::Nova::Server) and deployed-server (OS::Heat::DeployedServer) server resources in the same deployment. blueprint pluggable-server-type-per-role Change-Id: Ib9e9abe2ba5103db221f0b485c46704b1e260dbf
2017-04-10Add composable role support for NetApp Cinder back endAlan Bishop1-1/+0
Convert NetApp Cinder back end to support composable roles via new "CinderBackendNetApp" service. Closes-Bug: #1680568 Change-Id: Ia3a78a48c32997c9d3cbe1629c2043cfc5249e1c
2017-03-01Make UpdateDeployment depend on NetworkDeploymentSteven Hardy1-0/+1
Prior to https://review.openstack.org/#/c/271450/ os-net-config was applied via os-refresh-config directly, which meant that even though UpdateDeployment and NetworkDeployment can be created concurrently, we'd always do the os-net-config step first. However now that we apply both steps via scripts (which are both handled via the same heat-config hook) we should add an explicit dependency to ensure the network is always fully configured before attempting to run any update. This should avoid the risk of e.g running an update on initial deployment before the network connectivity to access yum repos is in place. Change-Id: Idff7a95afe7b49b6384b1d0c78e76522fb1f8eb7 Related-Bug: #1666227
2017-02-16Add Newton to Ocata UpgradeInitCommonCommandmarios1-0/+9
This adds the UpgradeInitCommonCommand for newton..ocata common UpgradeInit commands. This comes before the ansible upgrade steps so we need to do things like remove the old newton hieradata and install the ansible-pacemaker module and ansible heat-agent plugin This defaults to '' and is set in the major-upgrade-composable-steps and unset in the major-upgrade-converge environment files. Change-Id: I0c7a32194c0069b63a501a913c17907b47c9cc16
2017-02-08Merge "Composable service support for Cinder Dell EMC Storage Center"Jenkins1-1/+0
2017-02-07Composable services support for Cinder Dell EMC PS Seriesrajinir1-1/+0
Updated the heat templates for Cinder Dell EMC PS Series backend to use composable services and rebranding of EQLX to Dell EMC PS Series Closes-Bug: #1661313 Change-Id: Id9d6f172f3f79a31788b26c7776d738fda5a30fa
2017-02-03Composable service support for Cinder Dell EMC Storage Centerrajinir1-1/+0
Updated the heat templates for Cinder Dell EMC Storage Center Backend to use composable services Closes-Bug: #1661314 Change-Id: I454549c45da7388f0e42975c9f4637dde9ec51e3
2017-01-04Merge "Add pre-network hook and example showing config-then-reboot"Jenkins1-0/+6
2016-12-23Bump template version for all templates to "ocata"Steven Hardy1-1/+1
Heat now supports release name aliases, so we can replace the inconsistent mix of date related versions with one consistent version that aligns with the supported version of heat for this t-h-t branch. This should also help new users who sometimes copy/paste old templates and discover intrinsic functions in the t-h-t docs don't work because their template version is too old. Change-Id: Ib415e7290fea27447460baa280291492df197e54
2016-12-22Merge "Introduce role-specific NodeUserData, use for docker"Jenkins1-0/+7
2016-12-22Add hook to generate metadata from service profilesJuan Antonio Osorio Robles1-0/+4
This enables the deployer to dynamically add nova metadata to the servers based on the output of service profiles that implement the metadata_settings key in the role_data output for the profiles. One can set an implementation via the OS::TripleO::ServerMetadataHook resource, which currently is set as OS::Heat::None. So, because of the default implementation, if left untouched it actually does nothing. Currently, besides the list, which is metadata_settings, this hook also takes the name of the node that it's setting the metadata for. This is useful for nova vendordata plugins that can parse said metadata. Change-Id: I8a937f711f0b90156fbb6c4632760435ef846474
2016-12-19Introduce role-specific NodeUserData, use for dockerSteve Baker1-0/+7
Currently when the docker environments are invoked, every node has the boot script run which replaces os-collect-config with the heat-agents container. This should only be happening on Compute nodes currently, and each role will be converted to heat-agents one at a time. This change implements a role-specific NodeUserData resource and uses that mechanism to run docker/firstboot/install_docker_agents.yaml only on Compute nodes. Change-Id: Id81811dbcaf0e661c3980aa25f3ca80db5ef0954
2016-12-19Move UpgradeInitCommand to role templatesSteven Hardy1-1/+27
We can't run this during the upgrade steps, because there are things which need to happen before any role configuration happens, e.g installing the new hiera heat-config hook, which must be done before e.g "ControllerDeployment" runs or the stack update hangs. Partially-Implements: blueprint overcloud-upgrades-per-service Change-Id: I365b57513590662c3f78a33dc625747f457c48c5
2016-12-16Introduce role-specific nova-server-metadataJuan Antonio Osorio Robles1-2/+14
We could already pass metadata to the nova server instances (on creation) via the ServerMetadata parameter, however, there was no way of doing this per-role. This introduces that by adding a {{role}}ServerMetadata parameter for each role. This parameter gets merged with the ServerMetadata parameter and allows this functionality. Note that both default to {}, and so does the result of merging those parameters with their default values. So nothing changes for the default settings. Change-Id: I334edcc51ce7ee82fc13b6cf4c0d74ccb7db099c
2016-12-15Add pre-network hook and example showing config-then-rebootSteven Hardy1-0/+6
There are some requirements for early configuration that involves e.g setting kernel parameters then rebooting. Currently this can be done via cloud-init, e.g firstboot templates, but there's been discussion around enabling a SoftwareDeployment approach instead. The main advantage of doing it this way is there's an error path if something goes wrong with the config (except triggering the reboot as we have to use NO_SIGNAL for that). Change-Id: Ia54ee654f755631b8062eb5c209a60c6f9161500