summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorTim Rozet <trozet@redhat.com>2016-11-22 10:53:05 -0500
committerTim Rozet <trozet@redhat.com>2016-11-22 11:27:42 -0500
commit052e43fa14e84228f6b66609c8aa5758df1493f0 (patch)
treea95b5178b6da33737990be18c41bc318c723f20c
parenta75a950bd921bf4db78138fac6bca3cc17c13e44 (diff)
Fixes broken puppet-keystone calls to openstackclient
Issue where openstackclient is trying to parse keystone options, and os-client-config is converting 'url' to 'endpoint' which openstackclient does not support. Also updates mitaka delorean repo. JIRA: APEX-349 Change-Id: I137a069c297d7d00fed64746b5de261d7b8f7d93 Signed-off-by: Tim Rozet <trozet@redhat.com>
-rw-r--r--build/os-client-config/config.py1294
-rwxr-xr-xbuild/overcloud-full.sh5
-rwxr-xr-xbuild/undercloud.sh4
3 files changed, 1303 insertions, 0 deletions
diff --git a/build/os-client-config/config.py b/build/os-client-config/config.py
new file mode 100644
index 00000000..f684dfb7
--- /dev/null
+++ b/build/os-client-config/config.py
@@ -0,0 +1,1294 @@
+# Copyright (c) 2014 Hewlett-Packard Development Company, L.P.
+#
+# Licensed under the Apache License, Version 2.0 (the "License"); you may
+# not use this file except in compliance with the License. You may obtain
+# a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+# License for the specific language governing permissions and limitations
+# under the License.
+
+
+# alias because we already had an option named argparse
+import argparse as argparse_mod
+import collections
+import copy
+import json
+import os
+import sys
+import warnings
+
+import appdirs
+from keystoneauth1 import adapter
+from keystoneauth1 import loading
+import yaml
+
+from os_client_config import _log
+from os_client_config import cloud_config
+from os_client_config import defaults
+from os_client_config import exceptions
+from os_client_config import vendors
+
+APPDIRS = appdirs.AppDirs('openstack', 'OpenStack', multipath='/etc')
+CONFIG_HOME = APPDIRS.user_config_dir
+CACHE_PATH = APPDIRS.user_cache_dir
+
+UNIX_CONFIG_HOME = os.path.join(
+ os.path.expanduser(os.path.join('~', '.config')), 'openstack')
+UNIX_SITE_CONFIG_HOME = '/etc/openstack'
+
+SITE_CONFIG_HOME = APPDIRS.site_config_dir
+
+CONFIG_SEARCH_PATH = [
+ os.getcwd(),
+ CONFIG_HOME, UNIX_CONFIG_HOME,
+ SITE_CONFIG_HOME, UNIX_SITE_CONFIG_HOME
+]
+YAML_SUFFIXES = ('.yaml', '.yml')
+JSON_SUFFIXES = ('.json',)
+CONFIG_FILES = [
+ os.path.join(d, 'clouds' + s)
+ for d in CONFIG_SEARCH_PATH
+ for s in YAML_SUFFIXES + JSON_SUFFIXES
+]
+SECURE_FILES = [
+ os.path.join(d, 'secure' + s)
+ for d in CONFIG_SEARCH_PATH
+ for s in YAML_SUFFIXES + JSON_SUFFIXES
+]
+VENDOR_FILES = [
+ os.path.join(d, 'clouds-public' + s)
+ for d in CONFIG_SEARCH_PATH
+ for s in YAML_SUFFIXES + JSON_SUFFIXES
+]
+
+BOOL_KEYS = ('insecure', 'cache')
+
+
+# NOTE(dtroyer): This turns out to be not the best idea so let's move
+# overriding defaults to a kwarg to OpenStackConfig.__init__()
+# Remove this sometime in June 2015 once OSC is comfortably
+# changed-over and global-defaults is updated.
+def set_default(key, value):
+ warnings.warn(
+ "Use of set_default() is deprecated. Defaults should be set with the "
+ "`override_defaults` parameter of OpenStackConfig."
+ )
+ defaults.get_defaults() # make sure the dict is initialized
+ defaults._defaults[key] = value
+
+
+def get_boolean(value):
+ if value is None:
+ return False
+ if type(value) is bool:
+ return value
+ if value.lower() == 'true':
+ return True
+ return False
+
+
+def _get_os_environ(envvar_prefix=None):
+ ret = defaults.get_defaults()
+ if not envvar_prefix:
+ # This makes the or below be OS_ or OS_ which is a no-op
+ envvar_prefix = 'OS_'
+ environkeys = [k for k in os.environ.keys()
+ if (k.startswith('OS_') or k.startswith(envvar_prefix))
+ and not k.startswith('OS_TEST') # infra CI var
+ and not k.startswith('OS_STD') # infra CI var
+ ]
+ for k in environkeys:
+ newkey = k.split('_', 1)[-1].lower()
+ ret[newkey] = os.environ[k]
+ # If the only environ keys are cloud and region_name, don't return anything
+ # because they are cloud selectors
+ if set(environkeys) - set(['OS_CLOUD', 'OS_REGION_NAME']):
+ return ret
+ return None
+
+
+def _merge_clouds(old_dict, new_dict):
+ """Like dict.update, except handling nested dicts."""
+ ret = old_dict.copy()
+ for (k, v) in new_dict.items():
+ if isinstance(v, dict):
+ if k in ret:
+ ret[k] = _merge_clouds(ret[k], v)
+ else:
+ ret[k] = v.copy()
+ else:
+ ret[k] = v
+ return ret
+
+
+def _auth_update(old_dict, new_dict_source):
+ """Like dict.update, except handling the nested dict called auth."""
+ new_dict = copy.deepcopy(new_dict_source)
+ for (k, v) in new_dict.items():
+ if k == 'auth':
+ if k in old_dict:
+ old_dict[k].update(v)
+ else:
+ old_dict[k] = v.copy()
+ else:
+ old_dict[k] = v
+ return old_dict
+
+
+def _fix_argv(argv):
+ # Transform any _ characters in arg names to - so that we don't
+ # have to throw billions of compat argparse arguments around all
+ # over the place.
+ processed = collections.defaultdict(list)
+ for index in range(0, len(argv)):
+ if argv[index].startswith('--'):
+ split_args = argv[index].split('=')
+ orig = split_args[0]
+ new = orig.replace('_', '-')
+ if orig != new:
+ split_args[0] = new
+ argv[index] = "=".join(split_args)
+ # Save both for later so we can throw an error about dupes
+ processed[new].append(orig)
+ overlap = []
+ for new, old in processed.items():
+ if len(old) > 1:
+ overlap.extend(old)
+ if overlap:
+ raise exceptions.OpenStackConfigException(
+ "The following options were given: '{options}' which contain"
+ " duplicates except that one has _ and one has -. There is"
+ " no sane way for us to know what you're doing. Remove the"
+ " duplicate option and try again".format(
+ options=','.join(overlap)))
+
+
+class OpenStackConfig(object):
+
+ def __init__(self, config_files=None, vendor_files=None,
+ override_defaults=None, force_ipv4=None,
+ envvar_prefix=None, secure_files=None,
+ pw_func=None, session_constructor=None):
+ self.log = _log.setup_logging(__name__)
+ self._session_constructor = session_constructor
+
+ self._config_files = config_files or CONFIG_FILES
+ self._secure_files = secure_files or SECURE_FILES
+ self._vendor_files = vendor_files or VENDOR_FILES
+
+ config_file_override = os.environ.pop('OS_CLIENT_CONFIG_FILE', None)
+ if config_file_override:
+ self._config_files.insert(0, config_file_override)
+
+ secure_file_override = os.environ.pop('OS_CLIENT_SECURE_FILE', None)
+ if secure_file_override:
+ self._secure_files.insert(0, secure_file_override)
+
+ self.defaults = defaults.get_defaults()
+ if override_defaults:
+ self.defaults.update(override_defaults)
+
+ # First, use a config file if it exists where expected
+ self.config_filename, self.cloud_config = self._load_config_file()
+ _, secure_config = self._load_secure_file()
+ if secure_config:
+ self.cloud_config = _merge_clouds(
+ self.cloud_config, secure_config)
+
+ if not self.cloud_config:
+ self.cloud_config = {'clouds': {}}
+ if 'clouds' not in self.cloud_config:
+ self.cloud_config['clouds'] = {}
+
+ # Grab ipv6 preference settings from env
+ client_config = self.cloud_config.get('client', {})
+
+ if force_ipv4 is not None:
+ # If it's passed in to the constructor, honor it.
+ self.force_ipv4 = force_ipv4
+ else:
+ # Get the backwards compat value
+ prefer_ipv6 = get_boolean(
+ os.environ.pop(
+ 'OS_PREFER_IPV6', client_config.get(
+ 'prefer_ipv6', client_config.get(
+ 'prefer-ipv6', True))))
+ force_ipv4 = get_boolean(
+ os.environ.pop(
+ 'OS_FORCE_IPV4', client_config.get(
+ 'force_ipv4', client_config.get(
+ 'broken-ipv6', False))))
+
+ self.force_ipv4 = force_ipv4
+ if not prefer_ipv6:
+ # this will only be false if someone set it explicitly
+ # honor their wishes
+ self.force_ipv4 = True
+
+ # Next, process environment variables and add them to the mix
+ self.envvar_key = os.environ.pop('OS_CLOUD_NAME', 'envvars')
+ if self.envvar_key in self.cloud_config['clouds']:
+ raise exceptions.OpenStackConfigException(
+ '"{0}" defines a cloud named "{1}", but'
+ ' OS_CLOUD_NAME is also set to "{1}". Please rename'
+ ' either your environment based cloud, or one of your'
+ ' file-based clouds.'.format(self.config_filename,
+ self.envvar_key))
+ # Pull out OS_CLOUD so that if it's the only thing set, do not
+ # make an envvars cloud
+ self.default_cloud = os.environ.pop('OS_CLOUD', None)
+
+ envvars = _get_os_environ(envvar_prefix=envvar_prefix)
+ if envvars:
+ self.cloud_config['clouds'][self.envvar_key] = envvars
+ if not self.default_cloud:
+ self.default_cloud = self.envvar_key
+
+ # Finally, fall through and make a cloud that starts with defaults
+ # because we need somewhere to put arguments, and there are neither
+ # config files or env vars
+ if not self.cloud_config['clouds']:
+ self.cloud_config = dict(
+ clouds=dict(defaults=dict(self.defaults)))
+ self.default_cloud = 'defaults'
+
+ self._cache_expiration_time = 0
+ self._cache_path = CACHE_PATH
+ self._cache_class = 'dogpile.cache.null'
+ self._cache_arguments = {}
+ self._cache_expiration = {}
+ if 'cache' in self.cloud_config:
+ cache_settings = self._normalize_keys(self.cloud_config['cache'])
+
+ # expiration_time used to be 'max_age' but the dogpile setting
+ # is expiration_time. Support max_age for backwards compat.
+ self._cache_expiration_time = cache_settings.get(
+ 'expiration_time', cache_settings.get(
+ 'max_age', self._cache_expiration_time))
+
+ # If cache class is given, use that. If not, but if cache time
+ # is given, default to memory. Otherwise, default to nothing.
+ # to memory.
+ if self._cache_expiration_time:
+ self._cache_class = 'dogpile.cache.memory'
+ self._cache_class = self.cloud_config['cache'].get(
+ 'class', self._cache_class)
+
+ self._cache_path = os.path.expanduser(
+ cache_settings.get('path', self._cache_path))
+ self._cache_arguments = cache_settings.get(
+ 'arguments', self._cache_arguments)
+ self._cache_expiration = cache_settings.get(
+ 'expiration', self._cache_expiration)
+
+ # Flag location to hold the peeked value of an argparse timeout value
+ self._argv_timeout = False
+
+ # Save the password callback
+ # password = self._pw_callback(prompt="Password: ")
+ self._pw_callback = pw_func
+
+ def get_extra_config(self, key, defaults=None):
+ """Fetch an arbitrary extra chunk of config, laying in defaults.
+
+ :param string key: name of the config section to fetch
+ :param dict defaults: (optional) default values to merge under the
+ found config
+ """
+ if not defaults:
+ defaults = {}
+ return _merge_clouds(
+ self._normalize_keys(defaults),
+ self._normalize_keys(self.cloud_config.get(key, {})))
+
+ def _load_config_file(self):
+ return self._load_yaml_json_file(self._config_files)
+
+ def _load_secure_file(self):
+ return self._load_yaml_json_file(self._secure_files)
+
+ def _load_vendor_file(self):
+ return self._load_yaml_json_file(self._vendor_files)
+
+ def _load_yaml_json_file(self, filelist):
+ for path in filelist:
+ if os.path.exists(path):
+ with open(path, 'r') as f:
+ if path.endswith('json'):
+ return path, json.load(f)
+ else:
+ return path, yaml.safe_load(f)
+ return (None, {})
+
+ def _normalize_keys(self, config):
+ new_config = {}
+ for key, value in config.items():
+ key = key.replace('-', '_')
+ if isinstance(value, dict):
+ new_config[key] = self._normalize_keys(value)
+ elif isinstance(value, bool):
+ new_config[key] = value
+ elif isinstance(value, int) and key != 'verbose_level':
+ new_config[key] = str(value)
+ elif isinstance(value, float):
+ new_config[key] = str(value)
+ else:
+ new_config[key] = value
+ return new_config
+
+ def get_cache_expiration_time(self):
+ return int(self._cache_expiration_time)
+
+ def get_cache_interval(self):
+ return self.get_cache_expiration_time()
+
+ def get_cache_max_age(self):
+ return self.get_cache_expiration_time()
+
+ def get_cache_path(self):
+ return self._cache_path
+
+ def get_cache_class(self):
+ return self._cache_class
+
+ def get_cache_arguments(self):
+ return copy.deepcopy(self._cache_arguments)
+
+ def get_cache_expiration(self):
+ return copy.deepcopy(self._cache_expiration)
+
+ def _expand_region_name(self, region_name):
+ return {'name': region_name, 'values': {}}
+
+ def _expand_regions(self, regions):
+ ret = []
+ for region in regions:
+ if isinstance(region, dict):
+ ret.append(copy.deepcopy(region))
+ else:
+ ret.append(self._expand_region_name(region))
+ return ret
+
+ def _get_regions(self, cloud):
+ if cloud not in self.cloud_config['clouds']:
+ return [self._expand_region_name('')]
+ regions = self._get_known_regions(cloud)
+ if not regions:
+ # We don't know of any regions use a workable default.
+ regions = [self._expand_region_name('')]
+ return regions
+
+ def _get_known_regions(self, cloud):
+ config = self._normalize_keys(self.cloud_config['clouds'][cloud])
+ if 'regions' in config:
+ return self._expand_regions(config['regions'])
+ elif 'region_name' in config:
+ if isinstance(config['region_name'], list):
+ regions = config['region_name']
+ else:
+ regions = config['region_name'].split(',')
+ if len(regions) > 1:
+ warnings.warn(
+ "Comma separated lists in region_name are deprecated."
+ " Please use a yaml list in the regions"
+ " parameter in {0} instead.".format(self.config_filename))
+ return self._expand_regions(regions)
+ else:
+ # crappit. we don't have a region defined.
+ new_cloud = dict()
+ our_cloud = self.cloud_config['clouds'].get(cloud, dict())
+ self._expand_vendor_profile(cloud, new_cloud, our_cloud)
+ if 'regions' in new_cloud and new_cloud['regions']:
+ return self._expand_regions(new_cloud['regions'])
+ elif 'region_name' in new_cloud and new_cloud['region_name']:
+ return [self._expand_region_name(new_cloud['region_name'])]
+
+ def _get_region(self, cloud=None, region_name=''):
+ if region_name is None:
+ region_name = ''
+ if not cloud:
+ return self._expand_region_name(region_name)
+
+ regions = self._get_known_regions(cloud)
+ if not regions:
+ return self._expand_region_name(region_name)
+
+ if not region_name:
+ return regions[0]
+
+ for region in regions:
+ if region['name'] == region_name:
+ return region
+
+ raise exceptions.OpenStackConfigException(
+ 'Region {region_name} is not a valid region name for cloud'
+ ' {cloud}. Valid choices are {region_list}. Please note that'
+ ' region names are case sensitive.'.format(
+ region_name=region_name,
+ region_list=','.join([r['name'] for r in regions]),
+ cloud=cloud))
+
+ def get_cloud_names(self):
+ return self.cloud_config['clouds'].keys()
+
+ def _get_base_cloud_config(self, name):
+ cloud = dict()
+
+ # Only validate cloud name if one was given
+ if name and name not in self.cloud_config['clouds']:
+ raise exceptions.OpenStackConfigException(
+ "Cloud {name} was not found.".format(
+ name=name))
+
+ our_cloud = self.cloud_config['clouds'].get(name, dict())
+
+ # Get the defaults
+ cloud.update(self.defaults)
+ self._expand_vendor_profile(name, cloud, our_cloud)
+
+ if 'auth' not in cloud:
+ cloud['auth'] = dict()
+
+ _auth_update(cloud, our_cloud)
+ if 'cloud' in cloud:
+ del cloud['cloud']
+
+ return cloud
+
+ def _expand_vendor_profile(self, name, cloud, our_cloud):
+ # Expand a profile if it exists. 'cloud' is an old confusing name
+ # for this.
+ profile_name = our_cloud.get('profile', our_cloud.get('cloud', None))
+ if profile_name and profile_name != self.envvar_key:
+ if 'cloud' in our_cloud:
+ warnings.warn(
+ "{0} use the keyword 'cloud' to reference a known "
+ "vendor profile. This has been deprecated in favor of the "
+ "'profile' keyword.".format(self.config_filename))
+ vendor_filename, vendor_file = self._load_vendor_file()
+ if vendor_file and profile_name in vendor_file['public-clouds']:
+ _auth_update(cloud, vendor_file['public-clouds'][profile_name])
+ else:
+ profile_data = vendors.get_profile(profile_name)
+ if profile_data:
+ status = profile_data.pop('status', 'active')
+ message = profile_data.pop('message', '')
+ if status == 'deprecated':
+ warnings.warn(
+ "{profile_name} is deprecated: {message}".format(
+ profile_name=profile_name, message=message))
+ elif status == 'shutdown':
+ raise exceptions.OpenStackConfigException(
+ "{profile_name} references a cloud that no longer"
+ " exists: {message}".format(
+ profile_name=profile_name, message=message))
+ _auth_update(cloud, profile_data)
+ else:
+ # Can't find the requested vendor config, go about business
+ warnings.warn("Couldn't find the vendor profile '{0}', for"
+ " the cloud '{1}'".format(profile_name,
+ name))
+
+ def _project_scoped(self, cloud):
+ return ('project_id' in cloud or 'project_name' in cloud
+ or 'project_id' in cloud['auth']
+ or 'project_name' in cloud['auth'])
+
+ def _validate_networks(self, networks, key):
+ value = None
+ for net in networks:
+ if value and net[key]:
+ raise exceptions.OpenStackConfigException(
+ "Duplicate network entries for {key}: {net1} and {net2}."
+ " Only one network can be flagged with {key}".format(
+ key=key,
+ net1=value['name'],
+ net2=net['name']))
+ if not value and net[key]:
+ value = net
+
+ def _fix_backwards_networks(self, cloud):
+ # Leave the external_network and internal_network keys in the
+ # dict because consuming code might be expecting them.
+ networks = []
+ # Normalize existing network entries
+ for net in cloud.get('networks', []):
+ name = net.get('name')
+ if not name:
+ raise exceptions.OpenStackConfigException(
+ 'Entry in network list is missing required field "name".')
+ network = dict(
+ name=name,
+ routes_externally=get_boolean(net.get('routes_externally')),
+ nat_destination=get_boolean(net.get('nat_destination')),
+ default_interface=get_boolean(net.get('default_interface')),
+ )
+ # routes_ipv4_externally defaults to the value of routes_externally
+ network['routes_ipv4_externally'] = get_boolean(
+ net.get(
+ 'routes_ipv4_externally', network['routes_externally']))
+ # routes_ipv6_externally defaults to the value of routes_externally
+ network['routes_ipv6_externally'] = get_boolean(
+ net.get(
+ 'routes_ipv6_externally', network['routes_externally']))
+ networks.append(network)
+
+ for key in ('external_network', 'internal_network'):
+ external = key.startswith('external')
+ if key in cloud and 'networks' in cloud:
+ raise exceptions.OpenStackConfigException(
+ "Both {key} and networks were specified in the config."
+ " Please remove {key} from the config and use the network"
+ " list to configure network behavior.".format(key=key))
+ if key in cloud:
+ warnings.warn(
+ "{key} is deprecated. Please replace with an entry in"
+ " a dict inside of the networks list with name: {name}"
+ " and routes_externally: {external}".format(
+ key=key, name=cloud[key], external=external))
+ networks.append(dict(
+ name=cloud[key],
+ routes_externally=external,
+ nat_destination=not external,
+ default_interface=external))
+
+ # Validate that we don't have duplicates
+ self._validate_networks(networks, 'nat_destination')
+ self._validate_networks(networks, 'default_interface')
+
+ cloud['networks'] = networks
+ return cloud
+
+ def _handle_domain_id(self, cloud):
+ # Allow people to just specify domain once if it's the same
+ mappings = {
+ 'domain_id': ('user_domain_id', 'project_domain_id'),
+ 'domain_name': ('user_domain_name', 'project_domain_name'),
+ }
+ for target_key, possible_values in mappings.items():
+ if not self._project_scoped(cloud):
+ if target_key in cloud and target_key not in cloud['auth']:
+ cloud['auth'][target_key] = cloud.pop(target_key)
+ continue
+ for key in possible_values:
+ if target_key in cloud['auth'] and key not in cloud['auth']:
+ cloud['auth'][key] = cloud['auth'][target_key]
+ cloud.pop(target_key, None)
+ cloud['auth'].pop(target_key, None)
+ return cloud
+
+ def _fix_backwards_project(self, cloud):
+ # Do the lists backwards so that project_name is the ultimate winner
+ # Also handle moving domain names into auth so that domain mapping
+ # is easier
+ mappings = {
+ 'domain_id': ('domain_id', 'domain-id'),
+ 'domain_name': ('domain_name', 'domain-name'),
+ 'user_domain_id': ('user_domain_id', 'user-domain-id'),
+ 'user_domain_name': ('user_domain_name', 'user-domain-name'),
+ 'project_domain_id': ('project_domain_id', 'project-domain-id'),
+ 'project_domain_name': (
+ 'project_domain_name', 'project-domain-name'),
+ 'token': ('auth-token', 'auth_token', 'token'),
+ }
+ if cloud.get('auth_type', None) == 'v2password':
+ # If v2password is explcitly requested, this is to deal with old
+ # clouds. That's fine - we need to map settings in the opposite
+ # direction
+ mappings['tenant_id'] = (
+ 'project_id', 'project-id', 'tenant_id', 'tenant-id')
+ mappings['tenant_name'] = (
+ 'project_name', 'project-name', 'tenant_name', 'tenant-name')
+ else:
+ mappings['project_id'] = (
+ 'tenant_id', 'tenant-id', 'project_id', 'project-id')
+ mappings['project_name'] = (
+ 'tenant_name', 'tenant-name', 'project_name', 'project-name')
+ for target_key, possible_values in mappings.items():
+ target = None
+ for key in possible_values:
+ if key in cloud:
+ target = str(cloud[key])
+ del cloud[key]
+ if key in cloud['auth']:
+ target = str(cloud['auth'][key])
+ del cloud['auth'][key]
+ if target:
+ cloud['auth'][target_key] = target
+ return cloud
+
+ def _fix_backwards_auth_plugin(self, cloud):
+ # Do the lists backwards so that auth_type is the ultimate winner
+ mappings = {
+ 'auth_type': ('auth_plugin', 'auth_type'),
+ }
+ for target_key, possible_values in mappings.items():
+ target = None
+ for key in possible_values:
+ if key in cloud:
+ target = cloud[key]
+ del cloud[key]
+ cloud[target_key] = target
+ # Because we force alignment to v3 nouns, we want to force
+ # use of the auth plugin that can do auto-selection and dealing
+ # with that based on auth parameters. v2password is basically
+ # completely broken
+ return cloud
+
+ def register_argparse_arguments(self, parser, argv, service_keys=None):
+ """Register all of the common argparse options needed.
+
+ Given an argparse parser, register the keystoneauth Session arguments,
+ the keystoneauth Auth Plugin Options and os-cloud. Also, peek in the
+ argv to see if all of the auth plugin options should be registered
+ or merely the ones already configured.
+ :param argparse.ArgumentParser: parser to attach argparse options to
+ :param list argv: the arguments provided to the application
+ :param string service_keys: Service or list of services this argparse
+ should be specialized for, if known.
+ The first item in the list will be used
+ as the default value for service_type
+ (optional)
+
+ :raises exceptions.OpenStackConfigException if an invalid auth-type
+ is requested
+ """
+
+ if service_keys is None:
+ service_keys = []
+
+ # Fix argv in place - mapping any keys with embedded _ in them to -
+ _fix_argv(argv)
+
+ local_parser = argparse_mod.ArgumentParser(add_help=False)
+
+ for p in (parser, local_parser):
+ p.add_argument(
+ '--os-cloud',
+ metavar='<name>',
+ default=os.environ.get('OS_CLOUD', None),
+ help='Named cloud to connect to')
+
+ # we need to peek to see if timeout was actually passed, since
+ # the keystoneauth declaration of it has a default, which means
+ # we have no clue if the value we get is from the ksa default
+ # for from the user passing it explicitly. We'll stash it for later
+ local_parser.add_argument('--timeout', metavar='<timeout>')
+
+ # We need for get_one_cloud to be able to peek at whether a token
+ # was passed so that we can swap the default from password to
+ # token if it was. And we need to also peek for --os-auth-token
+ # for novaclient backwards compat
+ local_parser.add_argument('--os-token')
+ local_parser.add_argument('--os-auth-token')
+
+ # Peek into the future and see if we have an auth-type set in
+ # config AND a cloud set, so that we know which command line
+ # arguments to register and show to the user (the user may want
+ # to say something like:
+ # openstack --os-cloud=foo --os-oidctoken=bar
+ # although I think that user is the cause of my personal pain
+ options, _args = local_parser.parse_known_args(argv)
+ if options.timeout:
+ self._argv_timeout = True
+
+ # validate = False because we're not _actually_ loading here
+ # we're only peeking, so it's the wrong time to assert that
+ # the rest of the arguments given are invalid for the plugin
+ # chosen (for instance, --help may be requested, so that the
+ # user can see what options he may want to give
+ cloud = self.get_one_cloud(argparse=options, validate=False)
+ default_auth_type = cloud.config['auth_type']
+
+ try:
+ loading.register_auth_argparse_arguments(
+ parser, argv, default=default_auth_type)
+ except Exception:
+ # Hidiing the keystoneauth exception because we're not actually
+ # loading the auth plugin at this point, so the error message
+ # from it doesn't actually make sense to os-client-config users
+ options, _args = parser.parse_known_args(argv)
+ plugin_names = loading.get_available_plugin_names()
+ raise exceptions.OpenStackConfigException(
+ "An invalid auth-type was specified: {auth_type}."
+ " Valid choices are: {plugin_names}.".format(
+ auth_type=options.os_auth_type,
+ plugin_names=",".join(plugin_names)))
+
+ if service_keys:
+ primary_service = service_keys[0]
+ else:
+ primary_service = None
+ loading.register_session_argparse_arguments(parser)
+ adapter.register_adapter_argparse_arguments(
+ parser, service_type=primary_service)
+ for service_key in service_keys:
+ # legacy clients have un-prefixed api-version options
+ parser.add_argument(
+ '--{service_key}-api-version'.format(
+ service_key=service_key.replace('_', '-'),
+ help=argparse_mod.SUPPRESS))
+ adapter.register_service_adapter_argparse_arguments(
+ parser, service_type=service_key)
+
+ # Backwards compat options for legacy clients
+ parser.add_argument('--http-timeout', help=argparse_mod.SUPPRESS)
+ parser.add_argument('--os-endpoint-type', help=argparse_mod.SUPPRESS)
+ parser.add_argument('--endpoint-type', help=argparse_mod.SUPPRESS)
+
+ def _fix_backwards_interface(self, cloud):
+ new_cloud = {}
+ for key in cloud.keys():
+ if key.endswith('endpoint_type'):
+ target_key = key.replace('endpoint_type', 'interface')
+ else:
+ target_key = key
+ new_cloud[target_key] = cloud[key]
+ return new_cloud
+
+ def _fix_backwards_api_timeout(self, cloud):
+ new_cloud = {}
+ # requests can only have one timeout, which means that in a single
+ # cloud there is no point in different timeout values. However,
+ # for some reason many of the legacy clients decided to shove their
+ # service name in to the arg name for reasons surpassin sanity. If
+ # we find any values that are not api_timeout, overwrite api_timeout
+ # with the value
+ service_timeout = None
+ for key in cloud.keys():
+ if key.endswith('timeout') and not (
+ key == 'timeout' or key == 'api_timeout'):
+ service_timeout = cloud[key]
+ else:
+ new_cloud[key] = cloud[key]
+ if service_timeout is not None:
+ new_cloud['api_timeout'] = service_timeout
+ # The common argparse arg from keystoneauth is called timeout, but
+ # os-client-config expects it to be called api_timeout
+ if self._argv_timeout:
+ if 'timeout' in new_cloud and new_cloud['timeout']:
+ new_cloud['api_timeout'] = new_cloud.pop('timeout')
+ return new_cloud
+
+ def get_all_clouds(self):
+
+ clouds = []
+
+ for cloud in self.get_cloud_names():
+ for region in self._get_regions(cloud):
+ if region:
+ clouds.append(self.get_one_cloud(
+ cloud, region_name=region['name']))
+ return clouds
+
+ def _fix_args(self, args=None, argparse=None):
+ """Massage the passed-in options
+
+ Replace - with _ and strip os_ prefixes.
+
+ Convert an argparse Namespace object to a dict, removing values
+ that are either None or ''.
+ """
+ if not args:
+ args = {}
+
+ if argparse:
+ # Convert the passed-in Namespace
+ o_dict = vars(argparse)
+ parsed_args = dict()
+ for k in o_dict:
+ if o_dict[k] is not None and o_dict[k] != '':
+ parsed_args[k] = o_dict[k]
+ args.update(parsed_args)
+
+ os_args = dict()
+ new_args = dict()
+ for (key, val) in iter(args.items()):
+ if type(args[key]) == dict:
+ # dive into the auth dict
+ new_args[key] = self._fix_args(args[key])
+ continue
+
+ key = key.replace('-', '_')
+ if key.startswith('os_'):
+ os_args[key[3:]] = val
+ else:
+ new_args[key] = val
+ new_args.update(os_args)
+ return new_args
+
+ def _find_winning_auth_value(self, opt, config):
+ opt_name = opt.name.replace('-', '_')
+ if opt_name in config:
+ return config[opt_name]
+ else:
+ deprecated = getattr(opt, 'deprecated', getattr(
+ opt, 'deprecated_opts', []))
+ for d_opt in deprecated:
+ d_opt_name = d_opt.name.replace('-', '_')
+ if d_opt_name in config:
+ return config[d_opt_name]
+
+ def auth_config_hook(self, config):
+ """Allow examination of config values before loading auth plugin
+
+ OpenStackClient will override this to perform additional checks
+ on auth_type.
+ """
+ return config
+
+ def _get_auth_loader(self, config):
+ # Re-use the admin_token plugin for the "None" plugin
+ # since it does not look up endpoints or tokens but rather
+ # does a passthrough. This is useful for things like Ironic
+ # that have a keystoneless operational mode, but means we're
+ # still dealing with a keystoneauth Session object, so all the
+ # _other_ things (SSL arg handling, timeout) all work consistently
+ if config['auth_type'] in (None, "None", ''):
+ config['auth_type'] = 'admin_token'
+ # Set to notused rather than None because validate_auth will
+ # strip the value if it's actually python None
+ config['auth']['token'] = 'notused'
+ elif config['auth_type'] == 'token_endpoint':
+ # Humans have been trained to use a thing called token_endpoint
+ # That it does not exist in keystoneauth is irrelvant- it not
+ # doing what they want causes them sorrow.
+ config['auth_type'] = 'admin_token'
+ return loading.get_plugin_loader(config['auth_type'])
+
+ def _validate_auth_ksc(self, config, cloud):
+ try:
+ import keystoneclient.auth as ksc_auth
+ except ImportError:
+ return config
+
+ # May throw a keystoneclient.exceptions.NoMatchingPlugin
+ plugin_options = ksc_auth.get_plugin_class(
+ config['auth_type']).get_options()
+
+ for p_opt in plugin_options:
+ # if it's in config.auth, win, kill it from config dict
+ # if it's in config and not in config.auth, move it
+ # deprecated loses to current
+ # provided beats default, deprecated or not
+ winning_value = self._find_winning_auth_value(
+ p_opt,
+ config['auth'],
+ )
+ if not winning_value:
+ winning_value = self._find_winning_auth_value(
+ p_opt,
+ config,
+ )
+
+ # if the plugin tells us that this value is required
+ # then error if it's doesn't exist now
+ if not winning_value and p_opt.required:
+ raise exceptions.OpenStackConfigException(
+ 'Unable to find auth information for cloud'
+ ' {cloud} in config files {files}'
+ ' or environment variables. Missing value {auth_key}'
+ ' required for auth plugin {plugin}'.format(
+ cloud=cloud, files=','.join(self._config_files),
+ auth_key=p_opt.name, plugin=config.get('auth_type')))
+
+ # Clean up after ourselves
+ for opt in [p_opt.name] + [o.name for o in p_opt.deprecated_opts]:
+ opt = opt.replace('-', '_')
+ config.pop(opt, None)
+ config['auth'].pop(opt, None)
+
+ if winning_value:
+ # Prefer the plugin configuration dest value if the value's key
+ # is marked as depreciated.
+ if p_opt.dest is None:
+ config['auth'][p_opt.name.replace('-', '_')] = (
+ winning_value)
+ else:
+ config['auth'][p_opt.dest] = winning_value
+ if p_opt.dest == 'endpoint':
+ config['auth']['url'] = winning_value
+
+ return config
+
+ def _validate_auth(self, config, loader):
+ # May throw a keystoneauth1.exceptions.NoMatchingPlugin
+
+ plugin_options = loader.get_options()
+
+ for p_opt in plugin_options:
+ # if it's in config.auth, win, kill it from config dict
+ # if it's in config and not in config.auth, move it
+ # deprecated loses to current
+ # provided beats default, deprecated or not
+ winning_value = self._find_winning_auth_value(
+ p_opt,
+ config['auth'],
+ )
+ if not winning_value:
+ winning_value = self._find_winning_auth_value(
+ p_opt,
+ config,
+ )
+
+ config = self._clean_up_after_ourselves(
+ config,
+ p_opt,
+ winning_value,
+ )
+
+ if winning_value:
+ # Prefer the plugin configuration dest value if the value's key
+ # is marked as deprecated.
+ if p_opt.dest is None:
+ good_name = p_opt.name.replace('-', '_')
+ config['auth'][good_name] = winning_value
+ else:
+ config['auth'][p_opt.dest] = winning_value
+ if p_opt.dest == 'endpoint':
+ config['auth']['url'] = winning_value
+
+
+ # See if this needs a prompting
+ config = self.option_prompt(config, p_opt)
+
+ return config
+
+ def _validate_auth_correctly(self, config, loader):
+ # May throw a keystoneauth1.exceptions.NoMatchingPlugin
+
+ plugin_options = loader.get_options()
+
+ for p_opt in plugin_options:
+ # if it's in config, win, move it and kill it from config dict
+ # if it's in config.auth but not in config it's good
+ # deprecated loses to current
+ # provided beats default, deprecated or not
+ winning_value = self._find_winning_auth_value(
+ p_opt,
+ config,
+ )
+ if not winning_value:
+ winning_value = self._find_winning_auth_value(
+ p_opt,
+ config['auth'],
+ )
+
+ config = self._clean_up_after_ourselves(
+ config,
+ p_opt,
+ winning_value,
+ )
+
+ # See if this needs a prompting
+ config = self.option_prompt(config, p_opt)
+
+ return config
+
+ def option_prompt(self, config, p_opt):
+ """Prompt user for option that requires a value"""
+ if (
+ p_opt.prompt is not None and
+ p_opt.dest not in config['auth'] and
+ self._pw_callback is not None
+ ):
+ config['auth'][p_opt.dest] = self._pw_callback(p_opt.prompt)
+ return config
+
+ def _clean_up_after_ourselves(self, config, p_opt, winning_value):
+
+ # Clean up after ourselves
+ for opt in [p_opt.name] + [o.name for o in p_opt.deprecated]:
+ opt = opt.replace('-', '_')
+ config.pop(opt, None)
+ config['auth'].pop(opt, None)
+
+ if winning_value:
+ # Prefer the plugin configuration dest value if the value's key
+ # is marked as depreciated.
+ if p_opt.dest is None:
+ config['auth'][p_opt.name.replace('-', '_')] = (
+ winning_value)
+ else:
+ config['auth'][p_opt.dest] = winning_value
+ return config
+
+ def magic_fixes(self, config):
+ """Perform the set of magic argument fixups"""
+
+ # Infer token plugin if a token was given
+ if (('auth' in config and 'token' in config['auth']) or
+ ('auth_token' in config and config['auth_token']) or
+ ('token' in config and config['token'])):
+ config.setdefault('token', config.pop('auth_token', None))
+
+ # These backwards compat values are only set via argparse. If it's
+ # there, it's because it was passed in explicitly, and should win
+ config = self._fix_backwards_api_timeout(config)
+ if 'endpoint_type' in config:
+ config['interface'] = config.pop('endpoint_type')
+
+ config = self._fix_backwards_auth_plugin(config)
+ config = self._fix_backwards_project(config)
+ config = self._fix_backwards_interface(config)
+ config = self._fix_backwards_networks(config)
+ config = self._handle_domain_id(config)
+
+ for key in BOOL_KEYS:
+ if key in config:
+ if type(config[key]) is not bool:
+ config[key] = get_boolean(config[key])
+
+ # TODO(mordred): Special casing auth_url here. We should
+ # come back to this betterer later so that it's
+ # more generalized
+ if 'auth' in config and 'auth_url' in config['auth']:
+ config['auth']['auth_url'] = config['auth']['auth_url'].format(
+ **config)
+
+ return config
+
+ def get_one_cloud(self, cloud=None, validate=True,
+ argparse=None, **kwargs):
+ """Retrieve a single cloud configuration and merge additional options
+
+ :param string cloud:
+ The name of the configuration to load from clouds.yaml
+ :param boolean validate:
+ Validate the config. Setting this to False causes no auth plugin
+ to be created. It's really only useful for testing.
+ :param Namespace argparse:
+ An argparse Namespace object; allows direct passing in of
+ argparse options to be added to the cloud config. Values
+ of None and '' will be removed.
+ :param region_name: Name of the region of the cloud.
+ :param kwargs: Additional configuration options
+
+ :raises: keystoneauth1.exceptions.MissingRequiredOptions
+ on missing required auth parameters
+ """
+
+ args = self._fix_args(kwargs, argparse=argparse)
+
+ if cloud is None:
+ if 'cloud' in args:
+ cloud = args['cloud']
+ else:
+ cloud = self.default_cloud
+
+ config = self._get_base_cloud_config(cloud)
+
+ # Get region specific settings
+ if 'region_name' not in args:
+ args['region_name'] = ''
+ region = self._get_region(cloud=cloud, region_name=args['region_name'])
+ args['region_name'] = region['name']
+ region_args = copy.deepcopy(region['values'])
+
+ # Regions is a list that we can use to create a list of cloud/region
+ # objects. It does not belong in the single-cloud dict
+ config.pop('regions', None)
+
+ # Can't just do update, because None values take over
+ for arg_list in region_args, args:
+ for (key, val) in iter(arg_list.items()):
+ if val is not None:
+ if key == 'auth' and config[key] is not None:
+ config[key] = _auth_update(config[key], val)
+ else:
+ config[key] = val
+
+ config = self.magic_fixes(config)
+ config = self._normalize_keys(config)
+
+ # NOTE(dtroyer): OSC needs a hook into the auth args before the
+ # plugin is loaded in order to maintain backward-
+ # compatible behaviour
+ config = self.auth_config_hook(config)
+
+ if validate:
+ try:
+ loader = self._get_auth_loader(config)
+ config = self._validate_auth(config, loader)
+ auth_plugin = loader.load_from_options(**config['auth'])
+ except Exception as e:
+ # We WANT the ksa exception normally
+ # but OSC can't handle it right now, so we try deferring
+ # to ksc. If that ALSO fails, it means there is likely
+ # a deeper issue, so we assume the ksa error was correct
+ self.log.debug("Deferring keystone exception: {e}".format(e=e))
+ auth_plugin = None
+ try:
+ config = self._validate_auth_ksc(config, cloud)
+ except Exception:
+ raise e
+ else:
+ auth_plugin = None
+
+ # If any of the defaults reference other values, we need to expand
+ for (key, value) in config.items():
+ if hasattr(value, 'format'):
+ config[key] = value.format(**config)
+
+ force_ipv4 = config.pop('force_ipv4', self.force_ipv4)
+ prefer_ipv6 = config.pop('prefer_ipv6', True)
+ if not prefer_ipv6:
+ force_ipv4 = True
+
+ if cloud is None:
+ cloud_name = ''
+ else:
+ cloud_name = str(cloud)
+ return cloud_config.CloudConfig(
+ name=cloud_name,
+ region=config['region_name'],
+ config=config,
+ force_ipv4=force_ipv4,
+ auth_plugin=auth_plugin,
+ openstack_config=self,
+ session_constructor=self._session_constructor,
+ )
+
+ def get_one_cloud_osc(
+ self,
+ cloud=None,
+ validate=True,
+ argparse=None,
+ **kwargs
+ ):
+ """Retrieve a single cloud configuration and merge additional options
+
+ :param string cloud:
+ The name of the configuration to load from clouds.yaml
+ :param boolean validate:
+ Validate the config. Setting this to False causes no auth plugin
+ to be created. It's really only useful for testing.
+ :param Namespace argparse:
+ An argparse Namespace object; allows direct passing in of
+ argparse options to be added to the cloud config. Values
+ of None and '' will be removed.
+ :param region_name: Name of the region of the cloud.
+ :param kwargs: Additional configuration options
+
+ :raises: keystoneauth1.exceptions.MissingRequiredOptions
+ on missing required auth parameters
+ """
+
+ args = self._fix_args(kwargs, argparse=argparse)
+
+ if cloud is None:
+ if 'cloud' in args:
+ cloud = args['cloud']
+ else:
+ cloud = self.default_cloud
+
+ config = self._get_base_cloud_config(cloud)
+
+ # Get region specific settings
+ if 'region_name' not in args:
+ args['region_name'] = ''
+ region = self._get_region(cloud=cloud, region_name=args['region_name'])
+ args['region_name'] = region['name']
+ region_args = copy.deepcopy(region['values'])
+
+ # Regions is a list that we can use to create a list of cloud/region
+ # objects. It does not belong in the single-cloud dict
+ config.pop('regions', None)
+
+ # Can't just do update, because None values take over
+ for arg_list in region_args, args:
+ for (key, val) in iter(arg_list.items()):
+ if val is not None:
+ if key == 'auth' and config[key] is not None:
+ config[key] = _auth_update(config[key], val)
+ else:
+ config[key] = val
+
+ config = self.magic_fixes(config)
+
+ # NOTE(dtroyer): OSC needs a hook into the auth args before the
+ # plugin is loaded in order to maintain backward-
+ # compatible behaviour
+ config = self.auth_config_hook(config)
+
+ if validate:
+ loader = self._get_auth_loader(config)
+ config = self._validate_auth_correctly(config, loader)
+ auth_plugin = loader.load_from_options(**config['auth'])
+ else:
+ auth_plugin = None
+
+ # If any of the defaults reference other values, we need to expand
+ for (key, value) in config.items():
+ if hasattr(value, 'format'):
+ config[key] = value.format(**config)
+
+ force_ipv4 = config.pop('force_ipv4', self.force_ipv4)
+ prefer_ipv6 = config.pop('prefer_ipv6', True)
+ if not prefer_ipv6:
+ force_ipv4 = True
+
+ if cloud is None:
+ cloud_name = ''
+ else:
+ cloud_name = str(cloud)
+ return cloud_config.CloudConfig(
+ name=cloud_name,
+ region=config['region_name'],
+ config=self._normalize_keys(config),
+ force_ipv4=force_ipv4,
+ auth_plugin=auth_plugin,
+ openstack_config=self,
+ )
+
+ @staticmethod
+ def set_one_cloud(config_file, cloud, set_config=None):
+ """Set a single cloud configuration.
+
+ :param string config_file:
+ The path to the config file to edit. If this file does not exist
+ it will be created.
+ :param string cloud:
+ The name of the configuration to save to clouds.yaml
+ :param dict set_config: Configuration options to be set
+ """
+
+ set_config = set_config or {}
+ cur_config = {}
+ try:
+ with open(config_file) as fh:
+ cur_config = yaml.safe_load(fh)
+ except IOError as e:
+ # Not no such file
+ if e.errno != 2:
+ raise
+ pass
+
+ clouds_config = cur_config.get('clouds', {})
+ cloud_config = _auth_update(clouds_config.get(cloud, {}), set_config)
+ clouds_config[cloud] = cloud_config
+ cur_config['clouds'] = clouds_config
+
+ with open(config_file, 'w') as fh:
+ yaml.safe_dump(cur_config, fh, default_flow_style=False)
+
+if __name__ == '__main__':
+ config = OpenStackConfig().get_all_clouds()
+ for cloud in config:
+ print_cloud = False
+ if len(sys.argv) == 1:
+ print_cloud = True
+ elif len(sys.argv) == 3 and (
+ sys.argv[1] == cloud.name and sys.argv[2] == cloud.region):
+ print_cloud = True
+ elif len(sys.argv) == 2 and (
+ sys.argv[1] == cloud.name):
+ print_cloud = True
+
+ if print_cloud:
+ print(cloud.name, cloud.region, cloud.config)
diff --git a/build/overcloud-full.sh b/build/overcloud-full.sh
index 798b5f43..8ffa2b40 100755
--- a/build/overcloud-full.sh
+++ b/build/overcloud-full.sh
@@ -92,6 +92,9 @@ pushd puppet-tacker > /dev/null
git archive --format=tar.gz --prefix=tacker/ HEAD > ../puppet-tacker.tar.gz
popd > /dev/null
+# replace repo with latest delorean mitaka
+wget https://trunk.rdoproject.org/centos7-mitaka/current-passed-ci/delorean.repo
+
# installing forked opnfv-puppet-tripleo
# enable connection tracking for protocal sctp
# upload dpdk rpms but do not install
@@ -104,6 +107,7 @@ popd > /dev/null
# git clone vsperf into the overcloud image
# upload the tacker puppet module and untar it
LIBGUESTFS_BACKEND=direct virt-customize \
+ --upload delorean.repo:/etc/yum.repos.d/ \
--upload ../opnfv-puppet-tripleo.tar.gz:/etc/puppet/modules \
--run-command "yum update -y python-ipaddress rabbitmq-server erlang*" \
--run-command "if ! rpm -qa | grep python-redis; then yum install -y python-redis; fi" \
@@ -156,6 +160,7 @@ LIBGUESTFS_BACKEND=direct virt-customize \
--upload ../puppet-neutron/lib/puppet/type/neutron_agent_vpp.rb:/etc/puppet/modules/neutron/lib/puppet/type/ \
--mkdir /etc/puppet/modules/neutron/lib/puppet/provider/neutron_agent_vpp \
--upload ../puppet-neutron/lib/puppet/provider/neutron_agent_vpp/ini_setting.rb:/etc/puppet/modules/neutron/lib/puppet/provider/neutron_agent_vpp/ \
+ --upload ../os-client-config/config.py:/usr/lib/python2.7/site-packages/os_client_config/ \
-a overcloud-full_build.qcow2
rm -rf ovs_nsh_patches
diff --git a/build/undercloud.sh b/build/undercloud.sh
index a4d008ee..af1b3455 100755
--- a/build/undercloud.sh
+++ b/build/undercloud.sh
@@ -22,6 +22,9 @@ pushd opnfv-tht > /dev/null
git archive --format=tar.gz --prefix=openstack-tripleo-heat-templates/ HEAD > ../opnfv-tht.tar.gz
popd > /dev/null
+# replace repo with latest delorean mitaka
+wget https://trunk.rdoproject.org/centos7-mitaka/current-passed-ci/delorean.repo
+
pushd images > /dev/null
# installing forked opnfv-tht
# enabling ceph OSDs to live on the controller
@@ -32,6 +35,7 @@ pushd images > /dev/null
# add tacker password to python-tripleoclient
# upload tacker repo and install the client package
LIBGUESTFS_BACKEND=direct virt-customize \
+ --upload ../delorean.repo:/etc/yum.repos.d/ \
--upload ../opnfv-tht.tar.gz:/usr/share \
--run-command "cd /usr/share && rm -rf openstack-tripleo-heat-templates && tar xzf opnfv-tht.tar.gz" \
--run-command "sed -i '/ControllerEnableCephStorage/c\\ ControllerEnableCephStorage: true' /usr/share/openstack-tripleo-heat-templates/environments/storage-environment.yaml" \