diff options
author | Rex Lee <limingjiang@huawei.com> | 2017-07-12 01:35:03 +0000 |
---|---|---|
committer | Gerrit Code Review <gerrit@opnfv.org> | 2017-07-12 01:35:03 +0000 |
commit | 7949507437f8b1b1d9100bd13cf2d90bae60279d (patch) | |
tree | f6030e60eeeebeae7670f2022c06f13296c6bb2a /api/resources/env_action.py | |
parent | 7ee54db221fbb4173de011585a425f0750dd6ccf (diff) | |
parent | 0d18f9d3299480cb43e6b335180e5cbdb58505c3 (diff) |
Merge "Yardstick API architecture improvement"
Diffstat (limited to 'api/resources/env_action.py')
-rw-r--r-- | api/resources/env_action.py | 427 |
1 files changed, 0 insertions, 427 deletions
diff --git a/api/resources/env_action.py b/api/resources/env_action.py deleted file mode 100644 index fed987063..000000000 --- a/api/resources/env_action.py +++ /dev/null @@ -1,427 +0,0 @@ -############################################################################## -# Copyright (c) 2016 Huawei Technologies Co.,Ltd and others. -# -# All rights reserved. This program and the accompanying materials -# are made available under the terms of the Apache License, Version 2.0 -# which accompanies this distribution, and is available at -# http://www.apache.org/licenses/LICENSE-2.0 -############################################################################## -from __future__ import absolute_import - -import errno -import logging -import os -import subprocess -import threading -import time -import uuid -import glob -import yaml -import collections -from subprocess import PIPE - -from six.moves import configparser -from oslo_serialization import jsonutils -from docker import Client - -from api.database.v1.handlers import AsyncTaskHandler -from api.utils import influx -from api.utils.common import result_handler -from yardstick.common import constants as consts -from yardstick.common import utils as common_utils -from yardstick.common import openstack_utils -from yardstick.common.httpClient import HttpClient - - -LOG = logging.getLogger(__name__) -LOG.setLevel(logging.DEBUG) - -async_handler = AsyncTaskHandler() - - -def create_grafana(args): - task_id = str(uuid.uuid4()) - - thread = threading.Thread(target=_create_grafana, args=(task_id,)) - thread.start() - - return result_handler(consts.API_SUCCESS, {'task_id': task_id}) - - -def _create_grafana(task_id): - _create_task(task_id) - - client = Client(base_url=consts.DOCKER_URL) - - try: - LOG.info('Checking if grafana image exist') - image = '{}:{}'.format(consts.GRAFANA_IMAGE, consts.GRAFANA_TAG) - if not _check_image_exist(client, image): - LOG.info('Grafana image not exist, start pulling') - client.pull(consts.GRAFANA_IMAGE, consts.GRAFANA_TAG) - - LOG.info('Createing grafana container') - _create_grafana_container(client) - LOG.info('Grafana container is created') - - time.sleep(5) - - LOG.info('Creating data source for grafana') - _create_data_source() - - LOG.info('Creating dashboard for grafana') - _create_dashboard() - - _update_task_status(task_id) - LOG.info('Finished') - except Exception as e: - _update_task_error(task_id, str(e)) - LOG.exception('Create grafana failed') - - -def _create_dashboard(): - url = 'http://admin:admin@%s:3000/api/dashboards/db' % consts.GRAFANA_IP - path = os.path.join(consts.REPOS_DIR, 'dashboard', '*dashboard.json') - - for i in sorted(glob.iglob(path)): - with open(i) as f: - data = jsonutils.load(f) - try: - HttpClient().post(url, data) - except Exception: - LOG.exception('Create dashboard %s failed', i) - raise - - -def _create_data_source(): - url = 'http://admin:admin@%s:3000/api/datasources' % consts.GRAFANA_IP - data = { - "name": "yardstick", - "type": "influxdb", - "access": "proxy", - "url": "http://%s:8086" % consts.INFLUXDB_IP, - "password": "root", - "user": "root", - "database": "yardstick", - "basicAuth": True, - "basicAuthUser": "admin", - "basicAuthPassword": "admin", - "isDefault": False, - } - try: - HttpClient().post(url, data) - except Exception: - LOG.exception('Create datasources failed') - raise - - -def _create_grafana_container(client): - ports = [3000] - port_bindings = {k: k for k in ports} - restart_policy = {"MaximumRetryCount": 0, "Name": "always"} - host_config = client.create_host_config(port_bindings=port_bindings, - restart_policy=restart_policy) - - LOG.info('Creating container') - container = client.create_container(image='%s:%s' % (consts.GRAFANA_IMAGE, - consts.GRAFANA_TAG), - ports=ports, - detach=True, - tty=True, - host_config=host_config) - LOG.info('Starting container') - client.start(container) - - -def _check_image_exist(client, t): - return any(t in a['RepoTags'][0] for a in client.images() if a['RepoTags']) - - -def create_influxdb(args): - task_id = str(uuid.uuid4()) - - thread = threading.Thread(target=_create_influxdb, args=(task_id,)) - thread.start() - - return result_handler(consts.API_SUCCESS, {'task_id': task_id}) - - -def _create_influxdb(task_id): - _create_task(task_id) - - client = Client(base_url=consts.DOCKER_URL) - - try: - LOG.info('Changing output to influxdb') - _change_output_to_influxdb() - - LOG.info('Checking if influxdb image exist') - if not _check_image_exist(client, '%s:%s' % (consts.INFLUXDB_IMAGE, - consts.INFLUXDB_TAG)): - LOG.info('Influxdb image not exist, start pulling') - client.pull(consts.INFLUXDB_IMAGE, tag=consts.INFLUXDB_TAG) - - LOG.info('Createing influxdb container') - _create_influxdb_container(client) - LOG.info('Influxdb container is created') - - time.sleep(5) - - LOG.info('Config influxdb') - _config_influxdb() - - _update_task_status(task_id) - - LOG.info('Finished') - except Exception as e: - _update_task_error(task_id, str(e)) - LOG.exception('Creating influxdb failed') - - -def _create_influxdb_container(client): - - ports = [8083, 8086] - port_bindings = {k: k for k in ports} - restart_policy = {"MaximumRetryCount": 0, "Name": "always"} - host_config = client.create_host_config(port_bindings=port_bindings, - restart_policy=restart_policy) - - LOG.info('Creating container') - container = client.create_container(image='%s:%s' % (consts.INFLUXDB_IMAGE, - consts.INFLUXDB_TAG), - ports=ports, - detach=True, - tty=True, - host_config=host_config) - LOG.info('Starting container') - client.start(container) - - -def _config_influxdb(): - try: - client = influx.get_data_db_client() - client.create_user(consts.INFLUXDB_USER, - consts.INFLUXDB_PASS, - consts.INFLUXDB_DB_NAME) - client.create_database(consts.INFLUXDB_DB_NAME) - LOG.info('Success to config influxDB') - except Exception: - LOG.exception('Config influxdb failed') - - -def _change_output_to_influxdb(): - common_utils.makedirs(consts.CONF_DIR) - - parser = configparser.ConfigParser() - LOG.info('Reading output sample configuration') - parser.read(consts.CONF_SAMPLE_FILE) - - LOG.info('Set dispatcher to influxdb') - parser.set('DEFAULT', 'dispatcher', 'influxdb') - parser.set('dispatcher_influxdb', 'target', - 'http://%s:8086' % consts.INFLUXDB_IP) - - LOG.info('Writing to %s', consts.CONF_FILE) - with open(consts.CONF_FILE, 'w') as f: - parser.write(f) - - -def prepare_env(args): - task_id = str(uuid.uuid4()) - - thread = threading.Thread(target=_prepare_env_daemon, args=(task_id,)) - thread.start() - - return result_handler(consts.API_SUCCESS, {'task_id': task_id}) - - -def _already_source_openrc(): - """Check if openrc is sourced already""" - return all(os.environ.get(k) for k in ['OS_AUTH_URL', 'OS_USERNAME', - 'OS_PASSWORD', 'EXTERNAL_NETWORK']) - - -def _prepare_env_daemon(task_id): - _create_task(task_id) - - try: - _create_directories() - - rc_file = consts.OPENRC - - LOG.info('Checkout Openrc Environment variable') - if not _already_source_openrc(): - LOG.info('Openrc variable not found in Environment') - if not os.path.exists(rc_file): - LOG.info('Openrc file not found') - installer_ip = os.environ.get('INSTALLER_IP', '192.168.200.2') - installer_type = os.environ.get('INSTALLER_TYPE', 'compass') - LOG.info('Getting openrc file from %s', installer_type) - _get_remote_rc_file(rc_file, installer_ip, installer_type) - LOG.info('Source openrc file') - _source_file(rc_file) - LOG.info('Appending external network') - _append_external_network(rc_file) - LOG.info('Openrc file exist, source openrc file') - _source_file(rc_file) - - LOG.info('Cleaning images') - _clean_images() - - LOG.info('Loading images') - _load_images() - - _update_task_status(task_id) - LOG.info('Finished') - except Exception as e: - _update_task_error(task_id, str(e)) - LOG.exception('Prepare env failed') - - -def _create_directories(): - common_utils.makedirs(consts.CONF_DIR) - - -def _source_file(rc_file): - common_utils.source_env(rc_file) - - -def _get_remote_rc_file(rc_file, installer_ip, installer_type): - - os_fetch_script = os.path.join(consts.RELENG_DIR, consts.FETCH_SCRIPT) - - try: - cmd = [os_fetch_script, '-d', rc_file, '-i', installer_type, - '-a', installer_ip] - p = subprocess.Popen(cmd, stdout=subprocess.PIPE) - p.communicate() - - if p.returncode != 0: - LOG.error('Failed to fetch credentials from installer') - except OSError as e: - if e.errno != errno.EEXIST: - raise - - -def _append_external_network(rc_file): - neutron_client = openstack_utils.get_neutron_client() - networks = neutron_client.list_networks()['networks'] - try: - ext_network = next(n['name'] for n in networks if n['router:external']) - except StopIteration: - LOG.warning("Can't find external network") - else: - cmd = 'export EXTERNAL_NETWORK=%s' % ext_network - try: - with open(rc_file, 'a') as f: - f.write(cmd + '\n') - except OSError as e: - if e.errno != errno.EEXIST: - raise - - -def _clean_images(): - cmd = [consts.CLEAN_IMAGES_SCRIPT] - p = subprocess.Popen(cmd, stdout=subprocess.PIPE, cwd=consts.REPOS_DIR) - output = p.communicate()[0] - LOG.debug(output) - - -def _load_images(): - cmd = [consts.LOAD_IMAGES_SCRIPT] - p = subprocess.Popen(cmd, stdout=subprocess.PIPE, cwd=consts.REPOS_DIR) - output = p.communicate()[0] - LOG.debug(output) - - -def _create_task(task_id): - async_handler.insert({'status': 0, 'task_id': task_id}) - - -def _update_task_status(task_id): - async_handler.update_attr(task_id, {'status': 1}) - - -def _update_task_error(task_id, error): - async_handler.update_attr(task_id, {'status': 2, 'error': error}) - - -def update_openrc(args): - try: - openrc_vars = args['openrc'] - except KeyError: - return result_handler(consts.API_ERROR, 'openrc must be provided') - else: - if not isinstance(openrc_vars, collections.Mapping): - return result_handler(consts.API_ERROR, 'args should be a dict') - - lines = ['export {}={}\n'.format(k, v) for k, v in openrc_vars.items()] - LOG.debug('Writing: %s', ''.join(lines)) - - LOG.info('Writing openrc: Writing') - common_utils.makedirs(consts.CONF_DIR) - - with open(consts.OPENRC, 'w') as f: - f.writelines(lines) - LOG.info('Writing openrc: Done') - - LOG.info('Source openrc: Sourcing') - try: - _source_file(consts.OPENRC) - except Exception as e: - LOG.exception('Failed to source openrc') - return result_handler(consts.API_ERROR, str(e)) - LOG.info('Source openrc: Done') - - return result_handler(consts.API_SUCCESS, {'openrc': openrc_vars}) - - -def upload_pod_file(args): - try: - pod_file = args['file'] - except KeyError: - return result_handler(consts.API_ERROR, 'file must be provided') - - LOG.info('Checking file') - data = yaml.load(pod_file.read()) - if not isinstance(data, collections.Mapping): - return result_handler(consts.API_ERROR, 'invalid yaml file') - - LOG.info('Writing file') - with open(consts.POD_FILE, 'w') as f: - yaml.dump(data, f, default_flow_style=False) - LOG.info('Writing finished') - - return result_handler(consts.API_SUCCESS, {'pod_info': data}) - - -def update_pod_file(args): - try: - pod_dic = args['pod'] - except KeyError: - return result_handler(consts.API_ERROR, 'pod must be provided') - else: - if not isinstance(pod_dic, collections.Mapping): - return result_handler(consts.API_ERROR, 'pod should be a dict') - - LOG.info('Writing file') - with open(consts.POD_FILE, 'w') as f: - yaml.dump(pod_dic, f, default_flow_style=False) - LOG.info('Writing finished') - - return result_handler(consts.API_SUCCESS, {'pod_info': pod_dic}) - - -def update_hosts(hosts_ip): - if not isinstance(hosts_ip, dict): - return result_handler(consts.API_ERROR, 'Error, args should be a dict') - LOG.info('Writing hosts: Writing') - LOG.debug('Writing: %s', hosts_ip) - cmd = ["sudo", "python", "write_hosts.py"] - p = subprocess.Popen(cmd, stdin=PIPE, stdout=PIPE, stderr=PIPE, - cwd = os.path.join(consts.REPOS_DIR, "api/resources")) - _, err = p.communicate(jsonutils.dumps(hosts_ip)) - if p.returncode != 0 : - return result_handler(consts.API_ERROR, err) - LOG.info('Writing hosts: Done') - return result_handler(consts.API_SUCCESS, 'success') |