############################################################################## # Copyright (c) 2016 Huawei Technologies Co.,Ltd and others. # # All rights reserved. This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is available at # http://www.apache.org/licenses/LICENSE-2.0 ############################################################################## from __future__ import absolute_import import errno import logging import os import subprocess import threading import time import uuid import glob import yaml import collections from six.moves import configparser from oslo_serialization import jsonutils from docker import Client from api.database.v1.handlers import AsyncTaskHandler from api.utils import influx from api.utils.common import result_handler from yardstick.common import constants as consts from yardstick.common import utils as common_utils from yardstick.common import openstack_utils from yardstick.common.httpClient import HttpClient LOG = logging.getLogger(__name__) LOG.setLevel(logging.DEBUG) async_handler = AsyncTaskHandler() def create_grafana(args): task_id = str(uuid.uuid4()) thread = threading.Thread(target=_create_grafana, args=(task_id,)) thread.start() return result_handler(consts.API_SUCCESS, {'task_id': task_id}) def _create_grafana(task_id): _create_task(task_id) client = Client(base_url=consts.DOCKER_URL) try: LOG.info('Checking if grafana image exist') image = '{}:{}'.format(consts.GRAFANA_IMAGE, consts.GRAFANA_TAG) if not _check_image_exist(client, image): LOG.info('Grafana image not exist, start pulling') client.pull(consts.GRAFANA_IMAGE, consts.GRAFANA_TAG) LOG.info('Createing grafana container') _create_grafana_container(client) LOG.info('Grafana container is created') time.sleep(5) LOG.info('Creating data source for grafana') _create_data_source() LOG.info('Creating dashboard for grafana') _create_dashboard() _update_task_status(task_id) LOG.info('Finished') except Exception as e: _update_task_error(task_id, str(e)) LOG.exception('Create grafana failed') def _create_dashboard(): url = 'http://admin:admin@%s:3000/api/dashboards/db' % consts.GRAFANA_IP path = os.path.join(consts.REPOS_DIR, 'dashboard', '*dashboard.json') for i in sorted(glob.iglob(path)): with open(i) as f: data = jsonutils.load(f) try: HttpClient().post(url, data) except Exception: LOG.exception('Create dashboard %s failed', i) raise def _create_data_source(): url = 'http://admin:admin@%s:3000/api/datasources' % consts.GRAFANA_IP data = { "name": "yardstick", "type": "influxdb", "access": "proxy", "url": "http://%s:8086" % consts.INFLUXDB_IP, "password": "root", "user": "root", "database": "yardstick", "basicAuth": True, "basicAuthUser": "admin", "basicAuthPassword": "admin", "isDefault": False, } try: HttpClient().post(url, data) except Exception: LOG.exception('Create datasources failed') raise def _create_grafana_container(client): ports = [3000] port_bindings = {k: k for k in ports} restart_policy = {"MaximumRetryCount": 0, "Name": "always"} host_config = client.create_host_config(port_bindings=port_bindings, restart_policy=restart_policy) LOG.info('Creating container') container = client.create_container(image='%s:%s' % (consts.GRAFANA_IMAGE, consts.GRAFANA_TAG), ports=ports, detach=True, tty=True, host_config=host_config) LOG.info('Starting container') client.start(container) def _check_image_exist(client, t): return any(t in a['RepoTags'][0] for a in client.images() if a['RepoTags']) def create_influxdb(args): task_id = str(uuid.uuid4()) thread = threading.Thread(target=_create_influxdb, args=(task_id,)) thread.start() return result_handler(consts.API_SUCCESS, {'task_id': task_id}) def _create_influxdb(task_id): _create_task(task_id) client = Client(base_url=consts.DOCKER_URL) try: LOG.info('Changing output to influxdb') _change_output_to_influxdb() LOG.info('Checking if influxdb image exist') if not _check_image_exist(client, '%s:%s' % (consts.INFLUXDB_IMAGE, consts.INFLUXDB_TAG)): LOG.info('Influxdb image not exist, start pulling') client.pull(consts.INFLUXDB_IMAGE, tag=consts.INFLUXDB_TAG) LOG.info('Createing influxdb container') _create_influxdb_container(client) LOG.info('Influxdb container is created') time.sleep(5) LOG.info('Config influxdb') _config_influxdb() _update_task_status(task_id) LOG.info('Finished') except Exception as e: _update_task_error(task_id, str(e)) LOG.exception('Creating influxdb failed') def _create_influxdb_container(client): ports = [8083, 8086] port_bindings = {k: k for k in ports} restart_policy = {"MaximumRetryCount": 0, "Name": "always"} host_config = client.create_host_config(port_bindings=port_bindings, restart_policy=restart_policy) LOG.info('Creating container') container = client.create_container(image='%s:%s' % (consts.INFLUXDB_IMAGE, consts.INFLUXDB_TAG), ports=ports, detach=True, tty=True, host_config=host_config) LOG.info('Starting container') client.start(container) def _config_influxdb(): try: client = influx.get_data_db_client() client.create_user(consts.INFLUXDB_USER, consts.INFLUXDB_PASS, consts.INFLUXDB_DB_NAME) client.create_database(consts.INFLUXDB_DB_NAME) LOG.info('Success to config influxDB') except Exception: LOG.exception('Config influxdb failed') def _change_output_to_influxdb(): common_utils.makedirs(consts.CONF_DIR) parser = configparser.ConfigParser() LOG.info('Reading output sample configuration') parser.read(consts.CONF_SAMPLE_FILE) LOG.info('Set dispatcher to influxdb') parser.set('DEFAULT', 'dispatcher', 'influxdb') parser.set('dispatcher_influxdb', 'target', 'http://%s:8086' % consts.INFLUXDB_IP) LOG.info('Writing to %s', consts.CONF_FILE) with open(consts.CONF_FILE, 'w') as f: parser.write(f) def prepare_env(args): task_id = str(uuid.uuid4()) thread = threading.Thread(target=_prepare_env_daemon, args=(task_id,)) thread.start() return result_handler(consts.API_SUCCESS, {'task_id': task_id}) def _already_source_openrc(): """Check if openrc is sourced already""" return all(os.environ.get(k) for k in ['OS_AUTH_URL', 'OS_USERNAME', 'OS_PASSWORD', 'EXTERNAL_NETWORK']) def _prepare_env_daemon(task_id): _create_task(task_id) try: _create_directories() rc_file = consts.OPENRC LOG.info('Checkout Openrc Environment variable') if not _already_source_openrc(): LOG.info('Openrc variable not found in Environment') if not os.path.exists(rc_file): LOG.info('Openrc file not found') installer_ip = os.environ.get('INSTALLER_IP', '192.168.200.2') installer_type = os.environ.get('INSTALLER_TYPE', 'compass') LOG.info('Getting openrc file from %s', installer_type) _get_remote_rc_file(rc_file, installer_ip, installer_type) LOG.info('Source openrc file') _source_file(rc_file) LOG.info('Appending external network') _append_external_network(rc_file) LOG.info('Openrc file exist, source openrc file') _source_file(rc_file) LOG.info('Cleaning images') _clean_images() LOG.info('Loading images') _load_images() _update_task_status(task_id) LOG.info('Finished') except Exception as e: _update_task_error(task_id, str(e)) LOG.exception('Prepare env failed') def _create_directories(): common_utils.makedirs(consts.CONF_DIR) def _source_file(rc_file): common_utils.source_env(rc_file) def _get_remote_rc_file(rc_file, installer_ip, installer_type): os_fetch_script = os.path.join(consts.RELENG_DIR, consts.FETCH_SCRIPT) try: cmd = [os_fetch_script, '-d', rc_file, '-i', installer_type, '-a', installer_ip] p = subprocess.Popen(cmd, stdout=subprocess.PIPE) p.communicate() if p.returncode != 0: LOG.error('Failed to fetch credentials from installer') except OSError as e: if e.errno != errno.EEXIST: raise def _append_external_network(rc_file): neutron_client = openstack_utils.get_neutron_client() networks = neutron_client.list_networks()['networks'] try: ext_network = next(n['name'] for n in networks if n['router:external']) except StopIteration: LOG.warning("Can't find external network") else: cmd = 'export EXTERNAL_NETWORK=%s' % ext_network try: with open(rc_file, 'a') as f: f.write(cmd + '\n') except OSError as e: if e.errno != errno.EEXIST: raise def _clean_images(): cmd = [consts.CLEAN_IMAGES_SCRIPT] p = subprocess.Popen(cmd, stdout=subprocess.PIPE, cwd=consts.REPOS_DIR) output = p.communicate()[0] LOG.debug(output) def _load_images(): cmd = [consts.LOAD_IMAGES_SCRIPT] p = subprocess.Popen(cmd, stdout=subprocess.PIPE, cwd=consts.REPOS_DIR) output = p.communicate()[0] LOG.debug(output) def _create_task(task_id): async_handler.insert({'status': 0, 'task_id': task_id}) def _update_task_status(task_id): async_handler.update_attr(task_id, {'status': 1}) def _update_task_error(task_id, error): async_handler.update_attr(task_id, {'status': 2, 'error': error}) def update_openrc(args): try: openrc_vars = args['openrc'] except KeyError: return result_handler(consts.API_ERROR, 'openrc must be provided') else: if not isinstance(openrc_vars, collections.Mapping): return result_handler(consts.API_ERROR, 'args should be a dict') lines = ['export {}={}\n'.format(k, v) for k, v in openrc_vars.items()] LOG.debug('Writing: %s', ''.join(lines)) LOG.info('Writing openrc: Writing') common_utils.makedirs(consts.CONF_DIR) with open(consts.OPENRC, 'w') as f: f.writelines(lines) LOG.info('Writing openrc: Done') LOG.info('Source openrc: Sourcing') try: _source_file(consts.OPENRC) except Exception as e: LOG.exception('Failed to source openrc') return result_handler(consts.API_ERROR, str(e)) LOG.info('Source openrc: Done') return result_handler(consts.API_SUCCESS, {'openrc': openrc_vars}) def upload_pod_file(args): try: pod_file = args['file'] except KeyError: return result_handler(consts.API_ERROR, 'file must be provided') LOG.info('Checking file') data = yaml.load(pod_file.read()) if not isinstance(data, collections.Mapping): return result_handler(consts.API_ERROR, 'invalid yaml file') LOG.info('Writing file') with open(consts.POD_FILE, 'w') as f: yaml.dump(data, f, default_flow_style=False) LOG.info('Writing finished') return result_handler(consts.API_SUCCESS, {'pod_info': data}) def update_pod_file(args): try: pod_dic = args['pod'] except KeyError: return result_handler(consts.API_ERROR, 'pod must be provided') else: if not isinstance(pod_dic, collections.Mapping): return result_handler(consts.API_ERROR, 'pod should be a dict') LOG.info('Writing file') with open(consts.POD_FILE, 'w') as f: yaml.dump(pod_dic, f, default_flow_style=False) LOG.info('Writing finished') return result_handler(consts.API_SUCCESS, {'pod_info': pod_dic}) def update_hosts(hosts_ip): if not isinstance(hosts_ip, dict): return result_handler(consts.API_ERROR, 'Error, args should be a dict') LOG.info('Writing hosts: Writing') hosts_list = ['\n{} {}'.format(ip, host_name) for host_name, ip in hosts_ip.items()] LOG.debug('Writing: %s', hosts_list) with open(consts.ETC_HOSTS, 'a') as f: f.writelines(hosts_list) LOG.info('Writing hosts: Done') return result_handler(consts.API_SUCCESS, 'success')