X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=yardstick%2Fnetwork_services%2Fnfvi%2Fresource.py;h=ce09b659782430885c8eb78853c32ac1dcadabd0;hb=ca0aaae662da454451e1efdbef2cd3f779f7c093;hp=d71e1e9951fc279297d6a222d2e5941f97cde636;hpb=61c9612197d9a204cc7efa0a324a2b136886506f;p=yardstick.git diff --git a/yardstick/network_services/nfvi/resource.py b/yardstick/network_services/nfvi/resource.py index d71e1e995..ce09b6597 100644 --- a/yardstick/network_services/nfvi/resource.py +++ b/yardstick/network_services/nfvi/resource.py @@ -14,19 +14,28 @@ """ Resource collection definitions """ from __future__ import absolute_import +from __future__ import print_function +import tempfile import logging +import os import os.path import re import multiprocessing +from collections import Sequence + from oslo_config import cfg from yardstick import ssh from yardstick.network_services.nfvi.collectd import AmqpConsumer from yardstick.network_services.utils import provision_tool +LOG = logging.getLogger(__name__) + CONF = cfg.CONF ZMQ_OVS_PORT = 5567 ZMQ_POLLING_TIME = 12000 +LIST_PLUGINS_ENABLED = ["amqp", "cpu", "cpufreq", "intel_rdt", "memory", + "hugepages", "dpdkstat", "virt", "ovs_stats"] class ResourceProfile(object): @@ -34,19 +43,18 @@ class ResourceProfile(object): This profile adds a resource at the beginning of the test session """ - def __init__(self, vnfd, cores): + def __init__(self, mgmt, interfaces=None, cores=None): self.enable = True self.connection = None - self.cores = cores - - mgmt_interface = vnfd.get("mgmt-interface") - user = mgmt_interface.get("user") - passwd = mgmt_interface.get("password") - ip_addr = mgmt_interface.get("ip") - self.vnfip = mgmt_interface.get("host", ip_addr) - ssh_port = mgmt_interface.get("ssh_port", ssh.DEFAULT_PORT) - self.connection = ssh.SSH(user, self.vnfip, - password=passwd, port=ssh_port) + self.cores = cores if isinstance(cores, Sequence) else [] + self._queue = multiprocessing.Queue() + self.amqp_client = None + self.interfaces = interfaces if isinstance(interfaces, Sequence) else [] + + # why the host or ip? + self.vnfip = mgmt.get("host", mgmt["ip"]) + self.connection = ssh.SSH.from_node(mgmt, overrides={"ip": self.vnfip}) + self.connection.wait() def check_if_sa_running(self, process): @@ -54,81 +62,147 @@ class ResourceProfile(object): err, pid, _ = self.connection.execute("pgrep -f %s" % process) return [err == 0, pid] - def run_collectd_amqp(self, queue): + def run_collectd_amqp(self): """ run amqp consumer to collect the NFVi data """ - amqp = \ - AmqpConsumer('amqp://admin:admin@{}:5672/%2F'.format(self.vnfip), - queue) + amqp_url = 'amqp://admin:admin@{}:5672/%2F'.format(self.vnfip) + amqp = AmqpConsumer(amqp_url, self._queue) try: amqp.run() except (AttributeError, RuntimeError, KeyboardInterrupt): amqp.stop() @classmethod - def get_cpu_data(cls, reskey, value): + def parse_simple_resource(cls, key, value): + return {'/'.join(key): value.split(":")[1]} + + @classmethod + def get_cpu_data(cls, key_split, value): """ Get cpu topology of the host """ pattern = r"-(\d+)" - if "cpufreq" in reskey[1]: - match = re.search(pattern, reskey[2], re.MULTILINE) - metric = reskey[1] + if "cpufreq" in key_split[0]: + metric = key_split[0] + source = key_split[1] else: - match = re.search(pattern, reskey[1], re.MULTILINE) - metric = reskey[2] + metric = key_split[1] + source = key_split[0] + + match = re.search(pattern, source, re.MULTILINE) + if not match: + return "error", "Invalid", "", "" - time, val = re.split(":", value) - if match: - return [str(match.group(1)), metric, val, time] + time, value = value.split(":") + return str(match.group(1)), metric, value, time - return ["error", "Invalid", ""] + @classmethod + def parse_hugepages(cls, key, value): + return cls.parse_simple_resource(key, value) - def parse_collectd_result(self, metrics, listcores): + @classmethod + def parse_dpdkstat(cls, key, value): + return cls.parse_simple_resource(key, value) + + @classmethod + def parse_virt(cls, key, value): + return cls.parse_simple_resource(key, value) + + @classmethod + def parse_ovs_stats(cls, key, value): + return cls.parse_simple_resource(key, value) + + def parse_collectd_result(self, metrics, core_list): """ convert collectd data into json""" - res = {"cpu": {}, "memory": {}} + result = { + "cpu": {}, + "memory": {}, + "hugepages": {}, + "dpdkstat": {}, + "virt": {}, + "ovs_stats": {}, + } testcase = "" for key, value in metrics.items(): - reskey = key.rsplit("/") - if "cpu" in reskey[1] or "intel_rdt" in reskey[1]: - cpu_key, name, metric, testcase = \ - self.get_cpu_data(reskey, value) - if cpu_key in listcores: - res["cpu"].setdefault(cpu_key, {}).update({name: metric}) - elif "memory" in reskey[1]: - val = re.split(":", value)[1] - res["memory"].update({reskey[2]: val}) - res["timestamp"] = testcase - - return res - - def amqp_collect_nfvi_kpi(self, _queue=multiprocessing.Queue()): + key_split = key.split("/") + res_key_iter = (key for key in key_split if "nsb_stats" not in key) + res_key0 = next(res_key_iter) + res_key1 = next(res_key_iter) + + if "cpu" in res_key0 or "intel_rdt" in res_key0: + cpu_key, name, metric, testcase = self.get_cpu_data(key_split, value) + if cpu_key in core_list: + result["cpu"].setdefault(cpu_key, {}).update({name: metric}) + + elif "memory" in res_key0: + result["memory"].update({res_key1: value.split(":")[0]}) + + elif "hugepages" in res_key0: + result["hugepages"].update(self.parse_hugepages(key, value)) + + elif "dpdkstat" in res_key0: + result["dpdkstat"].update(self.parse_dpdkstat(key, value)) + + elif "virt" in res_key1: + result["virt"].update(self.parse_virt(key, value)) + + elif "ovs_stats" in res_key0: + result["ovs_stats"].update(self.parse_ovs_stats(key, value)) + + result["timestamp"] = testcase + + return result + + def amqp_process_for_nfvi_kpi(self): """ amqp collect and return nfvi kpis """ - try: - metric = {} - amqp_client = \ - multiprocessing.Process(target=self.run_collectd_amqp, - args=(_queue,)) - amqp_client.start() - amqp_client.join(7) - amqp_client.terminate() - - while not _queue.empty(): - metric.update(_queue.get()) - except (AttributeError, RuntimeError, TypeError, ValueError): - logging.debug("Failed to get NFVi stats...") - msg = {} - else: - msg = self.parse_collectd_result(metric, self.cores) + if self.amqp_client is None: + self.amqp_client = \ + multiprocessing.Process(target=self.run_collectd_amqp) + self.amqp_client.start() + def amqp_collect_nfvi_kpi(self): + """ amqp collect and return nfvi kpis """ + metric = {} + while not self._queue.empty(): + metric.update(self._queue.get()) + msg = self.parse_collectd_result(metric, self.cores) return msg - @classmethod - def _start_collectd(cls, connection, bin_path): - connection.execute('pkill -9 collectd') + def _provide_config_file(self, bin_path, nfvi_cfg, kwargs): + with open(os.path.join(bin_path, nfvi_cfg), 'r') as cfg: + template = cfg.read() + cfg, cfg_content = tempfile.mkstemp() + with os.fdopen(cfg, "w+") as cfg: + cfg.write(template.format(**kwargs)) + cfg_file = os.path.join(bin_path, nfvi_cfg) + self.connection.put(cfg_content, cfg_file) + + def _prepare_collectd_conf(self, bin_path): + """ Prepare collectd conf """ + loadplugin = "\n".join("LoadPlugin {0}".format(plugin) + for plugin in LIST_PLUGINS_ENABLED) + + interfaces = "\n".join("PortName '{0[name]}'".format(interface) + for interface in self.interfaces) + + kwargs = { + "interval": '25', + "loadplugin": loadplugin, + "dpdk_interface": interfaces, + } + + self._provide_config_file(bin_path, 'collectd.conf', kwargs) + + def _start_collectd(self, connection, bin_path): + LOG.debug("Starting collectd to collect NFVi stats") + # temp disable + return + connection.execute('sudo pkill -9 collectd') collectd = os.path.join(bin_path, "collectd.sh") provision_tool(connection, collectd) - provision_tool(connection, os.path.join(bin_path, "collectd.conf")) + self._prepare_collectd_conf(bin_path) # Reset amqp queue + LOG.debug("reset and setup amqp to collect data from collectd") + connection.execute("sudo rm -rf /var/lib/rabbitmq/mnesia/rabbit*") connection.execute("sudo service rabbitmq-server start") connection.execute("sudo rabbitmqctl stop_app") connection.execute("sudo rabbitmqctl reset") @@ -136,8 +210,15 @@ class ResourceProfile(object): connection.execute("sudo service rabbitmq-server restart") # Run collectd - connection.execute(collectd) - connection.execute(os.path.join(bin_path, "collectd", "collectd")) + + http_proxy = os.environ.get('http_proxy', '') + https_proxy = os.environ.get('https_proxy', '') + connection.execute("sudo %s '%s' '%s'" % + (collectd, http_proxy, https_proxy)) + LOG.debug("Start collectd service.....") + connection.execute( + "sudo %s" % os.path.join(bin_path, "collectd", "collectd")) + LOG.debug("Done") def initiate_systemagent(self, bin_path): """ Start system agent for NFVi collection on host """ @@ -147,16 +228,24 @@ class ResourceProfile(object): def start(self): """ start nfvi collection """ if self.enable: - logging.debug("Start NVFi metric collection...") + LOG.debug("Start NVFi metric collection...") def stop(self): """ stop nfvi collection """ - if self.enable: - agent = "collectd" - logging.debug("Stop resource monitor...") - status, pid = self.check_if_sa_running(agent) - if status: - self.connection.execute('kill -9 %s' % pid) - self.connection.execute('pkill -9 %s' % agent) - self.connection.execute('service rabbitmq-server stop') - self.connection.execute("sudo rabbitmqctl stop_app") + if not self.enable: + return + + agent = "collectd" + LOG.debug("Stop resource monitor...") + + if self.amqp_client is not None: + self.amqp_client.terminate() + + status, pid = self.check_if_sa_running(agent) + if status == 0: + return + + self.connection.execute('sudo kill -9 %s' % pid) + self.connection.execute('sudo pkill -9 %s' % agent) + self.connection.execute('sudo service rabbitmq-server stop') + self.connection.execute("sudo rabbitmqctl stop_app")