X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=yardstick%2Fbenchmark%2Fscenarios%2Favailability%2Fserviceha.py;h=10134ea6d8e6c45a3bfa0e60d21386f555feffb5;hb=53a96f499680587dc2f6172a964af35f9ce1c74b;hp=3e03e1da5fb7fd5a2124865e42fde49e5e6bae0f;hpb=9dd5d48efae06f6b87504c8f3fbfad0ab770a25e;p=yardstick.git diff --git a/yardstick/benchmark/scenarios/availability/serviceha.py b/yardstick/benchmark/scenarios/availability/serviceha.py index 3e03e1da5..10134ea6d 100755 --- a/yardstick/benchmark/scenarios/availability/serviceha.py +++ b/yardstick/benchmark/scenarios/availability/serviceha.py @@ -6,13 +6,11 @@ # which accompanies this distribution, and is available at # http://www.apache.org/licenses/LICENSE-2.0 ############################################################################## -import pkg_resources import logging import time -import yaml -import yardstick.ssh as ssh from yardstick.benchmark.scenarios import base from yardstick.benchmark.scenarios.availability import monitor +from yardstick.benchmark.scenarios.availability.attacker import baseattacker LOG = logging.getLogger(__name__) @@ -22,87 +20,34 @@ class ServiceHA(base.Scenario): """ __scenario_type__ = "ServiceHA" - HA_CONF = "ha_tools/ha_conf.yaml" - def __init__(self, scenario_cfg, context_cfg): + LOG.debug( + "scenario_cfg:%s context_cfg:%s" % + (scenario_cfg, context_cfg)) self.scenario_cfg = scenario_cfg self.context_cfg = context_cfg - self.service_name = scenario_cfg["options"]["component"] - self.fault_type = scenario_cfg["options"]["fault_type"] - self.fault_time = scenario_cfg["options"].get("fault_time", 0) - self.fault_cfg = None self.setup_done = False - self.need_teardown = False def setup(self): '''scenario setup''' - self.ha_conf_file = pkg_resources.resource_filename( - "yardstick.benchmark.scenarios.availability", - ServiceHA.HA_CONF) - ha_cfg = [] - with open(self.ha_conf_file) as stream: - ha_cfg = yaml.load(stream) - LOG.debug("ha_cfg content:%s" % ha_cfg) - - # check the ha_conf contains the service defined in test cases yaml - service_cfg = ha_cfg.get(self.service_name, None) - if not service_cfg: - LOG.error( - "The component %s can not be supported!" % self.service_name) - return - - for fault in service_cfg: - if fault["type"] == self.fault_type: - self.fault_cfg = fault - break - if not self.fault_cfg: - LOG.error( - "The fualt_type %s can not be supproted!" % self.fault_type) - return - LOG.debug("the fault_cfg :%s" % self.fault_cfg) - - self.fault_script = pkg_resources.resource_filename( - "yardstick.benchmark.scenarios.availability", - self.fault_cfg["inject_script"]) - self.recovery_script = pkg_resources.resource_filename( - "yardstick.benchmark.scenarios.availability", - self.fault_cfg["recovery_script"]) - self.check_script = pkg_resources.resource_filename( - "yardstick.benchmark.scenarios.availability", - self.fault_cfg["check_script"]) - - host = self.context_cfg.get("host", None) - ip = host.get("ip", None) - user = host.get("user", "root") - key_filename = host.get("key_filename", "~/.ssh/id_rsa") - LOG.info("The host: %s the service: %s" % (ip, self.service_name)) - LOG.debug("The params, host:%s fault_cfg:%s" % (host, self.fault_cfg)) - - LOG.debug( - "ssh connection ip:%s, user:%s, key_file:%s", - ip, user, key_filename) - self.connection = ssh.SSH(user, ip, key_filename=key_filename) - self.connection.wait(timeout=600) - LOG.debug("ssh host success!") - - # check the host envrioment - exit_status, stdout, stderr = self.connection.execute( - "/bin/sh -s {0}".format(self.service_name), - stdin=open(self.check_script, "r")) - LOG.info( - "the exit_status:%s stdout:%s stderr:%s" % - (exit_status, stdout, stderr)) - if exit_status: - raise RuntimeError(stderr) - - if stdout and "running" in stdout: - LOG.info("check the envrioment success!") - else: - LOG.error( - "the host envrioment is error, stdout:%s, stderr:%s" % - (stdout, stderr)) + nodes = self.context_cfg.get("nodes", None) + if nodes is None: + LOG.error("the nodes info is none") return - + self.attackers = [] + attacker_cfgs = self.scenario_cfg["options"]["attackers"] + for attacker_cfg in attacker_cfgs: + attacker_cls = baseattacker.BaseAttacker.get_attacker_cls( + attacker_cfg) + attacker_ins = attacker_cls(attacker_cfg, nodes) + attacker_ins.setup() + self.attackers.append(attacker_ins) + + monitor_cfgs = self.scenario_cfg["options"]["monitors"] + + self.monitor_ins = monitor.Monitor() + self.monitor_ins.setup(monitor_cfgs[0]) + self.monitor_ins.monitor_time = monitor_cfgs[0]["monitor_time"] self.setup_done = True def run(self, result): @@ -111,27 +56,18 @@ class ServiceHA(base.Scenario): LOG.error("The setup not finished!") return - monitorInstance = monitor.Monitor() - monitorInstance.setup(self.fault_cfg) - monitorInstance.start() + self.monitor_ins.start() LOG.info("monitor start!") - LOG.info("Inject fault!") - exit_status, stdout, stderr = self.connection.execute( - "/bin/sh -s {0}".format(self.service_name), - stdin=open(self.fault_script, "r")) - - if exit_status != 0: - monitorInstance.stop() - raise RuntimeError(stderr) + for attacker in self.attackers: + attacker.inject_fault() - self.need_teardown = True - time.sleep(self.fault_time) + time.sleep(self.monitor_ins.monitor_time) - monitorInstance.stop() + self.monitor_ins.stop() LOG.info("monitor stop!") - ret = monitorInstance.get_result() + ret = self.monitor_ins.get_result() LOG.info("The monitor result:%s" % ret) outage_time = ret.get("outage_time") result["outage_time"] = outage_time @@ -146,34 +82,36 @@ class ServiceHA(base.Scenario): def teardown(self): '''scenario teardown''' - LOG.info("recory the everiment!") - - if self.need_teardown: - exit_status, stdout, stderr = self.connection.execute( - "/bin/sh -s {0} ".format(self.service_name), - stdin=open(self.recovery_script, "r")) + for attacker in self.attackers: + if not attacker.check(): + attacker.recover() - if exit_status: - raise RuntimeError(stderr) - else: - self.need_teardown = False -""" -def _test(): +def _test(): # pragma: no cover '''internal test function''' host = { "ip": "10.20.0.5", "user": "root", "key_filename": "/root/.ssh/id_rsa" } - ctx = {"host": host} - - logger = logging.getLogger("yardstick") - logger.setLevel(logging.DEBUG) + ctx = {"nodes": {"node1": host}} + attacker_cfg = { + "fault_type": "kill-process", + "process_name": "nova-api", + "host": "node1" + } + attacker_cfgs = [] + attacker_cfgs.append(attacker_cfg) + monitor_cfg = { + "monitor_cmd": "nova image-list", + "monitor_tme": 10 + } + monitor_cfgs = [] + monitor_cfgs.append(monitor_cfg) options = { - "component": "nova-api", - "fault_type": "stop-service" + "attackers": attacker_cfgs, + "monitors": monitor_cfgs } sla = {"outage_time": 5} args = {"options": options, "sla": sla} @@ -188,6 +126,5 @@ def _test(): terstInstance.teardown() -if __name__ == '__main__': +if __name__ == '__main__': # pragma: no cover _test() -"""