Rewrite the HA test case (1)
[yardstick.git] / yardstick / benchmark / scenarios / availability / serviceha.py
index 3e03e1d..10134ea 100755 (executable)
@@ -6,13 +6,11 @@
 # which accompanies this distribution, and is available at
 # http://www.apache.org/licenses/LICENSE-2.0
 ##############################################################################
-import pkg_resources
 import logging
 import time
-import yaml
-import yardstick.ssh as ssh
 from yardstick.benchmark.scenarios import base
 from yardstick.benchmark.scenarios.availability import monitor
+from yardstick.benchmark.scenarios.availability.attacker import baseattacker
 
 LOG = logging.getLogger(__name__)
 
@@ -22,87 +20,34 @@ class ServiceHA(base.Scenario):
     """
     __scenario_type__ = "ServiceHA"
 
-    HA_CONF = "ha_tools/ha_conf.yaml"
-
     def __init__(self, scenario_cfg, context_cfg):
+        LOG.debug(
+            "scenario_cfg:%s context_cfg:%s" %
+            (scenario_cfg, context_cfg))
         self.scenario_cfg = scenario_cfg
         self.context_cfg = context_cfg
-        self.service_name = scenario_cfg["options"]["component"]
-        self.fault_type = scenario_cfg["options"]["fault_type"]
-        self.fault_time = scenario_cfg["options"].get("fault_time", 0)
-        self.fault_cfg = None
         self.setup_done = False
-        self.need_teardown = False
 
     def setup(self):
         '''scenario setup'''
-        self.ha_conf_file = pkg_resources.resource_filename(
-            "yardstick.benchmark.scenarios.availability",
-            ServiceHA.HA_CONF)
-        ha_cfg = []
-        with open(self.ha_conf_file) as stream:
-                ha_cfg = yaml.load(stream)
-        LOG.debug("ha_cfg content:%s" % ha_cfg)
-
-        # check the ha_conf contains the service defined in test cases yaml
-        service_cfg = ha_cfg.get(self.service_name, None)
-        if not service_cfg:
-            LOG.error(
-                "The component %s can not be supported!" % self.service_name)
-            return
-
-        for fault in service_cfg:
-            if fault["type"] == self.fault_type:
-                self.fault_cfg = fault
-                break
-        if not self.fault_cfg:
-            LOG.error(
-                "The fualt_type %s can not be supproted!" % self.fault_type)
-            return
-        LOG.debug("the fault_cfg :%s" % self.fault_cfg)
-
-        self.fault_script = pkg_resources.resource_filename(
-            "yardstick.benchmark.scenarios.availability",
-            self.fault_cfg["inject_script"])
-        self.recovery_script = pkg_resources.resource_filename(
-            "yardstick.benchmark.scenarios.availability",
-            self.fault_cfg["recovery_script"])
-        self.check_script = pkg_resources.resource_filename(
-            "yardstick.benchmark.scenarios.availability",
-            self.fault_cfg["check_script"])
-
-        host = self.context_cfg.get("host", None)
-        ip = host.get("ip", None)
-        user = host.get("user", "root")
-        key_filename = host.get("key_filename", "~/.ssh/id_rsa")
-        LOG.info("The host: %s  the service: %s" % (ip, self.service_name))
-        LOG.debug("The params, host:%s  fault_cfg:%s" % (host, self.fault_cfg))
-
-        LOG.debug(
-            "ssh connection ip:%s, user:%s, key_file:%s",
-            ip, user, key_filename)
-        self.connection = ssh.SSH(user, ip, key_filename=key_filename)
-        self.connection.wait(timeout=600)
-        LOG.debug("ssh host success!")
-
-        # check the host envrioment
-        exit_status, stdout, stderr = self.connection.execute(
-            "/bin/sh -s {0}".format(self.service_name),
-            stdin=open(self.check_script, "r"))
-        LOG.info(
-            "the exit_status:%s stdout:%s stderr:%s" %
-            (exit_status, stdout, stderr))
-        if exit_status:
-            raise RuntimeError(stderr)
-
-        if stdout and "running" in stdout:
-            LOG.info("check the envrioment success!")
-        else:
-            LOG.error(
-                "the host envrioment is error, stdout:%s, stderr:%s" %
-                (stdout, stderr))
+        nodes = self.context_cfg.get("nodes", None)
+        if nodes is None:
+            LOG.error("the nodes info is none")
             return
-
+        self.attackers = []
+        attacker_cfgs = self.scenario_cfg["options"]["attackers"]
+        for attacker_cfg in attacker_cfgs:
+            attacker_cls = baseattacker.BaseAttacker.get_attacker_cls(
+                attacker_cfg)
+            attacker_ins = attacker_cls(attacker_cfg, nodes)
+            attacker_ins.setup()
+            self.attackers.append(attacker_ins)
+
+        monitor_cfgs = self.scenario_cfg["options"]["monitors"]
+
+        self.monitor_ins = monitor.Monitor()
+        self.monitor_ins.setup(monitor_cfgs[0])
+        self.monitor_ins.monitor_time = monitor_cfgs[0]["monitor_time"]
         self.setup_done = True
 
     def run(self, result):
@@ -111,27 +56,18 @@ class ServiceHA(base.Scenario):
             LOG.error("The setup not finished!")
             return
 
-        monitorInstance = monitor.Monitor()
-        monitorInstance.setup(self.fault_cfg)
-        monitorInstance.start()
+        self.monitor_ins.start()
         LOG.info("monitor start!")
 
-        LOG.info("Inject fault!")
-        exit_status, stdout, stderr = self.connection.execute(
-            "/bin/sh -s {0}".format(self.service_name),
-            stdin=open(self.fault_script, "r"))
-
-        if exit_status != 0:
-            monitorInstance.stop()
-            raise RuntimeError(stderr)
+        for attacker in self.attackers:
+            attacker.inject_fault()
 
-        self.need_teardown = True
-        time.sleep(self.fault_time)
+        time.sleep(self.monitor_ins.monitor_time)
 
-        monitorInstance.stop()
+        self.monitor_ins.stop()
         LOG.info("monitor stop!")
 
-        ret = monitorInstance.get_result()
+        ret = self.monitor_ins.get_result()
         LOG.info("The monitor result:%s" % ret)
         outage_time = ret.get("outage_time")
         result["outage_time"] = outage_time
@@ -146,34 +82,36 @@ class ServiceHA(base.Scenario):
 
     def teardown(self):
         '''scenario teardown'''
-        LOG.info("recory the everiment!")
-
-        if self.need_teardown:
-            exit_status, stdout, stderr = self.connection.execute(
-                "/bin/sh -s {0} ".format(self.service_name),
-                stdin=open(self.recovery_script, "r"))
+        for attacker in self.attackers:
+            if not attacker.check():
+                attacker.recover()
 
-            if exit_status:
-                raise RuntimeError(stderr)
-            else:
-                self.need_teardown = False
 
-"""
-def _test():
+def _test():    # pragma: no cover
     '''internal test function'''
     host = {
         "ip": "10.20.0.5",
         "user": "root",
         "key_filename": "/root/.ssh/id_rsa"
     }
-    ctx = {"host": host}
-
-    logger = logging.getLogger("yardstick")
-    logger.setLevel(logging.DEBUG)
+    ctx = {"nodes": {"node1": host}}
+    attacker_cfg = {
+        "fault_type": "kill-process",
+        "process_name": "nova-api",
+        "host": "node1"
+    }
+    attacker_cfgs = []
+    attacker_cfgs.append(attacker_cfg)
+    monitor_cfg = {
+        "monitor_cmd": "nova image-list",
+        "monitor_tme": 10
+    }
+    monitor_cfgs = []
+    monitor_cfgs.append(monitor_cfg)
 
     options = {
-        "component": "nova-api",
-        "fault_type": "stop-service"
+        "attackers": attacker_cfgs,
+        "monitors": monitor_cfgs
     }
     sla = {"outage_time": 5}
     args = {"options": options, "sla": sla}
@@ -188,6 +126,5 @@ def _test():
 
     terstInstance.teardown()
 
-if __name__ == '__main__':
+if __name__ == '__main__':    # pragma: no cover
     _test()
-"""