Merge "standardize ssh auth"
[yardstick.git] / yardstick / benchmark / runners / base.py
old mode 100644 (file)
new mode 100755 (executable)
index 8483226..7c76e42
-##############################################################################
-# Copyright (c) 2015 Ericsson AB and others.
+# Copyright 2014: Mirantis Inc.
+# All Rights Reserved.
 #
-# All rights reserved. This program and the accompanying materials
-# are made available under the terms of the Apache License, Version 2.0
-# which accompanies this distribution, and is available at
-# http://www.apache.org/licenses/LICENSE-2.0
-##############################################################################
+#    Licensed under the Apache License, Version 2.0 (the "License"); you may
+#    not use this file except in compliance with the License. You may obtain
+#    a copy of the License at
+#
+#         http://www.apache.org/licenses/LICENSE-2.0
+#
+#    Unless required by applicable law or agreed to in writing, software
+#    distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
+#    WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
+#    License for the specific language governing permissions and limitations
+#    under the License.
 
+# yardstick comment: this is a modified copy of
+# rally/rally/benchmark/runners/base.py
+
+from __future__ import absolute_import
 import importlib
 import logging
 import multiprocessing
 import subprocess
 import time
+import traceback
 
-log = logging.getLogger(__name__)
+from oslo_config import cfg
 
 import yardstick.common.utils as utils
 from yardstick.benchmark.scenarios import base as base_scenario
 from yardstick.dispatcher.base import Base as DispatcherBase
 
+log = logging.getLogger(__name__)
+
+CONF = cfg.CONF
 
-def _output_serializer_main(filename, queue):
-    '''entrypoint for the singleton subprocess writing to outfile
+
+def _output_serializer_main(filename, queue, config):
+    """entrypoint for the singleton subprocess writing to outfile
     Use of this process enables multiple instances of a scenario without
     messing up the output file.
-    '''
-    config = {}
-    config["type"] = "File"
-    config["file_path"] = filename
-    dispatcher = DispatcherBase.get(config)
+    """
+    out_type = config['yardstick'].get('DEFAULT', {}).get('dispatcher', 'file')
+    conf = {
+        'type': out_type.capitalize(),
+        'file_path': filename
+    }
+
+    dispatcher = DispatcherBase.get(conf, config)
 
     while True:
         # blocks until data becomes available
         record = queue.get()
         if record == '_TERMINATE_':
+            dispatcher.flush_result_data()
             break
         else:
             dispatcher.record_result_data(record)
 
 
+def _execute_shell_command(command):
+    """execute shell script with error handling"""
+    exitcode = 0
+    output = []
+    try:
+        output = subprocess.check_output(command, shell=True)
+    except Exception:
+        exitcode = -1
+        output = traceback.format_exc()
+        log.error("exec command '%s' error:\n ", command)
+        log.error(traceback.format_exc())
+
+    return exitcode, output
+
+
 def _single_action(seconds, command, queue):
-    '''entrypoint for the single action process'''
+    """entrypoint for the single action process"""
     log.debug("single action, fires after %d seconds (from now)", seconds)
     time.sleep(seconds)
     log.debug("single action: executing command: '%s'", command)
-    data = subprocess.check_output(command, shell=True)
-    log.debug("\n%s" % data)
+    ret_code, data = _execute_shell_command(command)
+    if ret_code < 0:
+        log.error("single action error! command:%s", command)
+        queue.put({'single-action-data': data})
+        return
+    log.debug("single action data: \n%s", data)
+    queue.put({'single-action-data': data})
 
 
 def _periodic_action(interval, command, queue):
-    '''entrypoint for the periodic action process'''
+    """entrypoint for the periodic action process"""
     log.debug("periodic action, fires every: %d seconds", interval)
     time_spent = 0
     while True:
         time.sleep(interval)
         time_spent += interval
         log.debug("periodic action, executing command: '%s'", command)
-        data = subprocess.check_output(command, shell=True)
-        log.debug("\n%s" % data)
+        ret_code, data = _execute_shell_command(command)
+        if ret_code < 0:
+            log.error("periodic action error! command:%s", command)
+            queue.put({'periodic-action-data': data})
+            break
+        log.debug("periodic action data: \n%s", data)
+        queue.put({'periodic-action-data': data})
 
 
 class Runner(object):
@@ -67,7 +111,7 @@ class Runner(object):
 
     @staticmethod
     def get_cls(runner_type):
-        '''return class of specified type'''
+        """return class of specified type"""
         for runner in utils.itersubclasses(Runner):
             if runner_type == runner.__execution_type__:
                 return runner
@@ -75,80 +119,98 @@ class Runner(object):
 
     @staticmethod
     def get_types():
-        '''return a list of known runner type (class) names'''
+        """return a list of known runner type (class) names"""
         types = []
         for runner in utils.itersubclasses(Runner):
             types.append(runner)
         return types
 
     @staticmethod
-    def get(config):
+    def get(runner_cfg, config):
         """Returns instance of a scenario runner for execution type.
         """
         # if there is no runner, start the output serializer subprocess
-        if len(Runner.runners) == 0:
-            log.debug("Starting dump process file '%s'" %
-                      config["output_filename"])
+        if not Runner.runners:
+            log.debug("Starting dump process file '%s'",
+                      runner_cfg["output_filename"])
             Runner.queue = multiprocessing.Queue()
             Runner.dump_process = multiprocessing.Process(
                 target=_output_serializer_main,
                 name="Dumper",
-                args=(config["output_filename"], Runner.queue))
+                args=(runner_cfg["output_filename"], Runner.queue, config))
             Runner.dump_process.start()
 
-        return Runner.get_cls(config["type"])(config, Runner.queue)
+        return Runner.get_cls(runner_cfg["type"])(runner_cfg, Runner.queue)
 
     @staticmethod
-    def release(runner):
-        '''Release the runner'''
-        Runner.runners.remove(runner)
-        # if this was the last runner, stop the output serializer subprocess
-        if len(Runner.runners) == 0:
-            log.debug("Stopping dump process")
+    def release_dump_process():
+        """Release the dumper process"""
+        log.debug("Stopping dump process")
+        if Runner.dump_process:
             Runner.queue.put('_TERMINATE_')
             Runner.dump_process.join()
+            Runner.dump_process = None
+
+    @staticmethod
+    def release(runner):
+        """Release the runner"""
+        if runner in Runner.runners:
+            Runner.runners.remove(runner)
+
+        # if this was the last runner, stop the output serializer subprocess
+        if not Runner.runners:
+            Runner.release_dump_process()
+
+    @staticmethod
+    def terminate(runner):
+        """Terminate the runner"""
+        if runner.process and runner.process.is_alive():
+            runner.process.terminate()
 
     @staticmethod
     def terminate_all():
-        '''Terminate all runners (subprocesses)'''
+        """Terminate all runners (subprocesses)"""
         log.debug("Terminating all runners")
+
+        # release dumper process as some errors before any runner is created
+        if not Runner.runners:
+            Runner.release_dump_process()
+            return
+
         for runner in Runner.runners:
+            log.debug("Terminating runner: %s", runner)
+            if runner.process:
+                runner.process.terminate()
+                runner.process.join()
             if runner.periodic_action_process:
                 log.debug("Terminating periodic action process")
                 runner.periodic_action_process.terminate()
                 runner.periodic_action_process = None
-            runner.process.terminate()
-            runner.process.join()
             Runner.release(runner)
 
     def __init__(self, config, queue):
-        self.context = {}
         self.config = config
         self.periodic_action_process = None
         self.result_queue = queue
+        self.process = None
+        self.aborted = multiprocessing.Event()
         Runner.runners.append(self)
 
-    def run_pre_start_action(self):
-        '''run a potentially configured pre-start action'''
-        if "pre-start-action" in self.config:
-            command = self.config["pre-start-action"]["command"]
-            log.debug("pre start action: command: '%s'" % command)
-            data = subprocess.check_output(command, shell=True)
-            log.debug("pre-start data: \n%s" % data)
-            output = "{'pre-start-action-data': %s}" % data
-            self.result_queue.put(output)
-
     def run_post_stop_action(self):
-        '''run a potentially configured post-stop action'''
+        """run a potentially configured post-stop action"""
         if "post-stop-action" in self.config:
             command = self.config["post-stop-action"]["command"]
-            log.debug("post stop action: command: '%s'" % command)
-            data = subprocess.check_output(command, shell=True)
-            log.debug("post-stop data: \n%s" % data)
-            output = "{'post-stop-action-data': %s}" % data
-            self.result_queue.put(output)
+            log.debug("post stop action: command: '%s'", command)
+            ret_code, data = _execute_shell_command(command)
+            if ret_code < 0:
+                log.error("post action error! command:%s", command)
+                self.result_queue.put({'post-stop-action-data': data})
+                return
+            log.debug("post-stop data: \n%s", data)
+            self.result_queue.put({'post-stop-action-data': data})
 
-    def run(self, scenario_type, scenario_args):
+    def run(self, scenario_cfg, context_cfg):
+        scenario_type = scenario_cfg["type"]
         class_name = base_scenario.Scenario.get(scenario_type)
         path_split = class_name.split(".")
         module_path = ".".join(path_split[:-1])
@@ -156,8 +218,19 @@ class Runner(object):
         cls = getattr(module, path_split[-1])
 
         self.config['object'] = class_name
+        self.aborted.clear()
 
-        self.run_pre_start_action()
+        # run a potentially configured pre-start action
+        if "pre-start-action" in self.config:
+            command = self.config["pre-start-action"]["command"]
+            log.debug("pre start action: command: '%s'", command)
+            ret_code, data = _execute_shell_command(command)
+            if ret_code < 0:
+                log.error("pre-start action error! command:%s", command)
+                self.result_queue.put({'pre-start-action-data': data})
+                return
+            log.debug("pre-start data: \n%s", data)
+            self.result_queue.put({'pre-start-action-data': data})
 
         if "single-shot-action" in self.config:
             single_action_process = multiprocessing.Process(
@@ -177,12 +250,17 @@ class Runner(object):
                       self.result_queue))
             self.periodic_action_process.start()
 
-        self._run_benchmark(cls, "run", scenario_args)
+        self._run_benchmark(cls, "run", scenario_cfg, context_cfg)
 
-    def join(self):
-        self.process.join()
+    def abort(self):
+        """Abort the execution of a scenario"""
+        self.aborted.set()
+
+    def join(self, timeout=None):
+        self.process.join(timeout)
         if self.periodic_action_process:
             self.periodic_action_process.terminate()
             self.periodic_action_process = None
+
         self.run_post_stop_action()
         return self.process.exitcode