X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=ci%2Frun_tests.py;h=982567217fa97d03bb45ddd7c629452f4f52f872;hb=53fd052b7fce64294c50ca170d42367d711ccc3f;hp=cb05d435d0b964860c0175dee773559e78afb558;hpb=9ca917e17b6d04bce32cc420954aada538a60ddb;p=functest.git diff --git a/ci/run_tests.py b/ci/run_tests.py index cb05d435d..982567217 100755 --- a/ci/run_tests.py +++ b/ci/run_tests.py @@ -13,7 +13,7 @@ import datetime import os import re import sys - +import functest.ci.generate_report as generate_report import functest.ci.tier_builder as tb import functest.utils.functest_logger as ft_logger import functest.utils.functest_utils as ft_utils @@ -44,6 +44,7 @@ FUNCTEST_REPO = ("%s/functest/" % REPOS_DIR) EXEC_SCRIPT = ("%sci/exec_test.sh" % FUNCTEST_REPO) CLEAN_FLAG = True REPORT_FLAG = False +EXECUTED_TEST_CASES = [] # This will be the return code of this script. If any of the tests fails, # this variable will change to -1 @@ -74,8 +75,16 @@ def cleanup(): os_clean.main() -def run_test(test): - global OVERALL_RESULT +def update_test_info(test_name, result, duration): + for test in EXECUTED_TEST_CASES: + if test['test_name'] == test_name: + test.update({"result": result, + "duration": duration}) + + +def run_test(test, tier_name): + global OVERALL_RESULT, EXECUTED_TEST_CASES + result_str = "PASS" start = datetime.datetime.now() test_name = test.get_name() logger.info("\n") # blank line @@ -93,29 +102,35 @@ def run_test(test): cmd = ("%s%s" % (EXEC_SCRIPT, flags)) logger.debug("Executing command '%s'" % cmd) - - result = ft_utils.execute_command(cmd, logger, exit_on_error=False) + result = ft_utils.execute_command(cmd, exit_on_error=False) if CLEAN_FLAG: cleanup() - end = datetime.datetime.now() duration = (end - start).seconds - str = ("%02d:%02d" % divmod(duration, 60)) - logger.info("Test execution time: %s" % str) + duration_str = ("%02d:%02d" % divmod(duration, 60)) + logger.info("Test execution time: %s" % duration_str) if result != 0: logger.error("The test case '%s' failed. " % test_name) OVERALL_RESULT = -1 - - if test.get_blocking(): - logger.info("This test case is blocking. Exiting...") + result_str = "FAIL" + + if test.is_blocking(): + if not args.test or args.test == "all": + logger.info("This test case is blocking. Aborting overall " + "execution.") + # if it is a single test we don't print the whole results table + update_test_info(test_name, result_str, duration_str) + generate_report.main(EXECUTED_TEST_CASES) + logger.info("Execution exit value: %s" % OVERALL_RESULT) sys.exit(OVERALL_RESULT) - return result + update_test_info(test_name, result_str, duration_str) def run_tier(tier): + tier_name = tier.get_name() tests = tier.get_tests() if tests is None or len(tests) == 0: logger.info("There are no supported test cases in this tier " @@ -123,14 +138,15 @@ def run_tier(tier): return 0 logger.info("\n\n") # blank line print_separator("#") - logger.info("Running tier '%s'" % tier.get_name()) + logger.info("Running tier '%s'" % tier_name) print_separator("#") logger.debug("\n%s" % tier) for test in tests: - run_test(test) + run_test(test, tier_name) def run_all(tiers): + global EXECUTED_TEST_CASES summary = "" BUILD_TAG = os.getenv('BUILD_TAG') if BUILD_TAG is not None and re.search("daily", BUILD_TAG) is not None: @@ -149,10 +165,12 @@ def run_all(tiers): tier.get_test_names())) logger.info("Tests to be executed:%s" % summary) - + EXECUTED_TEST_CASES = generate_report.init(tiers_to_run) for tier in tiers_to_run: run_tier(tier) + generate_report.main(EXECUTED_TEST_CASES) + def main(): global CLEAN_FLAG @@ -176,7 +194,7 @@ def main(): run_tier(_tiers.get_tier(args.test)) elif _tiers.get_test(args.test): - run_test(_tiers.get_test(args.test)) + run_test(_tiers.get_test(args.test), _tiers.get_tier(args.test)) elif args.test == "all": run_all(_tiers) @@ -190,6 +208,7 @@ def main(): else: run_all(_tiers) + logger.info("Execution exit value: %s" % OVERALL_RESULT) sys.exit(OVERALL_RESULT) if __name__ == '__main__':