From 934a16a7766a5f06b046d882eee408a725b03139 Mon Sep 17 00:00:00 2001 From: "jose.lausuch" Date: Fri, 22 Jul 2016 17:49:34 +0200 Subject: [PATCH] Generate Functest report at the end of the execution JIRA: FUNCTEST-89 Change-Id: I10c41efdd6a7f737f57ab15158c8c8d49f392271 Signed-off-by: jose.lausuch --- ci/generate_report.py | 146 ++++++++++++++++++++++++++++++++++++++++++++++++++ ci/run_tests.py | 40 ++++++++++---- 2 files changed, 175 insertions(+), 11 deletions(-) create mode 100644 ci/generate_report.py diff --git a/ci/generate_report.py b/ci/generate_report.py new file mode 100644 index 00000000..16962fe0 --- /dev/null +++ b/ci/generate_report.py @@ -0,0 +1,146 @@ +import json +import os +import re +import urllib2 +import functest.utils.functest_logger as ft_logger + + +COL_1_LEN = 25 +COL_2_LEN = 15 +COL_3_LEN = 12 +COL_4_LEN = 10 +COL_5_LEN = 75 + +# If we run from CI (Jenkins) we will push the results to the DB +# and then we can print the url to the specific test result +IS_CI_RUN = False +BUILD_TAG = None + +logger = ft_logger.Logger("generate_report").getLogger() + + +def init(tiers_to_run): + test_cases_arr = [] + for tier in tiers_to_run: + for test in tier.get_tests(): + test_cases_arr.append({'test_name': test.get_name(), + 'tier_name': tier.get_name(), + 'result': 'Not executed', + 'duration': '0', + 'url': ''}) + return test_cases_arr + + +def get_results_from_db(): + url = 'http://testresults.opnfv.org/test/api/v1/results?build_tag=' + \ + BUILD_TAG + logger.debug("Query to rest api: %s" % url) + data = json.load(urllib2.urlopen(url)) + return data['results'] + + +def get_data(test, results): + test_result = test['result'] + url = '' + for test_db in results: + if test['test_name'] in test_db['case_name']: + id = test_db['_id'] + url = 'http://testresults.opnfv.org/test/api/v1/results/' + id + test_result = test_db['criteria'] + + return {"url": url, "result": test_result} + + +def print_line(w1, w2='', w3='', w4='', w5=''): + str = ('| ' + w1.ljust(COL_1_LEN - 1) + + '| ' + w2.ljust(COL_2_LEN - 1) + + '| ' + w3.ljust(COL_3_LEN - 1) + + '| ' + w4.ljust(COL_4_LEN - 1)) + if IS_CI_RUN: + str += ('| ' + w5.ljust(COL_5_LEN - 1)) + str += '|\n' + return str + + +def print_line_no_columns(str): + TOTAL_LEN = COL_1_LEN + COL_2_LEN + COL_3_LEN + COL_4_LEN + 2 + if IS_CI_RUN: + TOTAL_LEN += COL_5_LEN + 1 + return ('| ' + str.ljust(TOTAL_LEN) + "|\n") + + +def print_separator(char="=", delimiter="+"): + str = ("+" + char * COL_1_LEN + + delimiter + char * COL_2_LEN + + delimiter + char * COL_3_LEN + + delimiter + char * COL_4_LEN) + if IS_CI_RUN: + str += (delimiter + char * COL_5_LEN) + str += '+\n' + return str + + +def main(args): + global BUILD_TAG, IS_CI_RUN + executed_test_cases = args + + BUILD_TAG = os.getenv("BUILD_TAG") + if BUILD_TAG is not None: + IS_CI_RUN = True + + if IS_CI_RUN: + results = get_results_from_db() + for test in executed_test_cases: + data = get_data(test, results) + test.update({"url": data['url'], + "result": data['result']}) + + TOTAL_LEN = COL_1_LEN + COL_2_LEN + COL_3_LEN + COL_4_LEN + if IS_CI_RUN: + TOTAL_LEN += COL_5_LEN + MID = TOTAL_LEN / 2 + + INSTALLER = os.getenv('INSTALLER_TYPE', 'unknown') + CI_LOOP = os.getenv('CI_LOOP') + SCENARIO = os.getenv('DEPLOY_SCENARIO') + CI_LOOP = None + if BUILD_TAG is not None: + if re.search("daily", BUILD_TAG) is not None: + CI_LOOP = "daily" + else: + CI_LOOP = "weekly" + + str = '' + str += print_separator('=', delimiter="=") + str += print_line_no_columns(' ' * (MID - 8) + 'FUNCTEST REPORT') + str += print_separator('=', delimiter="=") + str += print_line_no_columns(' ') + str += print_line_no_columns(" Deployment description:") + str += print_line_no_columns(" INSTALLER: %s" % INSTALLER) + if SCENARIO is not None: + str += print_line_no_columns(" SCENARIO: %s" % SCENARIO) + if BUILD_TAG is not None: + str += print_line_no_columns(" BUILD TAG: %s" % BUILD_TAG) + if CI_LOOP is not None: + str += print_line_no_columns(" CI LOOP: %s" % CI_LOOP) + str += print_line_no_columns(' ') + str += print_separator('=') + if IS_CI_RUN: + str += print_line('TEST CASE', 'TIER', 'DURATION', 'RESULT', 'URL') + else: + str += print_line('TEST CASE', 'TIER', 'DURATION', 'RESULT') + str += print_separator('=') + for test in executed_test_cases: + str += print_line(test['test_name'], + test['tier_name'], + test['duration'], + test['result'], + test['url']) + str += print_separator('-') + + logger.info("\n\n\n%s" % str) + + +if __name__ == '__main__': + import sys + main(sys.argv[1:]) diff --git a/ci/run_tests.py b/ci/run_tests.py index cb05d435..ca7ff2b5 100755 --- a/ci/run_tests.py +++ b/ci/run_tests.py @@ -13,7 +13,7 @@ import datetime import os import re import sys - +import functest.ci.generate_report as generate_report import functest.ci.tier_builder as tb import functest.utils.functest_logger as ft_logger import functest.utils.functest_utils as ft_utils @@ -44,6 +44,7 @@ FUNCTEST_REPO = ("%s/functest/" % REPOS_DIR) EXEC_SCRIPT = ("%sci/exec_test.sh" % FUNCTEST_REPO) CLEAN_FLAG = True REPORT_FLAG = False +EXECUTED_TEST_CASES = [] # This will be the return code of this script. If any of the tests fails, # this variable will change to -1 @@ -74,8 +75,9 @@ def cleanup(): os_clean.main() -def run_test(test): - global OVERALL_RESULT +def run_test(test, tier_name): + global OVERALL_RESULT, EXECUTED_TEST_CASES + result_str = "PASS" start = datetime.datetime.now() test_name = test.get_name() logger.info("\n") # blank line @@ -93,29 +95,35 @@ def run_test(test): cmd = ("%s%s" % (EXEC_SCRIPT, flags)) logger.debug("Executing command '%s'" % cmd) - result = ft_utils.execute_command(cmd, logger, exit_on_error=False) if CLEAN_FLAG: cleanup() - end = datetime.datetime.now() duration = (end - start).seconds - str = ("%02d:%02d" % divmod(duration, 60)) - logger.info("Test execution time: %s" % str) + duration_str = ("%02d:%02d" % divmod(duration, 60)) + logger.info("Test execution time: %s" % duration_str) + result = 0 if result != 0: logger.error("The test case '%s' failed. " % test_name) OVERALL_RESULT = -1 + result_str = "FAIL" if test.get_blocking(): logger.info("This test case is blocking. Exiting...") sys.exit(OVERALL_RESULT) + for test in EXECUTED_TEST_CASES: + if test['test_name'] == test_name: + test.update({"result": result_str, + "duration": duration_str}) + return result def run_tier(tier): + tier_name = tier.get_name() tests = tier.get_tests() if tests is None or len(tests) == 0: logger.info("There are no supported test cases in this tier " @@ -123,14 +131,19 @@ def run_tier(tier): return 0 logger.info("\n\n") # blank line print_separator("#") - logger.info("Running tier '%s'" % tier.get_name()) + logger.info("Running tier '%s'" % tier_name) print_separator("#") logger.debug("\n%s" % tier) for test in tests: - run_test(test) + res = run_test(test, tier_name) + if res != 0: + return res + + return 0 def run_all(tiers): + global EXECUTED_TEST_CASES summary = "" BUILD_TAG = os.getenv('BUILD_TAG') if BUILD_TAG is not None and re.search("daily", BUILD_TAG) is not None: @@ -149,9 +162,14 @@ def run_all(tiers): tier.get_test_names())) logger.info("Tests to be executed:%s" % summary) - + EXECUTED_TEST_CASES = generate_report.init(tiers_to_run) for tier in tiers_to_run: - run_tier(tier) + res = run_tier(tier) + if res != 0: + return res + generate_report.main(EXECUTED_TEST_CASES) + + return 0 def main(): -- 2.16.6