67a6f1429d715068e8b115f5408940b975ea4875
[functest.git] / functest / ci / run_tests.py
1 #!/usr/bin/env python
2
3 # Copyright (c) 2016 Ericsson AB and others.
4 #
5 # All rights reserved. This program and the accompanying materials
6 # are made available under the terms of the Apache License, Version 2.0
7 # which accompanies this distribution, and is available at
8 # http://www.apache.org/licenses/LICENSE-2.0
9
10 import argparse
11 import enum
12 import importlib
13 import logging
14 import logging.config
15 import os
16 import pkg_resources
17 import re
18 import sys
19 import textwrap
20
21 import prettytable
22 import six
23 import yaml
24
25 import functest.ci.tier_builder as tb
26 import functest.core.testcase as testcase
27 import functest.utils.functest_utils as ft_utils
28 import functest.utils.openstack_utils as os_utils
29 from functest.utils.constants import CONST
30
31 # __name__ cannot be used here
32 logger = logging.getLogger('functest.ci.run_tests')
33
34 CONFIG_FUNCTEST_PATH = pkg_resources.resource_filename(
35     'functest', 'ci/config_functest.yaml')
36 CONFIG_PATCH_PATH = pkg_resources.resource_filename(
37     'functest', 'ci/config_patch.yaml')
38 CONFIG_AARCH64_PATCH_PATH = pkg_resources.resource_filename(
39     'functest', 'ci/config_aarch64_patch.yaml')
40 # set the architecture to default
41 pod_arch = os.getenv("POD_ARCH", None)
42 arch_filter = ['aarch64']
43
44
45 class Result(enum.Enum):
46     EX_OK = os.EX_OK
47     EX_ERROR = -1
48
49
50 class BlockingTestFailed(Exception):
51     pass
52
53
54 class TestNotEnabled(Exception):
55     pass
56
57
58 class RunTestsParser(object):
59
60     def __init__(self):
61         self.parser = argparse.ArgumentParser()
62         self.parser.add_argument("-t", "--test", dest="test", action='store',
63                                  help="Test case or tier (group of tests) "
64                                  "to be executed. It will run all the test "
65                                  "if not specified.")
66         self.parser.add_argument("-n", "--noclean", help="Do not clean "
67                                  "OpenStack resources after running each "
68                                  "test (default=false).",
69                                  action="store_true")
70         self.parser.add_argument("-r", "--report", help="Push results to "
71                                  "database (default=false).",
72                                  action="store_true")
73
74     def parse_args(self, argv=[]):
75         return vars(self.parser.parse_args(argv))
76
77
78 class Runner(object):
79
80     def __init__(self):
81         self.executed_test_cases = {}
82         self.overall_result = Result.EX_OK
83         self.clean_flag = True
84         self.report_flag = False
85         self._tiers = tb.TierBuilder(
86             CONST.__getattribute__('INSTALLER_TYPE'),
87             CONST.__getattribute__('DEPLOY_SCENARIO'),
88             pkg_resources.resource_filename('functest', 'ci/testcases.yaml'))
89
90     @staticmethod
91     def update_config_file():
92         Runner.patch_file(CONFIG_PATCH_PATH)
93
94         if pod_arch and pod_arch in arch_filter:
95             Runner.patch_file(CONFIG_AARCH64_PATCH_PATH)
96
97     @staticmethod
98     def patch_file(patch_file_path):
99         logger.debug('Updating file: %s', patch_file_path)
100         with open(patch_file_path) as f:
101             patch_file = yaml.safe_load(f)
102
103         updated = False
104         for key in patch_file:
105             if key in CONST.__getattribute__('DEPLOY_SCENARIO'):
106                 new_functest_yaml = dict(ft_utils.merge_dicts(
107                     ft_utils.get_functest_yaml(), patch_file[key]))
108                 updated = True
109
110         if updated:
111             os.remove(CONFIG_FUNCTEST_PATH)
112             with open(CONFIG_FUNCTEST_PATH, "w") as f:
113                 f.write(yaml.dump(new_functest_yaml, default_style='"'))
114
115     @staticmethod
116     def source_rc_file():
117         rc_file = CONST.__getattribute__('openstack_creds')
118         if not os.path.isfile(rc_file):
119             raise Exception("RC file %s does not exist..." % rc_file)
120         logger.debug("Sourcing the OpenStack RC file...")
121         os_utils.source_credentials(rc_file)
122         for key, value in six.iteritems(os.environ):
123             if re.search("OS_", key):
124                 if key == 'OS_AUTH_URL':
125                     CONST.__setattr__('OS_AUTH_URL', value)
126                 elif key == 'OS_USERNAME':
127                     CONST.__setattr__('OS_USERNAME', value)
128                 elif key == 'OS_TENANT_NAME':
129                     CONST.__setattr__('OS_TENANT_NAME', value)
130                 elif key == 'OS_PASSWORD':
131                     CONST.__setattr__('OS_PASSWORD', value)
132                 elif key == "OS_PROJECT_DOMAIN_NAME":
133                     CONST.__setattr__('OS_PROJECT_DOMAIN_NAME', value)
134
135     @staticmethod
136     def get_run_dict(testname):
137         try:
138             dict = ft_utils.get_dict_by_test(testname)
139             if not dict:
140                 logger.error("Cannot get {}'s config options".format(testname))
141             elif 'run' in dict:
142                 return dict['run']
143             return None
144         except Exception:
145             logger.exception("Cannot get {}'s config options".format(testname))
146             return None
147
148     def run_test(self, test):
149         if not test.is_enabled():
150             raise TestNotEnabled(
151                 "The test case {} is not enabled".format(test.get_name()))
152         logger.info("Running test case '%s'...", test.get_name())
153         result = testcase.TestCase.EX_RUN_ERROR
154         run_dict = self.get_run_dict(test.get_name())
155         if run_dict:
156             try:
157                 module = importlib.import_module(run_dict['module'])
158                 cls = getattr(module, run_dict['class'])
159                 test_dict = ft_utils.get_dict_by_test(test.get_name())
160                 test_case = cls(**test_dict)
161                 self.executed_test_cases[test.get_name()] = test_case
162                 if self.clean_flag:
163                     if test_case.create_snapshot() != test_case.EX_OK:
164                         return testcase.TestCase.EX_RUN_ERROR
165                 try:
166                     kwargs = run_dict['args']
167                     test_case.run(**kwargs)
168                 except KeyError:
169                     test_case.run()
170                 if self.report_flag:
171                     test_case.push_to_db()
172                 if test.get_project() == "functest":
173                     result = test_case.is_successful()
174                 else:
175                     result = testcase.TestCase.EX_OK
176                 logger.info("Test result:\n\n%s\n", test_case)
177                 if self.clean_flag:
178                     test_case.clean()
179             except ImportError:
180                 logger.exception("Cannot import module {}".format(
181                     run_dict['module']))
182             except AttributeError:
183                 logger.exception("Cannot get class {}".format(
184                     run_dict['class']))
185         else:
186             raise Exception("Cannot import the class for the test case.")
187         return result
188
189     def run_tier(self, tier):
190         tier_name = tier.get_name()
191         tests = tier.get_tests()
192         if tests is None or len(tests) == 0:
193             logger.info("There are no supported test cases in this tier "
194                         "for the given scenario")
195             self.overall_result = Result.EX_ERROR
196         else:
197             logger.info("Running tier '%s'" % tier_name)
198             for test in tests:
199                 self.run_test(test)
200                 test_case = self.executed_test_cases[test.get_name()]
201                 if test_case.is_successful() != testcase.TestCase.EX_OK:
202                     logger.error("The test case '%s' failed.", test.get_name())
203                     if test.get_project() == "functest":
204                         self.overall_result = Result.EX_ERROR
205                     if test.is_blocking():
206                         raise BlockingTestFailed(
207                             "The test case {} failed and is blocking".format(
208                                 test.get_name()))
209         return self.overall_result
210
211     def run_all(self):
212         tiers_to_run = []
213         msg = prettytable.PrettyTable(
214             header_style='upper', padding_width=5,
215             field_names=['tiers', 'order', 'CI Loop', 'description',
216                          'testcases'])
217         for tier in self._tiers.get_tiers():
218             if (len(tier.get_tests()) != 0 and
219                     re.search(CONST.__getattribute__('CI_LOOP'),
220                               tier.get_ci_loop()) is not None):
221                 tiers_to_run.append(tier)
222                 msg.add_row([tier.get_name(), tier.get_order(),
223                              tier.get_ci_loop(),
224                              textwrap.fill(tier.description, width=40),
225                              textwrap.fill(' '.join([str(x.get_name(
226                                  )) for x in tier.get_tests()]), width=40)])
227         logger.info("TESTS TO BE EXECUTED:\n\n%s\n", msg)
228         for tier in tiers_to_run:
229             self.run_tier(tier)
230
231     def main(self, **kwargs):
232         Runner.update_config_file()
233
234         if 'noclean' in kwargs:
235             self.clean_flag = not kwargs['noclean']
236         if 'report' in kwargs:
237             self.report_flag = kwargs['report']
238         try:
239             if 'test' in kwargs:
240                 self.source_rc_file()
241                 logger.debug("Test args: %s", kwargs['test'])
242                 if self._tiers.get_tier(kwargs['test']):
243                     self.run_tier(self._tiers.get_tier(kwargs['test']))
244                 elif self._tiers.get_test(kwargs['test']):
245                     result = self.run_test(
246                         self._tiers.get_test(kwargs['test']))
247                     if result != testcase.TestCase.EX_OK:
248                         logger.error("The test case '%s' failed.",
249                                      kwargs['test'])
250                         self.overall_result = Result.EX_ERROR
251                 elif kwargs['test'] == "all":
252                     self.run_all()
253                 else:
254                     logger.error("Unknown test case or tier '%s', "
255                                  "or not supported by "
256                                  "the given scenario '%s'."
257                                  % (kwargs['test'],
258                                     CONST.__getattribute__('DEPLOY_SCENARIO')))
259                     logger.debug("Available tiers are:\n\n%s",
260                                  self._tiers)
261                     return Result.EX_ERROR
262             else:
263                 self.run_all()
264         except BlockingTestFailed:
265             pass
266         except Exception:
267             logger.exception("Failures when running testcase(s)")
268             self.overall_result = Result.EX_ERROR
269         if not self._tiers.get_test(kwargs['test']):
270             self.summary(self._tiers.get_tier(kwargs['test']))
271         logger.info("Execution exit value: %s" % self.overall_result)
272         return self.overall_result
273
274     def summary(self, tier=None):
275         msg = prettytable.PrettyTable(
276             header_style='upper', padding_width=5,
277             field_names=['env var', 'value'])
278         for env_var in ['INSTALLER_TYPE', 'DEPLOY_SCENARIO', 'BUILD_TAG',
279                         'CI_LOOP']:
280             msg.add_row([env_var, CONST.__getattribute__(env_var)])
281         logger.info("Deployment description:\n\n%s\n", msg)
282         msg = prettytable.PrettyTable(
283             header_style='upper', padding_width=5,
284             field_names=['test case', 'project', 'tier',
285                          'duration', 'result'])
286         tiers = [tier] if tier else self._tiers.get_tiers()
287         for tier in tiers:
288             for test in tier.get_tests():
289                 try:
290                     test_case = self.executed_test_cases[test.get_name()]
291                 except KeyError:
292                     msg.add_row([test.get_name(), test.get_project(),
293                                  tier.get_name(), "00:00", "SKIP"])
294                 else:
295                     result = 'PASS' if(test_case.is_successful(
296                         ) == test_case.EX_OK) else 'FAIL'
297                     msg.add_row(
298                         [test_case.case_name, test_case.project_name,
299                          self._tiers.get_tier_name(test_case.case_name),
300                          test_case.get_duration(), result])
301             for test in tier.get_skipped_test():
302                 msg.add_row([test.get_name(), test.get_project(),
303                              tier.get_name(), "00:00", "SKIP"])
304         logger.info("FUNCTEST REPORT:\n\n%s\n", msg)
305
306
307 def main():
308     logging.config.fileConfig(pkg_resources.resource_filename(
309         'functest', 'ci/logging.ini'))
310     logging.captureWarnings(True)
311     parser = RunTestsParser()
312     args = parser.parse_args(sys.argv[1:])
313     runner = Runner()
314     return runner.main(**args).value