3 # Copyright (c) 2016 Ericsson AB and others.
5 # All rights reserved. This program and the accompanying materials
6 # are made available under the terms of the Apache License, Version 2.0
7 # which accompanies this distribution, and is available at
8 # http://www.apache.org/licenses/LICENSE-2.0
25 import functest.ci.tier_builder as tb
26 import functest.core.testcase as testcase
27 import functest.utils.functest_utils as ft_utils
28 import functest.utils.openstack_utils as os_utils
29 from functest.utils.constants import CONST
31 # __name__ cannot be used here
32 logger = logging.getLogger('functest.ci.run_tests')
34 CONFIG_FUNCTEST_PATH = pkg_resources.resource_filename(
35 'functest', 'ci/config_functest.yaml')
36 CONFIG_PATCH_PATH = pkg_resources.resource_filename(
37 'functest', 'ci/config_patch.yaml')
38 CONFIG_AARCH64_PATCH_PATH = pkg_resources.resource_filename(
39 'functest', 'ci/config_aarch64_patch.yaml')
40 # set the architecture to default
41 pod_arch = os.getenv("POD_ARCH", None)
42 arch_filter = ['aarch64']
45 class Result(enum.Enum):
50 class BlockingTestFailed(Exception):
54 class TestNotEnabled(Exception):
58 class RunTestsParser(object):
61 self.parser = argparse.ArgumentParser()
62 self.parser.add_argument("-t", "--test", dest="test", action='store',
63 help="Test case or tier (group of tests) "
64 "to be executed. It will run all the test "
66 self.parser.add_argument("-n", "--noclean", help="Do not clean "
67 "OpenStack resources after running each "
68 "test (default=false).",
70 self.parser.add_argument("-r", "--report", help="Push results to "
71 "database (default=false).",
74 def parse_args(self, argv=[]):
75 return vars(self.parser.parse_args(argv))
81 self.executed_test_cases = {}
82 self.overall_result = Result.EX_OK
83 self.clean_flag = True
84 self.report_flag = False
85 self._tiers = tb.TierBuilder(
86 CONST.__getattribute__('INSTALLER_TYPE'),
87 CONST.__getattribute__('DEPLOY_SCENARIO'),
88 pkg_resources.resource_filename('functest', 'ci/testcases.yaml'))
91 def update_config_file():
92 Runner.patch_file(CONFIG_PATCH_PATH)
94 if pod_arch and pod_arch in arch_filter:
95 Runner.patch_file(CONFIG_AARCH64_PATCH_PATH)
98 def patch_file(patch_file_path):
99 logger.debug('Updating file: %s', patch_file_path)
100 with open(patch_file_path) as f:
101 patch_file = yaml.safe_load(f)
104 for key in patch_file:
105 if key in CONST.__getattribute__('DEPLOY_SCENARIO'):
106 new_functest_yaml = dict(ft_utils.merge_dicts(
107 ft_utils.get_functest_yaml(), patch_file[key]))
111 os.remove(CONFIG_FUNCTEST_PATH)
112 with open(CONFIG_FUNCTEST_PATH, "w") as f:
113 f.write(yaml.dump(new_functest_yaml, default_style='"'))
116 def source_rc_file():
117 rc_file = CONST.__getattribute__('openstack_creds')
118 if not os.path.isfile(rc_file):
119 raise Exception("RC file %s does not exist..." % rc_file)
120 logger.debug("Sourcing the OpenStack RC file...")
121 os_utils.source_credentials(rc_file)
122 for key, value in six.iteritems(os.environ):
123 if re.search("OS_", key):
124 if key == 'OS_AUTH_URL':
125 CONST.__setattr__('OS_AUTH_URL', value)
126 elif key == 'OS_USERNAME':
127 CONST.__setattr__('OS_USERNAME', value)
128 elif key == 'OS_TENANT_NAME':
129 CONST.__setattr__('OS_TENANT_NAME', value)
130 elif key == 'OS_PASSWORD':
131 CONST.__setattr__('OS_PASSWORD', value)
132 elif key == "OS_PROJECT_DOMAIN_NAME":
133 CONST.__setattr__('OS_PROJECT_DOMAIN_NAME', value)
136 def get_run_dict(testname):
138 dict = ft_utils.get_dict_by_test(testname)
140 logger.error("Cannot get {}'s config options".format(testname))
145 logger.exception("Cannot get {}'s config options".format(testname))
148 def run_test(self, test):
149 if not test.is_enabled():
150 raise TestNotEnabled(
151 "The test case {} is not enabled".format(test.get_name()))
152 logger.info("Running test case '%s'...", test.get_name())
153 result = testcase.TestCase.EX_RUN_ERROR
154 run_dict = self.get_run_dict(test.get_name())
157 module = importlib.import_module(run_dict['module'])
158 cls = getattr(module, run_dict['class'])
159 test_dict = ft_utils.get_dict_by_test(test.get_name())
160 test_case = cls(**test_dict)
161 self.executed_test_cases[test.get_name()] = test_case
163 if test_case.create_snapshot() != test_case.EX_OK:
164 return testcase.TestCase.EX_RUN_ERROR
166 kwargs = run_dict['args']
167 test_case.run(**kwargs)
171 test_case.push_to_db()
172 if test.get_project() == "functest":
173 result = test_case.is_successful()
175 result = testcase.TestCase.EX_OK
176 logger.info("Test result:\n\n%s\n", test_case)
180 logger.exception("Cannot import module {}".format(
182 except AttributeError:
183 logger.exception("Cannot get class {}".format(
186 raise Exception("Cannot import the class for the test case.")
189 def run_tier(self, tier):
190 tier_name = tier.get_name()
191 tests = tier.get_tests()
192 if tests is None or len(tests) == 0:
193 logger.info("There are no supported test cases in this tier "
194 "for the given scenario")
195 self.overall_result = Result.EX_ERROR
197 logger.info("Running tier '%s'" % tier_name)
200 test_case = self.executed_test_cases[test.get_name()]
201 if test_case.is_successful() != testcase.TestCase.EX_OK:
202 logger.error("The test case '%s' failed.", test.get_name())
203 if test.get_project() == "functest":
204 self.overall_result = Result.EX_ERROR
205 if test.is_blocking():
206 raise BlockingTestFailed(
207 "The test case {} failed and is blocking".format(
209 return self.overall_result
213 msg = prettytable.PrettyTable(
214 header_style='upper', padding_width=5,
215 field_names=['tiers', 'order', 'CI Loop', 'description',
217 for tier in self._tiers.get_tiers():
218 if (len(tier.get_tests()) != 0 and
219 re.search(CONST.__getattribute__('CI_LOOP'),
220 tier.get_ci_loop()) is not None):
221 tiers_to_run.append(tier)
222 msg.add_row([tier.get_name(), tier.get_order(),
224 textwrap.fill(tier.description, width=40),
225 textwrap.fill(' '.join([str(x.get_name(
226 )) for x in tier.get_tests()]), width=40)])
227 logger.info("TESTS TO BE EXECUTED:\n\n%s\n", msg)
228 for tier in tiers_to_run:
231 def main(self, **kwargs):
232 Runner.update_config_file()
234 if 'noclean' in kwargs:
235 self.clean_flag = not kwargs['noclean']
236 if 'report' in kwargs:
237 self.report_flag = kwargs['report']
240 self.source_rc_file()
241 logger.debug("Test args: %s", kwargs['test'])
242 if self._tiers.get_tier(kwargs['test']):
243 self.run_tier(self._tiers.get_tier(kwargs['test']))
244 elif self._tiers.get_test(kwargs['test']):
245 result = self.run_test(
246 self._tiers.get_test(kwargs['test']))
247 if result != testcase.TestCase.EX_OK:
248 logger.error("The test case '%s' failed.",
250 self.overall_result = Result.EX_ERROR
251 elif kwargs['test'] == "all":
254 logger.error("Unknown test case or tier '%s', "
255 "or not supported by "
256 "the given scenario '%s'."
258 CONST.__getattribute__('DEPLOY_SCENARIO')))
259 logger.debug("Available tiers are:\n\n%s",
261 return Result.EX_ERROR
264 except BlockingTestFailed:
267 logger.exception("Failures when running testcase(s)")
268 self.overall_result = Result.EX_ERROR
269 if not self._tiers.get_test(kwargs['test']):
270 self.summary(self._tiers.get_tier(kwargs['test']))
271 logger.info("Execution exit value: %s" % self.overall_result)
272 return self.overall_result
274 def summary(self, tier=None):
275 msg = prettytable.PrettyTable(
276 header_style='upper', padding_width=5,
277 field_names=['env var', 'value'])
278 for env_var in ['INSTALLER_TYPE', 'DEPLOY_SCENARIO', 'BUILD_TAG',
280 msg.add_row([env_var, CONST.__getattribute__(env_var)])
281 logger.info("Deployment description:\n\n%s\n", msg)
282 msg = prettytable.PrettyTable(
283 header_style='upper', padding_width=5,
284 field_names=['test case', 'project', 'tier',
285 'duration', 'result'])
286 tiers = [tier] if tier else self._tiers.get_tiers()
288 for test in tier.get_tests():
290 test_case = self.executed_test_cases[test.get_name()]
292 msg.add_row([test.get_name(), test.get_project(),
293 tier.get_name(), "00:00", "SKIP"])
295 result = 'PASS' if(test_case.is_successful(
296 ) == test_case.EX_OK) else 'FAIL'
298 [test_case.case_name, test_case.project_name,
299 self._tiers.get_tier_name(test_case.case_name),
300 test_case.get_duration(), result])
301 for test in tier.get_skipped_test():
302 msg.add_row([test.get_name(), test.get_project(),
303 tier.get_name(), "00:00", "SKIP"])
304 logger.info("FUNCTEST REPORT:\n\n%s\n", msg)
308 logging.config.fileConfig(pkg_resources.resource_filename(
309 'functest', 'ci/logging.ini'))
310 logging.captureWarnings(True)
311 parser = RunTestsParser()
312 args = parser.parse_args(sys.argv[1:])
314 return runner.main(**args).value