3 # Copyright (c) 2016 Ericsson AB and others.
5 # All rights reserved. This program and the accompanying materials
6 # are made available under the terms of the Apache License, Version 2.0
7 # which accompanies this distribution, and is available at
8 # http://www.apache.org/licenses/LICENSE-2.0
23 import functest.ci.tier_builder as tb
24 import functest.core.testcase as testcase
25 import functest.utils.functest_utils as ft_utils
26 import functest.utils.openstack_utils as os_utils
27 from functest.utils.constants import CONST
29 # __name__ cannot be used here
30 logger = logging.getLogger('functest.ci.run_tests')
33 class Result(enum.Enum):
38 class BlockingTestFailed(Exception):
42 class TestNotEnabled(Exception):
46 class RunTestsParser(object):
49 self.parser = argparse.ArgumentParser()
50 self.parser.add_argument("-t", "--test", dest="test", action='store',
51 help="Test case or tier (group of tests) "
52 "to be executed. It will run all the test "
54 self.parser.add_argument("-n", "--noclean", help="Do not clean "
55 "OpenStack resources after running each "
56 "test (default=false).",
58 self.parser.add_argument("-r", "--report", help="Push results to "
59 "database (default=false).",
62 def parse_args(self, argv=[]):
63 return vars(self.parser.parse_args(argv))
69 self.executed_test_cases = {}
70 self.overall_result = Result.EX_OK
71 self.clean_flag = True
72 self.report_flag = False
73 self._tiers = tb.TierBuilder(
74 CONST.__getattribute__('INSTALLER_TYPE'),
75 CONST.__getattribute__('DEPLOY_SCENARIO'),
76 pkg_resources.resource_filename('functest', 'ci/testcases.yaml'))
80 rc_file = CONST.__getattribute__('openstack_creds')
81 if not os.path.isfile(rc_file):
82 raise Exception("RC file %s does not exist..." % rc_file)
83 logger.debug("Sourcing the OpenStack RC file...")
84 os_utils.source_credentials(rc_file)
85 for key, value in os.environ.iteritems():
86 if re.search("OS_", key):
87 if key == 'OS_AUTH_URL':
88 CONST.__setattr__('OS_AUTH_URL', value)
89 elif key == 'OS_USERNAME':
90 CONST.__setattr__('OS_USERNAME', value)
91 elif key == 'OS_TENANT_NAME':
92 CONST.__setattr__('OS_TENANT_NAME', value)
93 elif key == 'OS_PASSWORD':
94 CONST.__setattr__('OS_PASSWORD', value)
95 elif key == "OS_PROJECT_DOMAIN_NAME":
96 CONST.__setattr__('OS_PROJECT_DOMAIN_NAME', value)
99 def get_run_dict(testname):
101 dict = ft_utils.get_dict_by_test(testname)
103 logger.error("Cannot get {}'s config options".format(testname))
108 logger.exception("Cannot get {}'s config options".format(testname))
111 def run_test(self, test):
112 if not test.is_enabled():
113 raise TestNotEnabled(
114 "The test case {} is not enabled".format(test.get_name()))
115 logger.info("Running test case '%s'...", test.get_name())
116 result = testcase.TestCase.EX_RUN_ERROR
117 run_dict = self.get_run_dict(test.get_name())
120 module = importlib.import_module(run_dict['module'])
121 cls = getattr(module, run_dict['class'])
122 test_dict = ft_utils.get_dict_by_test(test.get_name())
123 test_case = cls(**test_dict)
124 self.executed_test_cases[test.get_name()] = test_case
126 if test_case.create_snapshot() != test_case.EX_OK:
129 kwargs = run_dict['args']
130 result = test_case.run(**kwargs)
132 result = test_case.run()
133 if result == testcase.TestCase.EX_OK:
135 test_case.push_to_db()
136 result = test_case.is_successful()
137 logger.info("Test result:\n\n%s\n", test_case)
141 logger.exception("Cannot import module {}".format(
143 except AttributeError:
144 logger.exception("Cannot get class {}".format(
147 raise Exception("Cannot import the class for the test case.")
150 def run_tier(self, tier):
151 tier_name = tier.get_name()
152 tests = tier.get_tests()
153 if tests is None or len(tests) == 0:
154 logger.info("There are no supported test cases in this tier "
155 "for the given scenario")
156 self.overall_result = Result.EX_ERROR
158 logger.info("Running tier '%s'" % tier_name)
160 result = self.run_test(test)
161 if result != testcase.TestCase.EX_OK:
162 logger.error("The test case '%s' failed.", test.get_name())
163 self.overall_result = Result.EX_ERROR
164 if test.is_blocking():
165 raise BlockingTestFailed(
166 "The test case {} failed and is blocking".format(
168 return self.overall_result
172 msg = prettytable.PrettyTable(
173 header_style='upper', padding_width=5,
174 field_names=['tiers', 'order', 'CI Loop', 'description',
176 for tier in self._tiers.get_tiers():
177 if (len(tier.get_tests()) != 0 and
178 re.search(CONST.__getattribute__('CI_LOOP'),
179 tier.get_ci_loop()) is not None):
180 tiers_to_run.append(tier)
181 msg.add_row([tier.get_name(), tier.get_order(),
183 textwrap.fill(tier.description, width=40),
184 textwrap.fill(' '.join([str(x.get_name(
185 )) for x in tier.get_tests()]), width=40)])
186 logger.info("TESTS TO BE EXECUTED:\n\n%s\n", msg)
187 for tier in tiers_to_run:
190 def main(self, **kwargs):
191 if 'noclean' in kwargs:
192 self.clean_flag = not kwargs['noclean']
193 if 'report' in kwargs:
194 self.report_flag = kwargs['report']
197 self.source_rc_file()
198 logger.debug("Test args: %s", kwargs['test'])
199 if self._tiers.get_tier(kwargs['test']):
200 self.run_tier(self._tiers.get_tier(kwargs['test']))
201 elif self._tiers.get_test(kwargs['test']):
202 result = self.run_test(
203 self._tiers.get_test(kwargs['test']))
204 if result != testcase.TestCase.EX_OK:
205 logger.error("The test case '%s' failed.",
207 self.overall_result = Result.EX_ERROR
208 elif kwargs['test'] == "all":
211 logger.error("Unknown test case or tier '%s', "
212 "or not supported by "
213 "the given scenario '%s'."
215 CONST.__getattribute__('DEPLOY_SCENARIO')))
216 logger.debug("Available tiers are:\n\n%s",
218 return Result.EX_ERROR
221 except BlockingTestFailed:
224 logger.exception("Failures when running testcase(s)")
225 self.overall_result = Result.EX_ERROR
226 if not self._tiers.get_test(kwargs['test']):
227 self.summary(self._tiers.get_tier(kwargs['test']))
228 logger.info("Execution exit value: %s" % self.overall_result)
229 return self.overall_result
231 def summary(self, tier=None):
232 msg = prettytable.PrettyTable(
233 header_style='upper', padding_width=5,
234 field_names=['env var', 'value'])
235 for env_var in ['INSTALLER_TYPE', 'DEPLOY_SCENARIO', 'BUILD_TAG',
237 msg.add_row([env_var, CONST.__getattribute__(env_var)])
238 logger.info("Deployment description:\n\n%s\n", msg)
239 msg = prettytable.PrettyTable(
240 header_style='upper', padding_width=5,
241 field_names=['test case', 'project', 'tier',
242 'duration', 'result'])
243 tiers = [tier] if tier else self._tiers.get_tiers()
245 for test in tier.get_tests():
247 test_case = self.executed_test_cases[test.get_name()]
249 msg.add_row([test.get_name(), test.get_project(),
250 tier.get_name(), "00:00", "SKIP"])
252 result = 'PASS' if(test_case.is_successful(
253 ) == test_case.EX_OK) else 'FAIL'
255 [test_case.case_name, test_case.project_name,
256 self._tiers.get_tier_name(test_case.case_name),
257 test_case.get_duration(), result])
258 for test in tier.get_skipped_test():
259 msg.add_row([test.get_name(), test.get_project(),
260 tier.get_name(), "00:00", "SKIP"])
261 logger.info("FUNCTEST REPORT:\n\n%s\n", msg)
265 logging.config.fileConfig(pkg_resources.resource_filename(
266 'functest', 'ci/logging.ini'))
267 parser = RunTestsParser()
268 args = parser.parse_args(sys.argv[1:])
270 return runner.main(**args).value