Merge "API proposal for functest"
authorCedric Ollivier <cedric.ollivier@orange.com>
Thu, 17 Aug 2017 11:12:52 +0000 (11:12 +0000)
committerGerrit Code Review <gerrit@opnfv.org>
Thu, 17 Aug 2017 11:12:52 +0000 (11:12 +0000)
docker/Dockerfile
docker/smoke/Dockerfile
functest/ci/run_tests.py
functest/ci/tier_builder.py
functest/ci/tier_handler.py
functest/tests/unit/ci/test_run_tests.py
functest/tests/unit/ci/test_tier_builder.py

index 1811e93..910be83 100644 (file)
@@ -14,7 +14,7 @@ LABEL version="0.1" description="OPNFV Functest Docker container"
 # Environment variables
 ARG BRANCH=master
 ARG RALLY_TAG=0.8.1
-ARG ODL_TAG=release/beryllium-sr4
+ARG ODL_TAG=release/carbon
 ARG OPENSTACK_TAG=stable/ocata
 ARG KINGBIRD_TAG=1.1.0
 ARG VIMS_TAG=stable
index 223ed64..1038547 100644 (file)
@@ -2,7 +2,7 @@ FROM opnfv/functest-core
 
 ARG BRANCH=master
 ARG OPENSTACK_TAG=stable/ocata
-ARG ODL_TAG=release/beryllium-sr4
+ARG ODL_TAG=release/carbon
 
 COPY thirdparty-requirements.txt thirdparty-requirements.txt
 RUN apk --no-cache add --virtual .build-deps --update \
index b95e100..e26f430 100644 (file)
@@ -1,12 +1,11 @@
 #!/usr/bin/env python
-#
-# Author: Jose Lausuch (jose.lausuch@ericsson.com)
+
+# Copyright (c) 2016 Ericsson AB and others.
 #
 # All rights reserved. This program and the accompanying materials
 # are made available under the terms of the Apache License, Version 2.0
 # which accompanies this distribution, and is available at
 # http://www.apache.org/licenses/LICENSE-2.0
-#
 
 import argparse
 import enum
@@ -17,6 +16,7 @@ import os
 import pkg_resources
 import re
 import sys
+import textwrap
 
 import prettytable
 
@@ -66,17 +66,14 @@ class RunTestsParser(object):
 class Runner(object):
 
     def __init__(self):
-        self.executed_test_cases = []
+        self.executed_test_cases = {}
         self.overall_result = Result.EX_OK
         self.clean_flag = True
         self.report_flag = False
-
-    @staticmethod
-    def print_separator(str, count=45):
-        line = ""
-        for i in range(0, count - 1):
-            line += str
-        logger.info("%s" % line)
+        self._tiers = tb.TierBuilder(
+            CONST.__getattribute__('INSTALLER_TYPE'),
+            CONST.__getattribute__('DEPLOY_SCENARIO'),
+            pkg_resources.resource_filename('functest', 'ci/testcases.yaml'))
 
     @staticmethod
     def source_rc_file():
@@ -109,21 +106,11 @@ class Runner(object):
             logger.exception("Cannot get {}'s config options".format(testname))
             return None
 
-    def run_test(self, test, tier_name, testcases=None):
+    def run_test(self, test):
         if not test.is_enabled():
             raise TestNotEnabled(
                 "The test case {} is not enabled".format(test.get_name()))
-        logger.info("\n")  # blank line
-        self.print_separator("=")
         logger.info("Running test case '%s'...", test.get_name())
-        self.print_separator("=")
-        logger.debug("\n%s" % test)
-        self.source_rc_file()
-
-        flags = " -t %s" % test.get_name()
-        if self.report_flag:
-            flags += " -r"
-
         result = testcase.TestCase.EX_RUN_ERROR
         run_dict = self.get_run_dict(test.get_name())
         if run_dict:
@@ -132,7 +119,7 @@ class Runner(object):
                 cls = getattr(module, run_dict['class'])
                 test_dict = ft_utils.get_dict_by_test(test.get_name())
                 test_case = cls(**test_dict)
-                self.executed_test_cases.append(test_case)
+                self.executed_test_cases[test.get_name()] = test_case
                 if self.clean_flag:
                     if test_case.create_snapshot() != test_case.EX_OK:
                         return result
@@ -156,7 +143,6 @@ class Runner(object):
                     run_dict['class']))
         else:
             raise Exception("Cannot import the class for the test case.")
-
         return result
 
     def run_tier(self, tier):
@@ -165,68 +151,60 @@ class Runner(object):
         if tests is None or len(tests) == 0:
             logger.info("There are no supported test cases in this tier "
                         "for the given scenario")
-            return 0
-        logger.info("\n\n")  # blank line
-        self.print_separator("#")
-        logger.info("Running tier '%s'" % tier_name)
-        self.print_separator("#")
-        logger.debug("\n%s" % tier)
-        for test in tests:
-            result = self.run_test(test, tier_name)
-            if result != testcase.TestCase.EX_OK:
-                logger.error("The test case '%s' failed.", test.get_name())
-                self.overall_result = Result.EX_ERROR
-                if test.is_blocking():
-                    raise BlockingTestFailed(
-                        "The test case {} failed and is blocking".format(
-                            test.get_name()))
+            self.overall_result = Result.EX_ERROR
+        else:
+            logger.info("Running tier '%s'" % tier_name)
+            for test in tests:
+                result = self.run_test(test)
+                if result != testcase.TestCase.EX_OK:
+                    logger.error("The test case '%s' failed.", test.get_name())
+                    self.overall_result = Result.EX_ERROR
+                    if test.is_blocking():
+                        raise BlockingTestFailed(
+                            "The test case {} failed and is blocking".format(
+                                test.get_name()))
+        return self.overall_result
 
-    def run_all(self, tiers):
-        summary = ""
+    def run_all(self):
         tiers_to_run = []
-
-        for tier in tiers.get_tiers():
+        msg = prettytable.PrettyTable(
+            header_style='upper', padding_width=5,
+            field_names=['tiers', 'order', 'CI Loop', 'description',
+                         'testcases'])
+        for tier in self._tiers.get_tiers():
             if (len(tier.get_tests()) != 0 and
                     re.search(CONST.__getattribute__('CI_LOOP'),
                               tier.get_ci_loop()) is not None):
                 tiers_to_run.append(tier)
-                summary += ("\n    - %s:\n\t   %s"
-                            % (tier.get_name(),
-                               tier.get_test_names()))
-
-        logger.info("Tests to be executed:%s" % summary)
+                msg.add_row([tier.get_name(), tier.get_order(),
+                             tier.get_ci_loop(),
+                             textwrap.fill(tier.description, width=40),
+                             textwrap.fill(' '.join([str(x.get_name(
+                                 )) for x in tier.get_tests()]), width=40)])
+        logger.info("TESTS TO BE EXECUTED:\n\n%s\n", msg)
         for tier in tiers_to_run:
             self.run_tier(tier)
 
     def main(self, **kwargs):
-        _tiers = tb.TierBuilder(
-            CONST.__getattribute__('INSTALLER_TYPE'),
-            CONST.__getattribute__('DEPLOY_SCENARIO'),
-            pkg_resources.resource_filename('functest', 'ci/testcases.yaml'))
-
         if kwargs['noclean']:
             self.clean_flag = False
-
         if kwargs['report']:
             self.report_flag = True
-
         try:
             if kwargs['test']:
                 self.source_rc_file()
                 logger.debug("Test args: %s", kwargs['test'])
-                if _tiers.get_tier(kwargs['test']):
-                    self.run_tier(_tiers.get_tier(kwargs['test']))
-                elif _tiers.get_test(kwargs['test']):
+                if self._tiers.get_tier(kwargs['test']):
+                    self.run_tier(self._tiers.get_tier(kwargs['test']))
+                elif self._tiers.get_test(kwargs['test']):
                     result = self.run_test(
-                        _tiers.get_test(kwargs['test']),
-                        _tiers.get_tier_name(kwargs['test']),
-                        kwargs['test'])
+                        self._tiers.get_test(kwargs['test']))
                     if result != testcase.TestCase.EX_OK:
                         logger.error("The test case '%s' failed.",
                                      kwargs['test'])
                         self.overall_result = Result.EX_ERROR
                 elif kwargs['test'] == "all":
-                    self.run_all(_tiers)
+                    self.run_all()
                 else:
                     logger.error("Unknown test case or tier '%s', "
                                  "or not supported by "
@@ -234,39 +212,45 @@ class Runner(object):
                                  % (kwargs['test'],
                                     CONST.__getattribute__('DEPLOY_SCENARIO')))
                     logger.debug("Available tiers are:\n\n%s",
-                                 _tiers)
+                                 self._tiers)
                     return Result.EX_ERROR
             else:
-                self.run_all(_tiers)
+                self.run_all()
         except BlockingTestFailed:
             pass
         except Exception:
             logger.exception("Failures when running testcase(s)")
             self.overall_result = Result.EX_ERROR
+        if not self._tiers.get_test(kwargs['test']):
+            self.summary(self._tiers.get_tier(kwargs['test']))
+        logger.info("Execution exit value: %s" % self.overall_result)
+        return self.overall_result
 
+    def summary(self, tier=None):
         msg = prettytable.PrettyTable(
             header_style='upper', padding_width=5,
             field_names=['env var', 'value'])
         for env_var in ['INSTALLER_TYPE', 'DEPLOY_SCENARIO', 'BUILD_TAG',
                         'CI_LOOP']:
             msg.add_row([env_var, CONST.__getattribute__(env_var)])
-        logger.info("Deployment description: \n\n%s\n", msg)
-
-        if len(self.executed_test_cases) > 1:
-            msg = prettytable.PrettyTable(
-                header_style='upper', padding_width=5,
-                field_names=['test case', 'project', 'tier',
-                             'duration', 'result'])
-            for test_case in self.executed_test_cases:
+        logger.info("Deployment description:\n\n%s\n", msg)
+        msg = prettytable.PrettyTable(
+            header_style='upper', padding_width=5,
+            field_names=['test case', 'project', 'tier',
+                         'duration', 'result'])
+        tiers = [tier] if tier else self._tiers.get_tiers()
+        for tier in tiers:
+            for test in tier.get_tests():
+                test_case = self.executed_test_cases[test.get_name()]
                 result = 'PASS' if(test_case.is_successful(
-                        ) == test_case.EX_OK) else 'FAIL'
+                    ) == test_case.EX_OK) else 'FAIL'
                 msg.add_row([test_case.case_name, test_case.project_name,
-                             _tiers.get_tier_name(test_case.case_name),
+                             self._tiers.get_tier_name(test_case.case_name),
                              test_case.get_duration(), result])
-            logger.info("FUNCTEST REPORT: \n\n%s\n", msg)
-
-        logger.info("Execution exit value: %s" % self.overall_result)
-        return self.overall_result
+            for test in tier.get_skipped_test():
+                msg.add_row([test.get_name(), test.get_project(),
+                             tier.get_name(), "00:00", "SKIP"])
+        logger.info("FUNCTEST REPORT:\n\n%s\n", msg)
 
 
 def main():
index f803846..d2722dc 100644 (file)
@@ -1,11 +1,11 @@
 #!/usr/bin/env python
+
+# Copyright (c) 2016 Ericsson AB and others.
 #
-# jose.lausuch@ericsson.com
 # All rights reserved. This program and the accompanying materials
 # are made available under the terms of the Apache License, Version 2.0
 # which accompanies this distribution, and is available at
 # http://www.apache.org/licenses/LICENSE-2.0
-#
 
 import tier_handler as th
 import yaml
@@ -52,11 +52,14 @@ class TierBuilder(object):
                                        dependency=dep,
                                        criteria=dic_testcase['criteria'],
                                        blocking=dic_testcase['blocking'],
-                                       description=dic_testcase['description'])
+                                       description=dic_testcase['description'],
+                                       project=dic_testcase['project_name'])
                 if (testcase.is_compatible(self.ci_installer,
                                            self.ci_scenario) and
                         testcase.is_enabled()):
                     tier.add_test(testcase)
+                else:
+                    tier.skip_test(testcase)
 
             self.tier_objects.append(tier)
 
index 4f2f14e..dd3e77c 100644 (file)
@@ -1,14 +1,18 @@
 #!/usr/bin/env python
+
+# Copyright (c) 2016 Ericsson AB and others.
 #
-# jose.lausuch@ericsson.com
 # All rights reserved. This program and the accompanying materials
 # are made available under the terms of the Apache License, Version 2.0
 # which accompanies this distribution, and is available at
 # http://www.apache.org/licenses/LICENSE-2.0
-#
 
 
 import re
+import textwrap
+
+import prettytable
+
 
 LINE_LENGTH = 72
 
@@ -32,6 +36,7 @@ class Tier(object):
 
     def __init__(self, name, order, ci_loop, description=""):
         self.tests_array = []
+        self.skipped_tests_array = []
         self.name = name
         self.order = order
         self.ci_loop = ci_loop
@@ -40,12 +45,18 @@ class Tier(object):
     def add_test(self, testcase):
         self.tests_array.append(testcase)
 
+    def skip_test(self, testcase):
+        self.skipped_tests_array.append(testcase)
+
     def get_tests(self):
         array_tests = []
         for test in self.tests_array:
             array_tests.append(test)
         return array_tests
 
+    def get_skipped_test(self):
+        return self.skipped_tests_array
+
     def get_test_names(self):
         array_tests = []
         for test in self.tests_array:
@@ -75,31 +86,16 @@ class Tier(object):
         return self.ci_loop
 
     def __str__(self):
-        lines = split_text(self.description, LINE_LENGTH - 6)
-
-        out = ""
-        out += ("+%s+\n" % ("=" * (LINE_LENGTH - 2)))
-        out += ("| Tier:  " + self.name.ljust(LINE_LENGTH - 10) + "|\n")
-        out += ("+%s+\n" % ("=" * (LINE_LENGTH - 2)))
-        out += ("| Order: " + str(self.order).ljust(LINE_LENGTH - 10) + "|\n")
-        out += ("| CI Loop: " + str(self.ci_loop).ljust(LINE_LENGTH - 12) +
-                "|\n")
-        out += ("| Description:".ljust(LINE_LENGTH - 1) + "|\n")
-        for line in lines:
-            out += ("|    " + line.ljust(LINE_LENGTH - 7) + " |\n")
-        out += ("| Test cases:".ljust(LINE_LENGTH - 1) + "|\n")
-        tests = self.get_test_names()
-        if len(tests) > 0:
-            for i in range(len(tests)):
-                out += ("|    - %s |\n" % tests[i].ljust(LINE_LENGTH - 9))
-        else:
-            out += ("|    (There are no supported test cases "
-                    .ljust(LINE_LENGTH - 1) + "|\n")
-            out += ("|    in this tier for the given scenario) "
-                    .ljust(LINE_LENGTH - 1) + "|\n")
-        out += ("|".ljust(LINE_LENGTH - 1) + "|\n")
-        out += ("+%s+\n" % ("-" * (LINE_LENGTH - 2)))
-        return out
+        msg = prettytable.PrettyTable(
+            header_style='upper', padding_width=5,
+            field_names=['tiers', 'order', 'CI Loop', 'description',
+                         'testcases'])
+        msg.add_row(
+            [self.name, self.order, self.ci_loop,
+             textwrap.fill(self.description, width=40),
+             textwrap.fill(' '.join([str(x.get_name(
+                 )) for x in self.get_tests()]), width=40)])
+        return msg.get_string()
 
 
 class TestCase(object):
@@ -109,13 +105,15 @@ class TestCase(object):
                  dependency,
                  criteria,
                  blocking,
-                 description=""):
+                 description="",
+                 project=""):
         self.name = name
         self.enabled = enabled
         self.dependency = dependency
         self.criteria = criteria
         self.blocking = blocking
         self.description = description
+        self.project = project
 
     @staticmethod
     def is_none(item):
@@ -147,26 +145,16 @@ class TestCase(object):
     def is_blocking(self):
         return self.blocking
 
+    def get_project(self):
+        return self.project
+
     def __str__(self):
-        lines = split_text(self.description, LINE_LENGTH - 6)
-
-        out = ""
-        out += ("+%s+\n" % ("=" * (LINE_LENGTH - 2)))
-        out += ("| Testcase:  " + self.name.ljust(LINE_LENGTH - 14) + "|\n")
-        out += ("+%s+\n" % ("=" * (LINE_LENGTH - 2)))
-        out += ("| Description:".ljust(LINE_LENGTH - 1) + "|\n")
-        for line in lines:
-            out += ("|    " + line.ljust(LINE_LENGTH - 7) + " |\n")
-        out += ("| Criteria:  " +
-                str(self.criteria).ljust(LINE_LENGTH - 14) + "|\n")
-        out += ("| Dependencies:".ljust(LINE_LENGTH - 1) + "|\n")
-        installer = self.dependency.get_installer()
-        scenario = self.dependency.get_scenario()
-        out += ("|   - Installer:" + installer.ljust(LINE_LENGTH - 17) + "|\n")
-        out += ("|   - Scenario :" + scenario.ljust(LINE_LENGTH - 17) + "|\n")
-        out += ("|".ljust(LINE_LENGTH - 1) + "|\n")
-        out += ("+%s+\n" % ("-" * (LINE_LENGTH - 2)))
-        return out
+        msg = prettytable.PrettyTable(
+            header_style='upper', padding_width=5,
+            field_names=['test case', 'description', 'criteria', 'dependency'])
+        msg.add_row([self.name, textwrap.fill(self.description, width=40),
+                     self.criteria, self.dependency])
+        return msg.get_string()
 
 
 class Dependency(object):
@@ -182,6 +170,7 @@ class Dependency(object):
         return self.scenario
 
     def __str__(self):
-        return ("Dependency info:\n"
-                "        installer: " + self.installer + "\n"
-                "        scenario:  " + self.scenario + "\n")
+        delimitator = "\n" if self.get_installer(
+            ) and self.get_scenario() else ""
+        return "{}{}{}".format(self.get_installer(), delimitator,
+                               self.get_scenario())
index fb8cb39..7495c40 100644 (file)
@@ -54,11 +54,6 @@ class RunTestsTesting(unittest.TestCase):
 
         self.run_tests_parser = run_tests.RunTestsParser()
 
-    @mock.patch('functest.ci.run_tests.logger.info')
-    def test_print_separator(self, mock_logger_info):
-        self.runner.print_separator(self.sep)
-        mock_logger_info.assert_called_once_with(self.sep * 44)
-
     @mock.patch('functest.ci.run_tests.logger.error')
     def test_source_rc_file_missing_file(self, mock_logger_error):
         with mock.patch('functest.ci.run_tests.os.path.isfile',
@@ -120,8 +115,7 @@ class RunTestsTesting(unittest.TestCase):
         args = {'get_name.return_value': 'test_name',
                 'needs_clean.return_value': False}
         mock_test.configure_mock(**args)
-        with mock.patch('functest.ci.run_tests.Runner.print_separator'),\
-            mock.patch('functest.ci.run_tests.Runner.source_rc_file'), \
+        with mock.patch('functest.ci.run_tests.Runner.source_rc_file'), \
             mock.patch('functest.ci.run_tests.Runner.get_run_dict',
                        return_value=None), \
                 self.assertRaises(Exception) as context:
@@ -129,7 +123,6 @@ class RunTestsTesting(unittest.TestCase):
             msg = "Cannot import the class for the test case."
             self.assertTrue(msg in context)
 
-    @mock.patch('functest.ci.run_tests.Runner.print_separator')
     @mock.patch('functest.ci.run_tests.Runner.source_rc_file')
     @mock.patch('importlib.import_module', name="module",
                 return_value=mock.Mock(test_class=mock.Mock(
@@ -145,123 +138,107 @@ class RunTestsTesting(unittest.TestCase):
         with mock.patch('functest.ci.run_tests.Runner.get_run_dict',
                         return_value=test_run_dict):
             self.runner.clean_flag = True
-            self.runner.run_test(mock_test, 'tier_name')
+            self.runner.run_test(mock_test)
         self.assertEqual(self.runner.overall_result,
                          run_tests.Result.EX_OK)
 
-    @mock.patch('functest.ci.run_tests.logger.info')
-    def test_run_tier_default(self, mock_logger_info):
-        with mock.patch('functest.ci.run_tests.Runner.print_separator'), \
-                mock.patch(
-                    'functest.ci.run_tests.Runner.run_test',
-                    return_value=TestCase.EX_OK) as mock_method:
-            self.runner.run_tier(self.tier)
-            mock_method.assert_any_call(mock.ANY, 'test_tier')
-            self.assertTrue(mock_logger_info.called)
+    @mock.patch('functest.ci.run_tests.Runner.run_test',
+                return_value=TestCase.EX_OK)
+    def test_run_tier_default(self, *mock_methods):
+        self.assertEqual(self.runner.run_tier(self.tier),
+                         run_tests.Result.EX_OK)
+        mock_methods[0].assert_called_with(mock.ANY)
 
     @mock.patch('functest.ci.run_tests.logger.info')
     def test_run_tier_missing_test(self, mock_logger_info):
-        with mock.patch('functest.ci.run_tests.Runner.print_separator'):
-            self.tier.get_tests.return_value = None
-            self.assertEqual(self.runner.run_tier(self.tier), 0)
-            self.assertTrue(mock_logger_info.called)
+        self.tier.get_tests.return_value = None
+        self.assertEqual(self.runner.run_tier(self.tier),
+                         run_tests.Result.EX_ERROR)
+        self.assertTrue(mock_logger_info.called)
 
     @mock.patch('functest.ci.run_tests.logger.info')
-    def test_run_all_default(self, mock_logger_info):
-        with mock.patch(
-                'functest.ci.run_tests.Runner.run_tier') as mock_method:
-            CONST.__setattr__('CI_LOOP', 'test_ci_loop')
-            self.runner.run_all(self.tiers)
-            mock_method.assert_any_call(self.tier)
-            self.assertTrue(mock_logger_info.called)
+    @mock.patch('functest.ci.run_tests.Runner.run_tier')
+    @mock.patch('functest.ci.run_tests.Runner.summary')
+    def test_run_all_default(self, *mock_methods):
+        CONST.__setattr__('CI_LOOP', 'test_ci_loop')
+        self.runner.run_all()
+        mock_methods[1].assert_not_called()
+        self.assertTrue(mock_methods[2].called)
 
     @mock.patch('functest.ci.run_tests.logger.info')
-    def test_run_all_missing_tier(self, mock_logger_info):
+    @mock.patch('functest.ci.run_tests.Runner.summary')
+    def test_run_all_missing_tier(self, *mock_methods):
         CONST.__setattr__('CI_LOOP', 'loop_re_not_available')
-        self.runner.run_all(self.tiers)
-        self.assertTrue(mock_logger_info.called)
+        self.runner.run_all()
+        self.assertTrue(mock_methods[1].called)
 
-    def test_main_failed(self):
+    @mock.patch('functest.ci.run_tests.Runner.source_rc_file',
+                side_effect=Exception)
+    @mock.patch('functest.ci.run_tests.Runner.summary')
+    def test_main_failed(self, *mock_methods):
         kwargs = {'test': 'test_name', 'noclean': True, 'report': True}
-        mock_obj = mock.Mock()
         args = {'get_tier.return_value': False,
                 'get_test.return_value': False}
-        mock_obj.configure_mock(**args)
-        with mock.patch('functest.ci.run_tests.tb.TierBuilder'), \
-            mock.patch('functest.ci.run_tests.Runner.source_rc_file',
-                       side_effect=Exception):
-            self.assertEqual(self.runner.main(**kwargs),
-                             run_tests.Result.EX_ERROR)
-        with mock.patch('functest.ci.run_tests.tb.TierBuilder',
-                        return_value=mock_obj), \
-            mock.patch('functest.ci.run_tests.Runner.source_rc_file',
-                       side_effect=Exception):
-            self.assertEqual(self.runner.main(**kwargs),
-                             run_tests.Result.EX_ERROR)
-
-    def test_main_tier(self, *args):
+        self.runner._tiers = mock.Mock()
+        self.runner._tiers.configure_mock(**args)
+        self.assertEqual(self.runner.main(**kwargs),
+                         run_tests.Result.EX_ERROR)
+        mock_methods[1].assert_called_once_with()
+
+    @mock.patch('functest.ci.run_tests.Runner.source_rc_file')
+    @mock.patch('functest.ci.run_tests.Runner.run_test',
+                return_value=TestCase.EX_OK)
+    @mock.patch('functest.ci.run_tests.Runner.summary')
+    def test_main_tier(self, *mock_methods):
         mock_tier = mock.Mock()
-        args = {'get_name.return_value': 'tier_name'}
+        args = {'get_name.return_value': 'tier_name',
+                'get_tests.return_value': ['test_name']}
         mock_tier.configure_mock(**args)
         kwargs = {'test': 'tier_name', 'noclean': True, 'report': True}
-        mock_obj = mock.Mock()
         args = {'get_tier.return_value': mock_tier,
                 'get_test.return_value': None}
-        mock_obj.configure_mock(**args)
-        with mock.patch('functest.ci.run_tests.tb.TierBuilder',
-                        return_value=mock_obj), \
-            mock.patch('functest.ci.run_tests.Runner.source_rc_file'), \
-                mock.patch('functest.ci.run_tests.Runner.run_tier') as m:
-            self.assertEqual(self.runner.main(**kwargs),
-                             run_tests.Result.EX_OK)
-            self.assertTrue(m.called)
-
-    def test_main_test(self, *args):
+        self.runner._tiers = mock.Mock()
+        self.runner._tiers.configure_mock(**args)
+        self.assertEqual(self.runner.main(**kwargs),
+                         run_tests.Result.EX_OK)
+        mock_methods[1].assert_called_once_with('test_name')
+
+    @mock.patch('functest.ci.run_tests.Runner.source_rc_file')
+    @mock.patch('functest.ci.run_tests.Runner.run_test',
+                return_value=TestCase.EX_OK)
+    def test_main_test(self, *mock_methods):
         kwargs = {'test': 'test_name', 'noclean': True, 'report': True}
-        mock_test = mock.Mock()
-        args = {'get_name.return_value': 'test_name',
-                'needs_clean.return_value': True}
-        mock_test.configure_mock(**args)
-        mock_obj = mock.Mock()
         args = {'get_tier.return_value': None,
-                'get_test.return_value': mock_test}
-        mock_obj.configure_mock(**args)
-        with mock.patch('functest.ci.run_tests.tb.TierBuilder',
-                        return_value=mock_obj), \
-            mock.patch('functest.ci.run_tests.Runner.source_rc_file'), \
-                mock.patch('functest.ci.run_tests.Runner.run_test',
-                           return_value=TestCase.EX_OK) as m:
-            self.assertEqual(self.runner.main(**kwargs),
-                             run_tests.Result.EX_OK)
-            self.assertTrue(m.called)
-
-    def test_main_all_tier(self, *args):
+                'get_test.return_value': 'test_name'}
+        self.runner._tiers = mock.Mock()
+        self.runner._tiers.configure_mock(**args)
+        self.assertEqual(self.runner.main(**kwargs),
+                         run_tests.Result.EX_OK)
+        mock_methods[0].assert_called_once_with('test_name')
+
+    @mock.patch('functest.ci.run_tests.Runner.source_rc_file')
+    @mock.patch('functest.ci.run_tests.Runner.run_all')
+    @mock.patch('functest.ci.run_tests.Runner.summary')
+    def test_main_all_tier(self, *mock_methods):
         kwargs = {'test': 'all', 'noclean': True, 'report': True}
-        mock_obj = mock.Mock()
         args = {'get_tier.return_value': None,
                 'get_test.return_value': None}
-        mock_obj.configure_mock(**args)
-        with mock.patch('functest.ci.run_tests.tb.TierBuilder',
-                        return_value=mock_obj), \
-            mock.patch('functest.ci.run_tests.Runner.source_rc_file'), \
-                mock.patch('functest.ci.run_tests.Runner.run_all') as m:
-            self.assertEqual(self.runner.main(**kwargs),
-                             run_tests.Result.EX_OK)
-            self.assertTrue(m.called)
-
-    def test_main_any_tier_test_ko(self, *args):
+        self.runner._tiers = mock.Mock()
+        self.runner._tiers.configure_mock(**args)
+        self.assertEqual(self.runner.main(**kwargs),
+                         run_tests.Result.EX_OK)
+        mock_methods[1].assert_called_once_with()
+
+    @mock.patch('functest.ci.run_tests.Runner.source_rc_file')
+    @mock.patch('functest.ci.run_tests.Runner.summary')
+    def test_main_any_tier_test_ko(self, *mock_methods):
         kwargs = {'test': 'any', 'noclean': True, 'report': True}
-        mock_obj = mock.Mock()
         args = {'get_tier.return_value': None,
                 'get_test.return_value': None}
-        mock_obj.configure_mock(**args)
-        with mock.patch('functest.ci.run_tests.tb.TierBuilder',
-                        return_value=mock_obj), \
-            mock.patch('functest.ci.run_tests.Runner.source_rc_file'), \
-                mock.patch('functest.ci.run_tests.logger.debug') as m:
-            self.assertEqual(self.runner.main(**kwargs),
-                             run_tests.Result.EX_ERROR)
-            self.assertTrue(m.called)
+        self.runner._tiers = mock.Mock()
+        self.runner._tiers.configure_mock(**args)
+        self.assertEqual(self.runner.main(**kwargs),
+                         run_tests.Result.EX_ERROR)
 
 
 if __name__ == "__main__":
index ab75e15..700c6e9 100644 (file)
@@ -24,7 +24,8 @@ class TierBuilderTesting(unittest.TestCase):
                          'case_name': 'test_name',
                          'criteria': 'test_criteria',
                          'blocking': 'test_blocking',
-                         'description': 'test_desc'}
+                         'description': 'test_desc',
+                         'project_name': 'project_name'}
 
         self.dic_tier = {'name': 'test_tier',
                          'order': 'test_order',