Merge "Cleanup ScenarioGeneral unit tests"
[yardstick.git] / yardstick / benchmark / scenarios / networking / vnf_generic.py
index 18a75d0..eb62d62 100644 (file)
 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 # See the License for the specific language governing permissions and
 # limitations under the License.
-""" NSPerf specific scenario definition """
-
-from __future__ import absolute_import
-import logging
-
-import errno
 
+import copy
 import ipaddress
-import os
-
-import re
 from itertools import chain
+import logging
+import os
+import sys
+import time
 
 import six
-from operator import itemgetter
-from collections import defaultdict
-
-from yardstick.benchmark.scenarios import base
-from yardstick.common.utils import import_modules_from_package, itersubclasses
-from yardstick.common.yaml_loader import yaml_load
+import yaml
+
+from yardstick.benchmark.contexts import base as context_base
+from yardstick.benchmark.scenarios import base as scenario_base
+from yardstick.common.constants import LOG_DIR
+from yardstick.common import exceptions
+from yardstick.common.process import terminate_children
+from yardstick.common import utils
 from yardstick.network_services.collector.subscriber import Collector
 from yardstick.network_services.vnf_generic import vnfdgen
 from yardstick.network_services.vnf_generic.vnf.base import GenericVNF
-from yardstick.network_services.traffic_profile.base import TrafficProfile
+from yardstick.network_services import traffic_profile
+from yardstick.network_services.traffic_profile import base as tprofile_base
+from yardstick.network_services.utils import get_nsb_option
 from yardstick import ssh
 
-LOG = logging.getLogger(__name__)
-
-
-class SSHError(Exception):
-    """Class handles ssh connection error exception"""
-    pass
 
-
-class SSHTimeout(SSHError):
-    """Class handles ssh connection timeout exception"""
-    pass
-
-
-class IncorrectConfig(Exception):
-    """Class handles incorrect configuration during setup"""
-    pass
-
-
-class IncorrectSetup(Exception):
-    """Class handles incorrect setup during setup"""
-    pass
-
-
-class SshManager(object):
-    def __init__(self, node):
-        super(SshManager, self).__init__()
-        self.node = node
-        self.conn = None
-
-    def __enter__(self):
-        """
-        args -> network device mappings
-        returns -> ssh connection ready to be used
-        """
-        try:
-            self.conn = ssh.SSH.from_node(self.node)
-            self.conn.wait()
-        except SSHError as error:
-            LOG.info("connect failed to %s, due to %s", self.node["ip"], error)
-        # self.conn defaults to None
-        return self.conn
-
-    def __exit__(self, exc_type, exc_val, exc_tb):
-        if self.conn:
-            self.conn.close()
-
-
-def find_relative_file(path, task_path):
-    # fixme: create schema to validate all fields have been provided
-    try:
-        with open(path):
-            pass
-        return path
-    except IOError as e:
-        if e.errno != errno.ENOENT:
-            raise
-        else:
-            rel_path = os.path.join(task_path, path)
-            with open(rel_path):
-                pass
-            return rel_path
+traffic_profile.register_modules()
 
 
-def open_relative_file(path, task_path):
-    try:
-        return open(path)
-    except IOError as e:
-        if e.errno == errno.ENOENT:
-            return open(os.path.join(task_path, path))
-        raise
+LOG = logging.getLogger(__name__)
 
 
-class NetworkServiceTestCase(base.Scenario):
+class NetworkServiceTestCase(scenario_base.Scenario):
     """Class handles Generic framework to do pre-deployment VNF &
        Network service testing  """
 
@@ -120,20 +55,24 @@ class NetworkServiceTestCase(base.Scenario):
         self.scenario_cfg = scenario_cfg
         self.context_cfg = context_cfg
 
-        # fixme: create schema to validate all fields have been provided
-        with open_relative_file(scenario_cfg["topology"],
-                                scenario_cfg['task_path']) as stream:
-            topology_yaml = yaml_load(stream)
-
-        self.topology = topology_yaml["nsd:nsd-catalog"]["nsd"][0]
+        self._render_topology()
         self.vnfs = []
         self.collector = None
         self.traffic_profile = None
+        self.node_netdevs = {}
+        self.bin_path = get_nsb_option('bin_path', '')
 
     def _get_ip_flow_range(self, ip_start_range):
 
+        # IP range is specified as 'x.x.x.x-y.y.y.y'
+        if isinstance(ip_start_range, six.string_types):
+            return ip_start_range
+
         node_name, range_or_interface = next(iter(ip_start_range.items()), (None, '0.0.0.0'))
-        if node_name is not None:
+        if node_name is None:
+            # we are manually specifying the range
+            ip_addr_range = range_or_interface
+        else:
             node = self.context_cfg["nodes"].get(node_name, {})
             try:
                 # the ip_range is the interface name
@@ -148,24 +87,37 @@ class NetworkServiceTestCase(base.Scenario):
 
             ipaddr = ipaddress.ip_network(six.text_type('{}/{}'.format(ip, mask)), strict=False)
             hosts = list(ipaddr.hosts())
-            ip_addr_range = "{}-{}".format(hosts[0], hosts[-1])
-        else:
-            # we are manually specifying the range
-            ip_addr_range = range_or_interface
+            if len(hosts) > 2:
+                # skip the first host in case of gateway
+                ip_addr_range = "{}-{}".format(hosts[1], hosts[-1])
+            else:
+                LOG.warning("Only single IP in range %s", ipaddr)
+                # fall back to single IP range
+                ip_addr_range = ip
         return ip_addr_range
 
     def _get_traffic_flow(self):
         flow = {}
         try:
+            # TODO: should be .0  or .1 so we can use list
+            # but this also roughly matches uplink_0, downlink_0
             fflow = self.scenario_cfg["options"]["flow"]
             for index, src in enumerate(fflow.get("src_ip", [])):
-                flow["src_ip{}".format(index)] = self._get_ip_flow_range(src)
+                flow["src_ip_{}".format(index)] = self._get_ip_flow_range(src)
 
             for index, dst in enumerate(fflow.get("dst_ip", [])):
-                flow["dst_ip{}".format(index)] = self._get_ip_flow_range(dst)
+                flow["dst_ip_{}".format(index)] = self._get_ip_flow_range(dst)
+
+            for index, publicip in enumerate(fflow.get("public_ip", [])):
+                flow["public_ip_{}".format(index)] = publicip
+
+            for index, src_port in enumerate(fflow.get("src_port", [])):
+                flow["src_port_{}".format(index)] = src_port
 
-            for index, publicip in enumerate(fflow.get("publicip", [])):
-                flow["public_ip{}".format(index)] = publicip
+            for index, dst_port in enumerate(fflow.get("dst_port", [])):
+                flow["dst_port_{}".format(index)] = dst_port
+
+            flow["count"] = fflow["count"]
         except KeyError:
             flow = {}
         return {"flow": flow}
@@ -180,38 +132,72 @@ class NetworkServiceTestCase(base.Scenario):
     def _get_traffic_profile(self):
         profile = self.scenario_cfg["traffic_profile"]
         path = self.scenario_cfg["task_path"]
-        with open_relative_file(profile, path) as infile:
+        with utils.open_relative_file(profile, path) as infile:
             return infile.read()
 
+    def _get_duration(self):
+        options = self.scenario_cfg.get('options', {})
+        return options.get('duration',
+                           tprofile_base.TrafficProfileConfig.DEFAULT_DURATION)
+
     def _fill_traffic_profile(self):
-        traffic_mapping = self._get_traffic_profile()
-        traffic_map_data = {
+        tprofile = self._get_traffic_profile()
+        extra_args = self.scenario_cfg.get('extra_args', {})
+        tprofile_data = {
             'flow': self._get_traffic_flow(),
             'imix': self._get_traffic_imix(),
-            'private': {},
-            'public': {},
-        }
+            tprofile_base.TrafficProfile.UPLINK: {},
+            tprofile_base.TrafficProfile.DOWNLINK: {},
+            'extra_args': extra_args,
+            'duration': self._get_duration()}
+        traffic_vnfd = vnfdgen.generate_vnfd(tprofile, tprofile_data)
+        self.traffic_profile = tprofile_base.TrafficProfile.get(traffic_vnfd)
+
+    def _get_topology(self):
+        topology = self.scenario_cfg["topology"]
+        path = self.scenario_cfg["task_path"]
+        with utils.open_relative_file(topology, path) as infile:
+            return infile.read()
 
-        traffic_vnfd = vnfdgen.generate_vnfd(traffic_mapping, traffic_map_data)
-        self.traffic_profile = TrafficProfile.get(traffic_vnfd)
-        return self.traffic_profile
+    def _render_topology(self):
+        topology = self._get_topology()
+        topology_args = self.scenario_cfg.get('extra_args', {})
+        topolgy_data = {
+            'extra_args': topology_args
+        }
+        topology_yaml = vnfdgen.generate_vnfd(topology, topolgy_data)
+        self.topology = topology_yaml["nsd:nsd-catalog"]["nsd"][0]
 
     def _find_vnf_name_from_id(self, vnf_id):
         return next((vnfd["vnfd-id-ref"]
                      for vnfd in self.topology["constituent-vnfd"]
                      if vnf_id == vnfd["member-vnf-index"]), None)
 
+    def _find_vnfd_from_vnf_idx(self, vnf_id):
+        return next((vnfd
+                     for vnfd in self.topology["constituent-vnfd"]
+                     if vnf_id == vnfd["member-vnf-index"]), None)
+
     @staticmethod
-    def get_vld_networks(networks):
-        return {n['vld_id']: n for n in networks.values()}
+    def find_node_if(nodes, name, if_name, vld_id):
+        try:
+            # check for xe0, xe1
+            intf = nodes[name]["interfaces"][if_name]
+        except KeyError:
+            # if not xe0, then maybe vld_id,  uplink_0, downlink_0
+            # pop it and re-insert with the correct name from topology
+            intf = nodes[name]["interfaces"].pop(vld_id)
+            nodes[name]["interfaces"][if_name] = intf
+        return intf
 
     def _resolve_topology(self):
         for vld in self.topology["vld"]:
             try:
                 node0_data, node1_data = vld["vnfd-connection-point-ref"]
             except (ValueError, TypeError):
-                raise IncorrectConfig("Topology file corrupted, "
-                                      "wrong endpoint count for connection")
+                raise exceptions.IncorrectConfig(
+                    error_msg='Topology file corrupted, wrong endpoint count '
+                              'for connection')
 
             node0_name = self._find_vnf_name_from_id(node0_data["member-vnf-index-ref"])
             node1_name = self._find_vnf_name_from_id(node1_data["member-vnf-index-ref"])
@@ -221,8 +207,8 @@ class NetworkServiceTestCase(base.Scenario):
 
             try:
                 nodes = self.context_cfg["nodes"]
-                node0_if = nodes[node0_name]["interfaces"][node0_if_name]
-                node1_if = nodes[node1_name]["interfaces"][node1_if_name]
+                node0_if = self.find_node_if(nodes, node0_name, node0_if_name, vld["id"])
+                node1_if = self.find_node_if(nodes, node1_name, node1_if_name, vld["id"])
 
                 # names so we can do reverse lookups
                 node0_if["ifname"] = node0_if_name
@@ -231,7 +217,6 @@ class NetworkServiceTestCase(base.Scenario):
                 node0_if["node_name"] = node0_name
                 node1_if["node_name"] = node1_name
 
-                vld_networks = self.get_vld_networks(self.context_cfg["networks"])
                 node0_if["vld_id"] = vld["id"]
                 node1_if["vld_id"] = vld["id"]
 
@@ -244,6 +229,9 @@ class NetworkServiceTestCase(base.Scenario):
                 node1_if["peer_ifname"] = node0_if_name
 
                 # just load the network
+                vld_networks = {n.get('vld_id', name): n for name, n in
+                                self.context_cfg["networks"].items()}
+
                 node0_if["network"] = vld_networks.get(vld["id"], {})
                 node1_if["network"] = vld_networks.get(vld["id"], {})
 
@@ -255,15 +243,17 @@ class NetworkServiceTestCase(base.Scenario):
 
             except KeyError:
                 LOG.exception("")
-                raise IncorrectConfig("Required interface not found, "
-                                      "topology file corrupted")
+                raise exceptions.IncorrectConfig(
+                    error_msg='Required interface not found, topology file '
+                              'corrupted')
 
         for vld in self.topology['vld']:
             try:
                 node0_data, node1_data = vld["vnfd-connection-point-ref"]
             except (ValueError, TypeError):
-                raise IncorrectConfig("Topology file corrupted, "
-                                      "wrong endpoint count for connection")
+                raise exceptions.IncorrectConfig(
+                    error_msg='Topology file corrupted, wrong endpoint count '
+                              'for connection')
 
             node0_name = self._find_vnf_name_from_id(node0_data["member-vnf-index-ref"])
             node1_name = self._find_vnf_name_from_id(node1_data["member-vnf-index-ref"])
@@ -272,8 +262,8 @@ class NetworkServiceTestCase(base.Scenario):
             node1_if_name = node1_data["vnfd-connection-point-ref"]
 
             nodes = self.context_cfg["nodes"]
-            node0_if = nodes[node0_name]["interfaces"][node0_if_name]
-            node1_if = nodes[node1_name]["interfaces"][node1_if_name]
+            node0_if = self.find_node_if(nodes, node0_name, node0_if_name, vld["id"])
+            node1_if = self.find_node_if(nodes, node1_name, node1_if_name, vld["id"])
 
             # add peer interface dict, but remove circular link
             # TODO: don't waste memory
@@ -282,10 +272,6 @@ class NetworkServiceTestCase(base.Scenario):
             node0_if["peer_intf"] = node1_copy
             node1_if["peer_intf"] = node0_copy
 
-    def _find_vnfd_from_vnf_idx(self, vnf_idx):
-        return next((vnfd for vnfd in self.topology["constituent-vnfd"]
-                     if vnf_idx == vnfd["member-vnf-index"]), None)
-
     def _update_context_with_topology(self):
         for vnfd in self.topology["constituent-vnfd"]:
             vnf_idx = vnfd["member-vnf-index"]
@@ -293,44 +279,30 @@ class NetworkServiceTestCase(base.Scenario):
             vnfd = self._find_vnfd_from_vnf_idx(vnf_idx)
             self.context_cfg["nodes"][vnf_name].update(vnfd)
 
-    @staticmethod
-    def _sort_dpdk_port_num(netdevs):
-        # dpdk_port_num is PCI BUS ID ordering, lowest first
-        s = sorted(netdevs.values(), key=itemgetter('pci_bus_id'))
-        for dpdk_port_num, netdev in enumerate(s):
-            netdev['dpdk_port_num'] = dpdk_port_num
-
-    def _probe_netdevs(self, node, node_dict):
-        cmd = "PATH=$PATH:/sbin:/usr/sbin ip addr show"
-        netdevs = {}
-        with SshManager(node_dict) as conn:
-            if conn:
-                exit_status = conn.execute(cmd)[0]
-                if exit_status != 0:
-                    raise IncorrectSetup("Node's %s lacks ip tool." % node)
-                exit_status, stdout, _ = conn.execute(
-                    self.FIND_NETDEVICE_STRING)
-                if exit_status != 0:
-                    raise IncorrectSetup(
-                        "Cannot find netdev info in sysfs" % node)
-                netdevs = node_dict['netdevs'] = self.parse_netdev_info(stdout)
-        return netdevs
-
-    @classmethod
-    def _probe_missing_values(cls, netdevs, network):
-
-        mac_lower = network['local_mac'].lower()
-        for netdev in netdevs.values():
-            if netdev['address'].lower() != mac_lower:
-                continue
-            network.update({
-                'driver': netdev['driver'],
-                'vpci': netdev['pci_bus_id'],
-                'ifindex': netdev['ifindex'],
-            })
+    def _generate_pod_yaml(self):
+        context_yaml = os.path.join(LOG_DIR, "pod-{}.yaml".format(self.scenario_cfg['task_id']))
+        # convert OrderedDict to a list
+        # pod.yaml nodes is a list
+        nodes = [self._serialize_node(node) for node in self.context_cfg["nodes"].values()]
+        pod_dict = {
+            "nodes": nodes,
+            "networks": self.context_cfg["networks"]
+        }
+        with open(context_yaml, "w") as context_out:
+            yaml.safe_dump(pod_dict, context_out, default_flow_style=False,
+                           explicit_start=True)
 
-    TOPOLOGY_REQUIRED_KEYS = frozenset({
-        "vpci", "local_ip", "netmask", "local_mac", "driver"})
+    @staticmethod
+    def _serialize_node(node):
+        new_node = copy.deepcopy(node)
+        # name field is required
+        # remove context suffix
+        new_node["name"] = node['name'].split('.')[0]
+        try:
+            new_node["pkey"] = ssh.convert_key_to_str(node["pkey"])
+        except KeyError:
+            pass
+        return new_node
 
     def map_topology_to_infrastructure(self):
         """ This method should verify if the available resources defined in pod.yaml
@@ -338,67 +310,10 @@ class NetworkServiceTestCase(base.Scenario):
 
         :return: None. Side effect: context_cfg is updated
         """
-        for node, node_dict in self.context_cfg["nodes"].items():
-
-            for network in node_dict["interfaces"].values():
-                missing = self.TOPOLOGY_REQUIRED_KEYS.difference(network)
-                if not missing:
-                    continue
-
-                # only ssh probe if there are missing values
-                # ssh probe won't work on Ixia, so we had better define all our values
-                try:
-                    netdevs = self._probe_netdevs(node, node_dict)
-                except (SSHError, SSHTimeout):
-                    raise IncorrectConfig(
-                        "Unable to probe missing interface fields '%s', on node %s "
-                        "SSH Error" % (', '.join(missing), node))
-                try:
-                    self._probe_missing_values(netdevs, network)
-                except KeyError:
-                    pass
-                else:
-                    missing = self.TOPOLOGY_REQUIRED_KEYS.difference(
-                        network)
-                if missing:
-                    raise IncorrectConfig(
-                        "Require interface fields '%s' not found, topology file "
-                        "corrupted" % ', '.join(missing))
-
         # 3. Use topology file to find connections & resolve dest address
         self._resolve_topology()
         self._update_context_with_topology()
 
-    FIND_NETDEVICE_STRING = r"""find /sys/devices/pci* -type d -name net -exec sh -c '{ grep -sH ^ \
-$1/ifindex $1/address $1/operstate $1/device/vendor $1/device/device \
-$1/device/subsystem_vendor $1/device/subsystem_device ; \
-printf "%s/driver:" $1 ; basename $(readlink -s $1/device/driver); } \
-' sh  \{\}/* \;
-"""
-    BASE_ADAPTER_RE = re.compile(
-        '^/sys/devices/(.*)/net/([^/]*)/([^:]*):(.*)$', re.M)
-
-    @classmethod
-    def parse_netdev_info(cls, stdout):
-        network_devices = defaultdict(dict)
-        matches = cls.BASE_ADAPTER_RE.findall(stdout)
-        for bus_path, interface_name, name, value in matches:
-            dirname, bus_id = os.path.split(bus_path)
-            if 'virtio' in bus_id:
-                # for some stupid reason VMs include virtio1/
-                # in PCI device path
-                bus_id = os.path.basename(dirname)
-            # remove extra 'device/' from 'device/vendor,
-            # device/subsystem_vendor', etc.
-            if 'device/' in name:
-                name = name.split('/')[1]
-            network_devices[interface_name][name] = value
-            network_devices[interface_name][
-                'interface_name'] = interface_name
-            network_devices[interface_name]['pci_bus_id'] = bus_id
-        # convert back to regular dict
-        return dict(network_devices)
-
     @classmethod
     def get_vnf_impl(cls, vnf_model_id):
         """ Find the implementing class from vnf_model["vnf"]["name"] field
@@ -406,13 +321,14 @@ printf "%s/driver:" $1 ; basename $(readlink -s $1/device/driver); } \
         :param vnf_model_id: parsed vnfd model ID field
         :return: subclass of GenericVNF
         """
-        import_modules_from_package(
+        utils.import_modules_from_package(
             "yardstick.network_services.vnf_generic.vnf")
         expected_name = vnf_model_id
         classes_found = []
 
         def impl():
-            for name, class_ in ((c.__name__, c) for c in itersubclasses(GenericVNF)):
+            for name, class_ in ((c.__name__, c) for c in
+                                 utils.itersubclasses(GenericVNF)):
                 if name == expected_name:
                     yield class_
                 classes_found.append(name)
@@ -422,14 +338,31 @@ printf "%s/driver:" $1 ; basename $(readlink -s $1/device/driver); } \
         except StopIteration:
             pass
 
-        raise IncorrectConfig("No implementation for %s found in %s" %
-                              (expected_name, classes_found))
+        message = ('No implementation for %s found in %s'
+                   % (expected_name, classes_found))
+        raise exceptions.IncorrectConfig(error_msg=message)
 
     @staticmethod
-    def update_interfaces_from_node(vnfd, node):
-        for intf in vnfd["vdu"][0]["external-interface"]:
-            node_intf = node['interfaces'][intf['name']]
-            intf['virtual-interface'].update(node_intf)
+    def create_interfaces_from_node(vnfd, node):
+        ext_intfs = vnfd["vdu"][0]["external-interface"] = []
+        # have to sort so xe0 goes first
+        for intf_name, intf in sorted(node['interfaces'].items()):
+            # only interfaces with vld_id are added.
+            # Thus there are two layers of filters, only intefaces with vld_id
+            # show up in interfaces, and only interfaces with traffic profiles
+            # are used by the generators
+            if intf.get('vld_id'):
+                # force dpkd_port_num to int so we can do reverse lookup
+                try:
+                    intf['dpdk_port_num'] = int(intf['dpdk_port_num'])
+                except KeyError:
+                    pass
+                ext_intf = {
+                    "name": intf_name,
+                    "virtual-interface": intf,
+                    "vnfd-connection-point-ref": intf_name,
+                }
+                ext_intfs.append(ext_intf)
 
     def load_vnf_models(self, scenario_cfg=None, context_cfg=None):
         """ Create VNF objects based on YAML descriptors
@@ -439,6 +372,9 @@ printf "%s/driver:" $1 ; basename $(readlink -s $1/device/driver); } \
         :param context_cfg:
         :return:
         """
+        trex_lib_path = get_nsb_option('trex_client_lib')
+        sys.path[:] = list(chain([trex_lib_path], (x for x in sys.path if x != trex_lib_path)))
+
         if scenario_cfg is None:
             scenario_cfg = self.scenario_cfg
 
@@ -446,17 +382,28 @@ printf "%s/driver:" $1 ; basename $(readlink -s $1/device/driver); } \
             context_cfg = self.context_cfg
 
         vnfs = []
-        # we assume OrderedDict for consistenct in instantiation
+        # we assume OrderedDict for consistency in instantiation
         for node_name, node in context_cfg["nodes"].items():
             LOG.debug(node)
-            file_name = node["VNF model"]
+            try:
+                file_name = node["VNF model"]
+            except KeyError:
+                LOG.debug("no model for %s, skipping", node_name)
+                continue
             file_path = scenario_cfg['task_path']
-            with open_relative_file(file_name, file_path) as stream:
+            with utils.open_relative_file(file_name, file_path) as stream:
                 vnf_model = stream.read()
             vnfd = vnfdgen.generate_vnfd(vnf_model, node)
             # TODO: here add extra context_cfg["nodes"] regardless of template
             vnfd = vnfd["vnfd:vnfd-catalog"]["vnfd"][0]
-            self.update_interfaces_from_node(vnfd, node)
+            # force inject pkey if it exists
+            # we want to standardize Heat using pkey as a string so we don't rely
+            # on the filesystem
+            try:
+                vnfd['mgmt-interface']['pkey'] = node['pkey']
+            except KeyError:
+                pass
+            self.create_interfaces_from_node(vnfd, node)
             vnf_impl = self.get_vnf_impl(vnfd['id'])
             vnf_instance = vnf_impl(node_name, vnfd)
             vnfs.append(vnf_instance)
@@ -488,18 +435,22 @@ printf "%s/driver:" $1 ; basename $(readlink -s $1/device/driver); } \
                 vnf.instantiate(self.scenario_cfg, self.context_cfg)
                 LOG.info("Waiting for %s to instantiate", vnf.name)
                 vnf.wait_for_instantiate()
-        except RuntimeError:
+        except:
+            LOG.exception("")
             for vnf in self.vnfs:
                 vnf.terminate()
             raise
 
+        # we have to generate pod.yaml here after VNF has probed so we know vpci and driver
+        self._generate_pod_yaml()
+
         # 3. Run experiment
         # Start listeners first to avoid losing packets
         for traffic_gen in traffic_runners:
             traffic_gen.listen_traffic(self.traffic_profile)
 
         # register collector with yardstick for KPI collection.
-        self.collector = Collector(self.vnfs, self.traffic_profile)
+        self.collector = Collector(self.vnfs, context_base.Context.get_physical_nodes())
         self.collector.start()
 
         # Start the actual traffic
@@ -515,11 +466,11 @@ printf "%s/driver:" $1 ; basename $(readlink -s $1/device/driver); } \
         :return: None
         """
 
-        for vnf in self.vnfs:
-            # Result example:
-            # {"VNF1: { "tput" : [1000, 999] }, "VNF2": { "latency": 100 }}
-            LOG.debug("vnf")
-            result.update(self.collector.get_kpi(vnf))
+        # this is the only method that is check from the runner
+        # so if we have any fatal error it must be raised via these methods
+        # otherwise we will not terminate
+
+        result.update(self.collector.get_kpi())
 
     def teardown(self):
         """ Stop the collector and terminate VNF & TG instance
@@ -527,7 +478,27 @@ printf "%s/driver:" $1 ; basename $(readlink -s $1/device/driver); } \
         :return
         """
 
-        self.collector.stop()
-        for vnf in self.vnfs:
-            LOG.info("Stopping %s", vnf.name)
-            vnf.terminate()
+        try:
+            try:
+                self.collector.stop()
+                for vnf in self.vnfs:
+                    LOG.info("Stopping %s", vnf.name)
+                    vnf.terminate()
+                LOG.debug("all VNFs terminated: %s", ", ".join(vnf.name for vnf in self.vnfs))
+            finally:
+                terminate_children()
+        except Exception:
+            # catch any exception in teardown and convert to simple exception
+            # never pass exceptions back to multiprocessing, because some exceptions can
+            # be unpicklable
+            # https://bugs.python.org/issue9400
+            LOG.exception("")
+            raise RuntimeError("Error in teardown")
+
+    def pre_run_wait_time(self, time_seconds):
+        """Time waited before executing the run method"""
+        time.sleep(time_seconds)
+
+    def post_run_wait_time(self, time_seconds):
+        """Time waited after executing the run method"""
+        pass