1 # Copyright (c) 2016-2017 Intel Corporation
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
7 # http://www.apache.org/licenses/LICENSE-2.0
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14 """ NSPerf specific scenario definition """
16 from __future__ import absolute_import
27 from itertools import chain
31 from collections import defaultdict
33 from yardstick.benchmark.scenarios import base
34 from yardstick.common.constants import LOG_DIR
35 from yardstick.common.process import terminate_children
36 from yardstick.common.utils import import_modules_from_package, itersubclasses
37 from yardstick.common.yaml_loader import yaml_load
38 from yardstick.network_services.collector.subscriber import Collector
39 from yardstick.network_services.vnf_generic import vnfdgen
40 from yardstick.network_services.vnf_generic.vnf.base import GenericVNF
41 from yardstick.network_services.traffic_profile.base import TrafficProfile
42 from yardstick.network_services.utils import get_nsb_option
43 from yardstick import ssh
46 LOG = logging.getLogger(__name__)
49 class SSHError(Exception):
50 """Class handles ssh connection error exception"""
54 class SSHTimeout(SSHError):
55 """Class handles ssh connection timeout exception"""
59 class IncorrectConfig(Exception):
60 """Class handles incorrect configuration during setup"""
64 class IncorrectSetup(Exception):
65 """Class handles incorrect setup during setup"""
69 class SshManager(object):
70 def __init__(self, node, timeout=120):
71 super(SshManager, self).__init__()
74 self.timeout = timeout
78 args -> network device mappings
79 returns -> ssh connection ready to be used
82 self.conn = ssh.SSH.from_node(self.node)
83 self.conn.wait(timeout=self.timeout)
84 except SSHError as error:
85 LOG.info("connect failed to %s, due to %s", self.node["ip"], error)
86 # self.conn defaults to None
89 def __exit__(self, exc_type, exc_val, exc_tb):
94 def find_relative_file(path, task_path):
96 Find file in one of places: in abs of path or
97 relative to TC scenario file. In this order.
101 :return str: full path to file
103 # fixme: create schema to validate all fields have been provided
104 for lookup in [os.path.abspath(path), os.path.join(task_path, path)]:
110 raise IOError(errno.ENOENT, 'Unable to find {} file'.format(path))
113 def open_relative_file(path, task_path):
117 if e.errno == errno.ENOENT:
118 return open(os.path.join(task_path, path))
122 class NetworkServiceTestCase(base.Scenario):
123 """Class handles Generic framework to do pre-deployment VNF &
124 Network service testing """
126 __scenario_type__ = "NSPerf"
128 def __init__(self, scenario_cfg, context_cfg): # Yardstick API
129 super(NetworkServiceTestCase, self).__init__()
130 self.scenario_cfg = scenario_cfg
131 self.context_cfg = context_cfg
133 # fixme: create schema to validate all fields have been provided
134 with open_relative_file(scenario_cfg["topology"],
135 scenario_cfg['task_path']) as stream:
136 topology_yaml = yaml_load(stream)
138 self.topology = topology_yaml["nsd:nsd-catalog"]["nsd"][0]
140 self.collector = None
141 self.traffic_profile = None
142 self.node_netdevs = {}
144 def _get_ip_flow_range(self, ip_start_range):
146 # IP range is specified as 'x.x.x.x-y.y.y.y'
147 if isinstance(ip_start_range, six.string_types):
148 return ip_start_range
150 node_name, range_or_interface = next(iter(ip_start_range.items()), (None, '0.0.0.0'))
151 if node_name is None:
152 # we are manually specifying the range
153 ip_addr_range = range_or_interface
155 node = self.context_cfg["nodes"].get(node_name, {})
157 # the ip_range is the interface name
158 interface = node.get("interfaces", {})[range_or_interface]
161 mask = "255.255.255.0"
163 ip = interface["local_ip"]
164 # we can't default these values, they must both exist to be valid
165 mask = interface["netmask"]
167 ipaddr = ipaddress.ip_network(six.text_type('{}/{}'.format(ip, mask)), strict=False)
168 hosts = list(ipaddr.hosts())
170 # skip the first host in case of gateway
171 ip_addr_range = "{}-{}".format(hosts[1], hosts[-1])
173 LOG.warning("Only single IP in range %s", ipaddr)
174 # fall back to single IP range
178 def _get_traffic_flow(self):
181 # TODO: should be .0 or .1 so we can use list
182 # but this also roughly matches uplink_0, downlink_0
183 fflow = self.scenario_cfg["options"]["flow"]
184 for index, src in enumerate(fflow.get("src_ip", [])):
185 flow["src_ip_{}".format(index)] = self._get_ip_flow_range(src)
187 for index, dst in enumerate(fflow.get("dst_ip", [])):
188 flow["dst_ip_{}".format(index)] = self._get_ip_flow_range(dst)
190 for index, publicip in enumerate(fflow.get("public_ip", [])):
191 flow["public_ip_{}".format(index)] = publicip
193 for index, src_port in enumerate(fflow.get("src_port", [])):
194 flow["src_port_{}".format(index)] = src_port
196 for index, dst_port in enumerate(fflow.get("dst_port", [])):
197 flow["dst_port_{}".format(index)] = dst_port
199 flow["count"] = fflow["count"]
202 return {"flow": flow}
204 def _get_traffic_imix(self):
206 imix = {"imix": self.scenario_cfg['options']['framesize']}
211 def _get_traffic_profile(self):
212 profile = self.scenario_cfg["traffic_profile"]
213 path = self.scenario_cfg["task_path"]
214 with open_relative_file(profile, path) as infile:
217 def _fill_traffic_profile(self):
218 traffic_mapping = self._get_traffic_profile()
220 'flow': self._get_traffic_flow(),
221 'imix': self._get_traffic_imix(),
222 TrafficProfile.UPLINK: {},
223 TrafficProfile.DOWNLINK: {},
226 traffic_vnfd = vnfdgen.generate_vnfd(traffic_mapping, traffic_map_data)
227 self.traffic_profile = TrafficProfile.get(traffic_vnfd)
228 return self.traffic_profile
230 def _find_vnf_name_from_id(self, vnf_id):
231 return next((vnfd["vnfd-id-ref"]
232 for vnfd in self.topology["constituent-vnfd"]
233 if vnf_id == vnfd["member-vnf-index"]), None)
236 def get_vld_networks(networks):
237 # network name is vld_id
239 for name, n in networks.items():
241 vld_map[n['vld_id']] = n
247 def find_node_if(nodes, name, if_name, vld_id):
250 intf = nodes[name]["interfaces"][if_name]
252 # if not xe0, then maybe vld_id, uplink_0, downlink_0
253 # pop it and re-insert with the correct name from topology
254 intf = nodes[name]["interfaces"].pop(vld_id)
255 nodes[name]["interfaces"][if_name] = intf
258 def _resolve_topology(self):
259 for vld in self.topology["vld"]:
261 node0_data, node1_data = vld["vnfd-connection-point-ref"]
262 except (ValueError, TypeError):
263 raise IncorrectConfig("Topology file corrupted, "
264 "wrong endpoint count for connection")
266 node0_name = self._find_vnf_name_from_id(node0_data["member-vnf-index-ref"])
267 node1_name = self._find_vnf_name_from_id(node1_data["member-vnf-index-ref"])
269 node0_if_name = node0_data["vnfd-connection-point-ref"]
270 node1_if_name = node1_data["vnfd-connection-point-ref"]
273 nodes = self.context_cfg["nodes"]
274 node0_if = self.find_node_if(nodes, node0_name, node0_if_name, vld["id"])
275 node1_if = self.find_node_if(nodes, node1_name, node1_if_name, vld["id"])
277 # names so we can do reverse lookups
278 node0_if["ifname"] = node0_if_name
279 node1_if["ifname"] = node1_if_name
281 node0_if["node_name"] = node0_name
282 node1_if["node_name"] = node1_name
284 node0_if["vld_id"] = vld["id"]
285 node1_if["vld_id"] = vld["id"]
288 node0_if["peer_name"] = node1_name
289 node1_if["peer_name"] = node0_name
291 # set peer interface name
292 node0_if["peer_ifname"] = node1_if_name
293 node1_if["peer_ifname"] = node0_if_name
295 # just load the network
296 vld_networks = self.get_vld_networks(self.context_cfg["networks"])
297 node0_if["network"] = vld_networks.get(vld["id"], {})
298 node1_if["network"] = vld_networks.get(vld["id"], {})
300 node0_if["dst_mac"] = node1_if["local_mac"]
301 node0_if["dst_ip"] = node1_if["local_ip"]
303 node1_if["dst_mac"] = node0_if["local_mac"]
304 node1_if["dst_ip"] = node0_if["local_ip"]
308 raise IncorrectConfig("Required interface not found, "
309 "topology file corrupted")
311 for vld in self.topology['vld']:
313 node0_data, node1_data = vld["vnfd-connection-point-ref"]
314 except (ValueError, TypeError):
315 raise IncorrectConfig("Topology file corrupted, "
316 "wrong endpoint count for connection")
318 node0_name = self._find_vnf_name_from_id(node0_data["member-vnf-index-ref"])
319 node1_name = self._find_vnf_name_from_id(node1_data["member-vnf-index-ref"])
321 node0_if_name = node0_data["vnfd-connection-point-ref"]
322 node1_if_name = node1_data["vnfd-connection-point-ref"]
324 nodes = self.context_cfg["nodes"]
325 node0_if = self.find_node_if(nodes, node0_name, node0_if_name, vld["id"])
326 node1_if = self.find_node_if(nodes, node1_name, node1_if_name, vld["id"])
328 # add peer interface dict, but remove circular link
329 # TODO: don't waste memory
330 node0_copy = node0_if.copy()
331 node1_copy = node1_if.copy()
332 node0_if["peer_intf"] = node1_copy
333 node1_if["peer_intf"] = node0_copy
335 def _find_vnfd_from_vnf_idx(self, vnf_idx):
336 return next((vnfd for vnfd in self.topology["constituent-vnfd"]
337 if vnf_idx == vnfd["member-vnf-index"]), None)
339 def _update_context_with_topology(self):
340 for vnfd in self.topology["constituent-vnfd"]:
341 vnf_idx = vnfd["member-vnf-index"]
342 vnf_name = self._find_vnf_name_from_id(vnf_idx)
343 vnfd = self._find_vnfd_from_vnf_idx(vnf_idx)
344 self.context_cfg["nodes"][vnf_name].update(vnfd)
346 def _probe_netdevs(self, node, node_dict, timeout=120):
348 return self.node_netdevs[node]
353 cmd = "PATH=$PATH:/sbin:/usr/sbin ip addr show"
355 with SshManager(node_dict, timeout=timeout) as conn:
357 exit_status = conn.execute(cmd)[0]
359 raise IncorrectSetup("Node's %s lacks ip tool." % node)
360 exit_status, stdout, _ = conn.execute(
361 self.FIND_NETDEVICE_STRING)
363 raise IncorrectSetup(
364 "Cannot find netdev info in sysfs" % node)
365 netdevs = node_dict['netdevs'] = self.parse_netdev_info(stdout)
367 self.node_netdevs[node] = netdevs
371 def _probe_missing_values(cls, netdevs, network):
373 mac_lower = network['local_mac'].lower()
374 for netdev in netdevs.values():
375 if netdev['address'].lower() != mac_lower:
378 'driver': netdev['driver'],
379 'vpci': netdev['pci_bus_id'],
380 'ifindex': netdev['ifindex'],
383 def _generate_pod_yaml(self):
384 context_yaml = os.path.join(LOG_DIR, "pod-{}.yaml".format(self.scenario_cfg['task_id']))
385 # convert OrderedDict to a list
386 # pod.yaml nodes is a list
387 nodes = [self._serialize_node(node) for node in self.context_cfg["nodes"].values()]
390 "networks": self.context_cfg["networks"]
392 with open(context_yaml, "w") as context_out:
393 yaml.safe_dump(pod_dict, context_out, default_flow_style=False,
397 def _serialize_node(node):
398 new_node = copy.deepcopy(node)
399 # name field is required
400 # remove context suffix
401 new_node["name"] = node['name'].split('.')[0]
403 new_node["pkey"] = ssh.convert_key_to_str(node["pkey"])
408 TOPOLOGY_REQUIRED_KEYS = frozenset({
409 "vpci", "local_ip", "netmask", "local_mac", "driver"})
411 def map_topology_to_infrastructure(self):
412 """ This method should verify if the available resources defined in pod.yaml
413 match the topology.yaml file.
415 :return: None. Side effect: context_cfg is updated
417 num_nodes = len(self.context_cfg["nodes"])
418 # OpenStack instance creation time is probably proportional to the number
420 timeout = 120 * num_nodes
421 for node, node_dict in self.context_cfg["nodes"].items():
423 for network in node_dict["interfaces"].values():
424 missing = self.TOPOLOGY_REQUIRED_KEYS.difference(network)
428 # only ssh probe if there are missing values
429 # ssh probe won't work on Ixia, so we had better define all our values
431 netdevs = self._probe_netdevs(node, node_dict, timeout=timeout)
432 except (SSHError, SSHTimeout):
433 raise IncorrectConfig(
434 "Unable to probe missing interface fields '%s', on node %s "
435 "SSH Error" % (', '.join(missing), node))
437 self._probe_missing_values(netdevs, network)
441 missing = self.TOPOLOGY_REQUIRED_KEYS.difference(
444 raise IncorrectConfig(
445 "Require interface fields '%s' not found, topology file "
446 "corrupted" % ', '.join(missing))
448 # we have to generate pod.yaml here so we have vpci and driver
449 self._generate_pod_yaml()
450 # 3. Use topology file to find connections & resolve dest address
451 self._resolve_topology()
452 self._update_context_with_topology()
454 FIND_NETDEVICE_STRING = r"""find /sys/devices/pci* -type d -name net -exec sh -c '{ grep -sH ^ \
455 $1/ifindex $1/address $1/operstate $1/device/vendor $1/device/device \
456 $1/device/subsystem_vendor $1/device/subsystem_device ; \
457 printf "%s/driver:" $1 ; basename $(readlink -s $1/device/driver); } \
460 BASE_ADAPTER_RE = re.compile(
461 '^/sys/devices/(.*)/net/([^/]*)/([^:]*):(.*)$', re.M)
464 def parse_netdev_info(cls, stdout):
465 network_devices = defaultdict(dict)
466 matches = cls.BASE_ADAPTER_RE.findall(stdout)
467 for bus_path, interface_name, name, value in matches:
468 dirname, bus_id = os.path.split(bus_path)
469 if 'virtio' in bus_id:
470 # for some stupid reason VMs include virtio1/
472 bus_id = os.path.basename(dirname)
473 # remove extra 'device/' from 'device/vendor,
474 # device/subsystem_vendor', etc.
475 if 'device/' in name:
476 name = name.split('/')[1]
477 network_devices[interface_name][name] = value
478 network_devices[interface_name][
479 'interface_name'] = interface_name
480 network_devices[interface_name]['pci_bus_id'] = bus_id
481 # convert back to regular dict
482 return dict(network_devices)
485 def get_vnf_impl(cls, vnf_model_id):
486 """ Find the implementing class from vnf_model["vnf"]["name"] field
488 :param vnf_model_id: parsed vnfd model ID field
489 :return: subclass of GenericVNF
491 import_modules_from_package(
492 "yardstick.network_services.vnf_generic.vnf")
493 expected_name = vnf_model_id
497 for name, class_ in ((c.__name__, c) for c in itersubclasses(GenericVNF)):
498 if name == expected_name:
500 classes_found.append(name)
504 except StopIteration:
507 raise IncorrectConfig("No implementation for %s found in %s" %
508 (expected_name, classes_found))
511 def create_interfaces_from_node(vnfd, node):
512 ext_intfs = vnfd["vdu"][0]["external-interface"] = []
513 # have to sort so xe0 goes first
514 for intf_name, intf in sorted(node['interfaces'].items()):
515 # only interfaces with vld_id are added.
516 # Thus there are two layers of filters, only intefaces with vld_id
517 # show up in interfaces, and only interfaces with traffic profiles
518 # are used by the generators
519 if intf.get('vld_id'):
520 # force dpkd_port_num to int so we can do reverse lookup
522 intf['dpdk_port_num'] = int(intf['dpdk_port_num'])
527 "virtual-interface": intf,
528 "vnfd-connection-point-ref": intf_name,
530 ext_intfs.append(ext_intf)
532 def load_vnf_models(self, scenario_cfg=None, context_cfg=None):
533 """ Create VNF objects based on YAML descriptors
540 trex_lib_path = get_nsb_option('trex_client_lib')
541 sys.path[:] = list(chain([trex_lib_path], (x for x in sys.path if x != trex_lib_path)))
543 if scenario_cfg is None:
544 scenario_cfg = self.scenario_cfg
546 if context_cfg is None:
547 context_cfg = self.context_cfg
550 # we assume OrderedDict for consistenct in instantiation
551 for node_name, node in context_cfg["nodes"].items():
554 file_name = node["VNF model"]
556 LOG.debug("no model for %s, skipping", node_name)
558 file_path = scenario_cfg['task_path']
559 with open_relative_file(file_name, file_path) as stream:
560 vnf_model = stream.read()
561 vnfd = vnfdgen.generate_vnfd(vnf_model, node)
562 # TODO: here add extra context_cfg["nodes"] regardless of template
563 vnfd = vnfd["vnfd:vnfd-catalog"]["vnfd"][0]
564 # force inject pkey if it exists
565 # we want to standardize Heat using pkey as a string so we don't rely
568 vnfd['mgmt-interface']['pkey'] = node['pkey']
571 self.create_interfaces_from_node(vnfd, node)
572 vnf_impl = self.get_vnf_impl(vnfd['id'])
573 vnf_instance = vnf_impl(node_name, vnfd)
574 vnfs.append(vnf_instance)
580 """ Setup infrastructure, provission VNFs & start traffic
584 # 1. Verify if infrastructure mapping can meet topology
585 self.map_topology_to_infrastructure()
586 # 1a. Load VNF models
587 self.load_vnf_models()
588 # 1b. Fill traffic profile with information from topology
589 self._fill_traffic_profile()
593 # link events will cause VNF application to exit
594 # so we should start traffic runners before VNFs
595 traffic_runners = [vnf for vnf in self.vnfs if vnf.runs_traffic]
596 non_traffic_runners = [vnf for vnf in self.vnfs if not vnf.runs_traffic]
598 for vnf in chain(traffic_runners, non_traffic_runners):
599 LOG.info("Instantiating %s", vnf.name)
600 vnf.instantiate(self.scenario_cfg, self.context_cfg)
601 LOG.info("Waiting for %s to instantiate", vnf.name)
602 vnf.wait_for_instantiate()
605 for vnf in self.vnfs:
610 # Start listeners first to avoid losing packets
611 for traffic_gen in traffic_runners:
612 traffic_gen.listen_traffic(self.traffic_profile)
614 # register collector with yardstick for KPI collection.
615 self.collector = Collector(self.vnfs, self.context_cfg["nodes"], self.traffic_profile)
616 self.collector.start()
618 # Start the actual traffic
619 for traffic_gen in traffic_runners:
620 LOG.info("Starting traffic on %s", traffic_gen.name)
621 traffic_gen.run_traffic(self.traffic_profile)
623 def run(self, result): # yardstick API
624 """ Yardstick calls run() at intervals defined in the yaml and
625 produces timestamped samples
627 :param result: dictionary with results to update
631 # this is the only method that is check from the runner
632 # so if we have any fatal error it must be raised via these methods
633 # otherwise we will not terminate
635 result.update(self.collector.get_kpi())
638 """ Stop the collector and terminate VNF & TG instance
645 self.collector.stop()
646 for vnf in self.vnfs:
647 LOG.info("Stopping %s", vnf.name)
649 LOG.debug("all VNFs terminated: %s", ", ".join(vnf.name for vnf in self.vnfs))
653 # catch any exception in teardown and convert to simple exception
654 # never pass exceptions back to multiprocessing, because some exceptions can
656 # https://bugs.python.org/issue9400
658 raise RuntimeError("Error in teardown")