summaryrefslogtreecommitdiffstats
path: root/yardstick/benchmark/contexts/standalone/ovs_dpdk.py
diff options
context:
space:
mode:
Diffstat (limited to 'yardstick/benchmark/contexts/standalone/ovs_dpdk.py')
-rw-r--r--yardstick/benchmark/contexts/standalone/ovs_dpdk.py383
1 files changed, 383 insertions, 0 deletions
diff --git a/yardstick/benchmark/contexts/standalone/ovs_dpdk.py b/yardstick/benchmark/contexts/standalone/ovs_dpdk.py
new file mode 100644
index 000000000..833c3fb80
--- /dev/null
+++ b/yardstick/benchmark/contexts/standalone/ovs_dpdk.py
@@ -0,0 +1,383 @@
+# Copyright (c) 2016-2017 Intel Corporation
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+from __future__ import absolute_import
+import os
+import logging
+import collections
+import time
+
+from collections import OrderedDict
+
+from yardstick import ssh
+from yardstick.network_services.utils import get_nsb_option
+from yardstick.network_services.utils import provision_tool
+from yardstick.benchmark.contexts.base import Context
+from yardstick.benchmark.contexts.standalone.model import Libvirt
+from yardstick.benchmark.contexts.standalone.model import StandaloneContextHelper
+from yardstick.benchmark.contexts.standalone.model import Server
+from yardstick.benchmark.contexts.standalone.model import OvsDeploy
+from yardstick.network_services.utils import PciAddress
+
+LOG = logging.getLogger(__name__)
+
+
+class OvsDpdkContext(Context):
+ """ This class handles OVS standalone nodes - VM running on Non-Managed NFVi
+ Configuration: ovs_dpdk
+ """
+
+ __context_type__ = "StandaloneOvsDpdk"
+
+ SUPPORTED_OVS_TO_DPDK_MAP = {
+ '2.6.0': '16.07.1',
+ '2.6.1': '16.07.2',
+ '2.7.0': '16.11.1',
+ '2.7.1': '16.11.2',
+ '2.7.2': '16.11.3',
+ '2.8.0': '17.05.2'
+ }
+
+ DEFAULT_OVS = '2.6.0'
+
+ PKILL_TEMPLATE = "pkill %s %s"
+
+ def __init__(self):
+ self.file_path = None
+ self.sriov = []
+ self.first_run = True
+ self.dpdk_nic_bind = ""
+ self.vm_names = []
+ self.name = None
+ self.nfvi_host = []
+ self.nodes = []
+ self.networks = {}
+ self.attrs = {}
+ self.vm_flavor = None
+ self.servers = None
+ self.helper = StandaloneContextHelper()
+ self.vnf_node = Server()
+ self.ovs_properties = {}
+ self.wait_for_vswitchd = 10
+ super(OvsDpdkContext, self).__init__()
+
+ def init(self, attrs):
+ """initializes itself from the supplied arguments"""
+
+ self.name = attrs["name"]
+ self.file_path = attrs.get("file", "pod.yaml")
+
+ self.nodes, self.nfvi_host, self.host_mgmt = \
+ self.helper.parse_pod_file(self.file_path, 'OvsDpdk')
+
+ self.attrs = attrs
+ self.vm_flavor = attrs.get('flavor', {})
+ self.servers = attrs.get('servers', {})
+ self.vm_deploy = attrs.get("vm_deploy", True)
+ self.ovs_properties = attrs.get('ovs_properties', {})
+ # add optional static network definition
+ self.networks = attrs.get("networks", {})
+
+ LOG.debug("Nodes: %r", self.nodes)
+ LOG.debug("NFVi Node: %r", self.nfvi_host)
+ LOG.debug("Networks: %r", self.networks)
+
+ def setup_ovs(self):
+ vpath = self.ovs_properties.get("vpath", "/usr/local")
+ xargs_kill_cmd = self.PKILL_TEMPLATE % ('-9', 'ovs')
+
+ create_from = os.path.join(vpath, 'etc/openvswitch/conf.db')
+ create_to = os.path.join(vpath, 'share/openvswitch/vswitch.ovsschema')
+
+ cmd_list = [
+ "chmod 0666 /dev/vfio/*",
+ "chmod a+x /dev/vfio",
+ "pkill -9 ovs",
+ xargs_kill_cmd,
+ "killall -r 'ovs*'",
+ "mkdir -p {0}/etc/openvswitch".format(vpath),
+ "mkdir -p {0}/var/run/openvswitch".format(vpath),
+ "rm {0}/etc/openvswitch/conf.db".format(vpath),
+ "ovsdb-tool create {0} {1}".format(create_from, create_to),
+ "modprobe vfio-pci",
+ "chmod a+x /dev/vfio",
+ "chmod 0666 /dev/vfio/*",
+ ]
+ for cmd in cmd_list:
+ self.connection.execute(cmd)
+ bind_cmd = "{dpdk_nic_bind} --force -b {driver} {port}"
+ phy_driver = "vfio-pci"
+ for key, port in self.networks.items():
+ vpci = port.get("phy_port")
+ self.connection.execute(bind_cmd.format(dpdk_nic_bind=self.dpdk_nic_bind,
+ driver=phy_driver, port=vpci))
+
+ def start_ovs_serverswitch(self):
+ vpath = self.ovs_properties.get("vpath")
+ pmd_nums = int(self.ovs_properties.get("pmd_threads", 2))
+ ovs_sock_path = '/var/run/openvswitch/db.sock'
+ log_path = '/var/log/openvswitch/ovs-vswitchd.log'
+
+ pmd_mask = hex(sum(2 ** num for num in range(pmd_nums)) << 1)
+ socket0 = self.ovs_properties.get("ram", {}).get("socket_0", "2048")
+ socket1 = self.ovs_properties.get("ram", {}).get("socket_1", "2048")
+
+ ovs_other_config = "ovs-vsctl {0}set Open_vSwitch . other_config:{1}"
+ detach_cmd = "ovs-vswitchd unix:{0}{1} --pidfile --detach --log-file={2}"
+
+ cmd_list = [
+ "mkdir -p /usr/local/var/run/openvswitch",
+ "ovsdb-server --remote=punix:/{0}/{1} --pidfile --detach".format(vpath,
+ ovs_sock_path),
+ ovs_other_config.format("--no-wait ", "dpdk-init=true"),
+ ovs_other_config.format("--no-wait ", "dpdk-socket-mem='%s,%s'" % (socket0, socket1)),
+ detach_cmd.format(vpath, ovs_sock_path, log_path),
+ ovs_other_config.format("", "pmd-cpu-mask=%s" % pmd_mask),
+ ]
+
+ for cmd in cmd_list:
+ LOG.info(cmd)
+ self.connection.execute(cmd)
+ time.sleep(self.wait_for_vswitchd)
+
+ def setup_ovs_bridge_add_flows(self):
+ dpdk_args = ""
+ dpdk_list = []
+ vpath = self.ovs_properties.get("vpath", "/usr/local")
+ version = self.ovs_properties.get('version', {})
+ ovs_ver = [int(x) for x in version.get('ovs', self.DEFAULT_OVS).split('.')]
+ ovs_add_port = \
+ "ovs-vsctl add-port {br} {port} -- set Interface {port} type={type_}{dpdk_args}"
+ ovs_add_queue = "ovs-vsctl set Interface {port} options:n_rxq={queue}"
+ chmod_vpath = "chmod 0777 {0}/var/run/openvswitch/dpdkvhostuser*"
+
+ cmd_dpdk_list = [
+ "ovs-vsctl del-br br0",
+ "rm -rf /usr/local/var/run/openvswitch/dpdkvhostuser*",
+ "ovs-vsctl add-br br0 -- set bridge br0 datapath_type=netdev",
+ ]
+
+ ordered_network = OrderedDict(self.networks)
+ for index, (key, vnf) in enumerate(ordered_network.items()):
+ if ovs_ver >= [2, 7, 0]:
+ dpdk_args = " options:dpdk-devargs=%s" % vnf.get("phy_port")
+ dpdk_list.append(ovs_add_port.format(br='br0', port='dpdk%s' % vnf.get("port_num", 0),
+ type_='dpdk', dpdk_args=dpdk_args))
+ dpdk_list.append(ovs_add_queue.format(port='dpdk%s' % vnf.get("port_num", 0),
+ queue=self.ovs_properties.get("queues", 4)))
+
+ # Sorting the array to make sure we execute dpdk0... in the order
+ list.sort(dpdk_list)
+ cmd_dpdk_list.extend(dpdk_list)
+
+ # Need to do two for loop to maintain the dpdk/vhost ports.
+ for index, _ in enumerate(ordered_network):
+ cmd_dpdk_list.append(ovs_add_port.format(br='br0', port='dpdkvhostuser%s' % index,
+ type_='dpdkvhostuser', dpdk_args=""))
+
+ for cmd in cmd_dpdk_list:
+ LOG.info(cmd)
+ self.connection.execute(cmd)
+
+ # Fixme: add flows code
+ ovs_flow = "ovs-ofctl add-flow br0 in_port=%s,action=output:%s"
+
+ network_count = len(ordered_network) + 1
+ for in_port, out_port in zip(range(1, network_count),
+ range(network_count, network_count * 2)):
+ self.connection.execute(ovs_flow % (in_port, out_port))
+ self.connection.execute(ovs_flow % (out_port, in_port))
+
+ self.connection.execute(chmod_vpath.format(vpath))
+
+ def cleanup_ovs_dpdk_env(self):
+ self.connection.execute("ovs-vsctl del-br br0")
+ self.connection.execute("pkill -9 ovs")
+
+ def check_ovs_dpdk_env(self):
+ self.cleanup_ovs_dpdk_env()
+
+ version = self.ovs_properties.get("version", {})
+ ovs_ver = version.get("ovs", self.DEFAULT_OVS)
+ dpdk_ver = version.get("dpdk", "16.07.2").split('.')
+
+ supported_version = self.SUPPORTED_OVS_TO_DPDK_MAP.get(ovs_ver, None)
+ if supported_version is None or supported_version.split('.')[:2] != dpdk_ver[:2]:
+ raise Exception("Unsupported ovs '{}'. Please check the config...".format(ovs_ver))
+
+ status = self.connection.execute("ovs-vsctl -V | grep -i '%s'" % ovs_ver)[0]
+ if status:
+ deploy = OvsDeploy(self.connection,
+ get_nsb_option("bin_path"),
+ self.ovs_properties)
+ deploy.ovs_deploy()
+
+ def deploy(self):
+ """don't need to deploy"""
+
+ # Todo: NFVi deploy (sriov, vswitch, ovs etc) based on the config.
+ if not self.vm_deploy:
+ return
+
+ self.connection = ssh.SSH.from_node(self.host_mgmt)
+ self.dpdk_nic_bind = provision_tool(
+ self.connection,
+ os.path.join(get_nsb_option("bin_path"), "dpdk-devbind.py"))
+
+ # Check dpdk/ovs version, if not present install
+ self.check_ovs_dpdk_env()
+ # Todo: NFVi deploy (sriov, vswitch, ovs etc) based on the config.
+ StandaloneContextHelper.install_req_libs(self.connection)
+ self.networks = StandaloneContextHelper.get_nic_details(self.connection,
+ self.networks,
+ self.dpdk_nic_bind)
+
+ self.setup_ovs()
+ self.start_ovs_serverswitch()
+ self.setup_ovs_bridge_add_flows()
+ self.nodes = self.setup_ovs_dpdk_context()
+ LOG.debug("Waiting for VM to come up...")
+ self.nodes = StandaloneContextHelper.wait_for_vnfs_to_start(self.connection,
+ self.servers,
+ self.nodes)
+
+ def undeploy(self):
+
+ if not self.vm_deploy:
+ return
+
+ # Cleanup the ovs installation...
+ self.cleanup_ovs_dpdk_env()
+
+ # Bind nics back to kernel
+ bind_cmd = "{dpdk_nic_bind} --force -b {driver} {port}"
+ for key, port in self.networks.items():
+ vpci = port.get("phy_port")
+ phy_driver = port.get("driver")
+ self.connection.execute(bind_cmd.format(dpdk_nic_bind=self.dpdk_nic_bind,
+ driver=phy_driver, port=vpci))
+
+ # Todo: NFVi undeploy (sriov, vswitch, ovs etc) based on the config.
+ for vm in self.vm_names:
+ Libvirt.check_if_vm_exists_and_delete(vm, self.connection)
+
+ def _get_server(self, attr_name):
+ """lookup server info by name from context
+
+ Keyword arguments:
+ attr_name -- A name for a server listed in nodes config file
+ """
+ node_name, name = self.split_name(attr_name)
+ if name is None or self.name != name:
+ return None
+
+ matching_nodes = (n for n in self.nodes if n["name"] == node_name)
+ try:
+ # A clone is created in order to avoid affecting the
+ # original one.
+ node = dict(next(matching_nodes))
+ except StopIteration:
+ return None
+
+ try:
+ duplicate = next(matching_nodes)
+ except StopIteration:
+ pass
+ else:
+ raise ValueError("Duplicate nodes!!! Nodes: %s %s",
+ (node, duplicate))
+
+ node["name"] = attr_name
+ return node
+
+ def _get_network(self, attr_name):
+ if not isinstance(attr_name, collections.Mapping):
+ network = self.networks.get(attr_name)
+
+ else:
+ # Don't generalize too much Just support vld_id
+ vld_id = attr_name.get('vld_id', {})
+ # for standalone context networks are dicts
+ iter1 = (n for n in self.networks.values() if n.get('vld_id') == vld_id)
+ network = next(iter1, None)
+
+ if network is None:
+ return None
+
+ result = {
+ # name is required
+ "name": network["name"],
+ "vld_id": network.get("vld_id"),
+ "segmentation_id": network.get("segmentation_id"),
+ "network_type": network.get("network_type"),
+ "physical_network": network.get("physical_network"),
+ }
+ return result
+
+ def configure_nics_for_ovs_dpdk(self):
+ portlist = OrderedDict(self.networks)
+ for key, ports in portlist.items():
+ mac = StandaloneContextHelper.get_mac_address()
+ portlist[key].update({'mac': mac})
+ self.networks = portlist
+ LOG.info("Ports %s" % self.networks)
+
+ def _enable_interfaces(self, index, vfs, cfg):
+ vpath = self.ovs_properties.get("vpath", "/usr/local")
+ vf = self.networks[vfs[0]]
+ port_num = vf.get('port_num', 0)
+ vpci = PciAddress.parse_address(vf['vpci'].strip(), multi_line=True)
+ # Generate the vpci for the interfaces
+ slot = index + port_num + 10
+ vf['vpci'] = \
+ "{}:{}:{:02x}.{}".format(vpci.domain, vpci.bus, slot, vpci.function)
+ Libvirt.add_ovs_interface(vpath, port_num, vf['vpci'], vf['mac'], str(cfg))
+
+ def setup_ovs_dpdk_context(self):
+ nodes = []
+
+ self.configure_nics_for_ovs_dpdk()
+
+ for index, (key, vnf) in enumerate(OrderedDict(self.servers).items()):
+ cfg = '/tmp/vm_ovs_%d.xml' % index
+ vm_name = "vm_%d" % index
+
+ # 1. Check and delete VM if already exists
+ Libvirt.check_if_vm_exists_and_delete(vm_name, self.connection)
+
+ vcpu, mac = Libvirt.build_vm_xml(self.connection, self.vm_flavor, cfg, vm_name, index)
+ # 2: Cleanup already available VMs
+ for idx, (vkey, vfs) in enumerate(OrderedDict(vnf["network_ports"]).items()):
+ if vkey == "mgmt":
+ continue
+ self._enable_interfaces(index, vfs, cfg)
+
+ # copy xml to target...
+ self.connection.put(cfg, cfg)
+
+ # FIXME: launch through libvirt
+ LOG.info("virsh create ...")
+ Libvirt.virsh_create_vm(self.connection, cfg)
+
+ # 5: Tunning for better performace
+ Libvirt.pin_vcpu_for_perf(self.connection, vm_name, vcpu)
+ self.vm_names.append(vm_name)
+
+ # build vnf node details
+ nodes.append(self.vnf_node.generate_vnf_instance(self.vm_flavor,
+ self.networks,
+ self.host_mgmt.get('ip'),
+ key, vnf, mac))
+
+ return nodes