diff options
Diffstat (limited to 'yardstick/benchmark')
-rw-r--r-- | yardstick/benchmark/contexts/ovsdpdk.py | 369 | ||||
-rw-r--r-- | yardstick/benchmark/contexts/sriov.py | 431 | ||||
-rw-r--r-- | yardstick/benchmark/contexts/standalone.py | 90 |
3 files changed, 878 insertions, 12 deletions
diff --git a/yardstick/benchmark/contexts/ovsdpdk.py b/yardstick/benchmark/contexts/ovsdpdk.py new file mode 100644 index 000000000..86610305e --- /dev/null +++ b/yardstick/benchmark/contexts/ovsdpdk.py @@ -0,0 +1,369 @@ +# Copyright (c) 2016-2017 Intel Corporation +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +from __future__ import absolute_import +import os +import yaml +import time +import glob +import itertools +import logging +from yardstick import ssh +from yardstick.benchmark.contexts.standalone import StandaloneContext + +BIN_PATH = "/opt/isb_bin/" +DPDK_NIC_BIND = "dpdk_nic_bind.py" + +log = logging.getLogger(__name__) + +VM_TEMPLATE = """ +<domain type='kvm'> + <name>vm1</name> + <uuid>18230c0c-635d-4c50-b2dc-a213d30acb34</uuid> + <memory unit='KiB'>20971520</memory> + <currentMemory unit="KiB">20971520</currentMemory> + <memoryBacking> + <hugepages/> + </memoryBacking> + <vcpu placement='static'>20</vcpu> + <os> + <type arch='x86_64' machine='pc'>hvm</type> + <boot dev='hd'/> + </os> + <features> + <acpi/> + <apic/> + </features> + <cpu match="exact" mode='host-model'> + <model fallback='allow'/> + <topology sockets='1' cores='10' threads='2'/> + </cpu> + <on_poweroff>destroy</on_poweroff> + <on_reboot>restart</on_reboot> + <on_crash>destroy</on_crash> + <devices> + <emulator>/usr/bin/qemu-system-x86_64</emulator> + <disk type='file' device='disk'> + <driver name='qemu' type='qcow2' cache='none'/> + <source file="{vm_image}"/> + <target dev='vda' bus='virtio'/> + <address bus="0x00" domain="0x0000" + function="0x0" slot="0x04" type="pci" /> + </disk> + <!--disk type='dir' device='disk'> + <driver name='qemu' type='fat'/> + <source dir='/opt/isb_bin/dpdk'/> + <target dev='vdb' bus='virtio'/> + <readonly/> + </disk--> + <interface type="bridge"> + <mac address="00:00:00:ab:cd:ef" /> + <source bridge="br-int" /> + </interface> + <interface type='vhostuser'> + <mac address='00:00:00:00:00:01'/> + <source type='unix' path='/usr/local/var/run/openvswitch/dpdkvhostuser0' mode='client'/> + <model type='virtio'/> + <driver queues='4'> + <host mrg_rxbuf='off'/> + </driver> + </interface> + <interface type='vhostuser'> + <mac address='00:00:00:00:00:02'/> + <source type='unix' path='/usr/local/var/run/openvswitch/dpdkvhostuser1' mode='client'/> + <model type='virtio'/> + <driver queues='4'> + <host mrg_rxbuf='off'/> + </driver> + </interface> + <serial type='pty'> + <target port='0'/> + </serial> + <console type='pty'> + <target type='serial' port='0'/> + </console> + <graphics autoport="yes" listen="0.0.0.0" port="1" type="vnc" /> + </devices> +</domain> +""" + + +class Ovsdpdk(StandaloneContext): + def __init__(self): + self.name = None + self.file_path = None + self.nodes = [] + self.vm_deploy = False + self.ovs = [] + self.first_run = True + self.dpdk_nic_bind = BIN_PATH + DPDK_NIC_BIND + self.user = "" + self.ssh_ip = "" + self.passwd = "" + self.ssh_port = "" + self.auth_type = "" + + def init(self): + '''initializes itself''' + log.debug("In init") + self.parse_pod_and_get_data() + + def parse_pod_and_get_data(self, file_path): + self.file_path = file_path + print("parsing pod file: {0}".format(self.file_path)) + try: + with open(self.file_path) as stream: + cfg = yaml.load(stream) + except IOError: + print("File {0} does not exist".format(self.file_path)) + raise + + self.ovs.extend([node for node in cfg["nodes"] + if node["role"] == "Ovsdpdk"]) + self.user = self.ovs[0]['user'] + self.ssh_ip = self.ovs[0]['ip'] + if self.ovs[0]['auth_type'] == "password": + self.passwd = self.ovs[0]['password'] + else: + self.ssh_port = self.ovs[0]['ssh_port'] + self.key_filename = self.ovs[0]['key_filename'] + + def ssh_remote_machine(self): + if self.ovs[0]['auth_type'] == "password": + self.connection = ssh.SSH( + self.user, + self.ssh_ip, + password=self.passwd) + self.connection.wait() + else: + if self.ssh_port is not None: + ssh_port = self.ssh_port + else: + ssh_port = ssh.DEFAULT_PORT + self.connection = ssh.SSH( + self.user, + self.ssh_ip, + port=ssh_port, + key_filename=self.key_filename) + self.connection.wait() + + def get_nic_details(self): + nic_details = {} + nic_details['interface'] = {} + nic_details['pci'] = self.ovs[0]['phy_ports'] + nic_details['phy_driver'] = self.ovs[0]['phy_driver'] + nic_details['vports_mac'] = self.ovs[0]['vports_mac'] + # Make sure that ports are bound to kernel drivers e.g. i40e/ixgbe + for i, _ in enumerate(nic_details['pci']): + err, out, _ = self.connection.execute( + "{dpdk_nic_bind} --force -b {driver} {port}".format( + dpdk_nic_bind=self.dpdk_nic_bind, + driver=self.ovs[0]['phy_driver'], + port=self.ovs[0]['phy_ports'][i])) + err, out, _ = self.connection.execute( + "lshw -c network -businfo | grep '{port}'".format( + port=self.ovs[0]['phy_ports'][i])) + a = out.split()[1] + err, out, _ = self.connection.execute( + "ip -s link show {interface}".format( + interface=out.split()[1])) + nic_details['interface'][i] = str(a) + print("{0}".format(nic_details)) + return nic_details + + def install_req_libs(self): + if self.first_run: + err, out, _ = self.connection.execute("apt-get update") + print("{0}".format(out)) + err, out, _ = self.connection.execute( + "apt-get -y install qemu-kvm libvirt-bin") + print("{0}".format(out)) + err, out, _ = self.connection.execute( + "apt-get -y install libvirt-dev bridge-utils numactl") + print("{0}".format(out)) + self.first_run = False + + def setup_ovs(self, vpcis): + self.connection.execute("/usr/bin/chmod 0666 /dev/vfio/*") + self.connection.execute("/usr/bin/chmod a+x /dev/vfio") + self.connection.execute("pkill -9 ovs") + self.connection.execute("ps -ef | grep ovs | grep -v grep | " + "awk '{print $2}' | xargs -r kill -9") + self.connection.execute("killall -r 'ovs*'") + self.connection.execute( + "mkdir -p {0}/etc/openvswitch".format(self.ovs[0]["vpath"])) + self.connection.execute( + "mkdir -p {0}/var/run/openvswitch".format(self.ovs[0]["vpath"])) + self.connection.execute( + "rm {0}/etc/openvswitch/conf.db".format(self.ovs[0]["vpath"])) + self.connection.execute( + "ovsdb-tool create {0}/etc/openvswitch/conf.db " + "{0}/share/openvswitch/" + "vswitch.ovsschema".format(self.ovs[0]["vpath"])) + self.connection.execute("modprobe vfio-pci") + self.connection.execute("chmod a+x /dev/vfio") + self.connection.execute("chmod 0666 /dev/vfio/*") + for vpci in vpcis: + self.connection.execute( + "/opt/isb_bin/dpdk_nic_bind.py " + "--bind=vfio-pci {0}".format(vpci)) + + def start_ovs_serverswitch(self): + self.connection.execute("mkdir -p /usr/local/var/run/openvswitch") + self.connection.execute( + "ovsdb-server --remote=punix:" + "/usr/local/var/run/openvswitch/db.sock --pidfile --detach") + self.connection.execute( + "ovs-vsctl --no-wait set " + "Open_vSwitch . other_config:dpdk-init=true") + self.connection.execute( + "ovs-vsctl --no-wait set " + "Open_vSwitch . other_config:dpdk-lcore-mask=0x3") + self.connection.execute( + "ovs-vsctl --no-wait set " + "Open_vSwitch . other_config:dpdk-socket-mem='2048,0'") + self.connection.execute( + "ovs-vswitchd unix:{0}/" + "var/run/openvswitch/db.sock --pidfile --detach " + "--log-file=/var/log/openvswitch/" + "ovs-vswitchd.log".format( + self.ovs[0]["vpath"])) + self.connection.execute( + "ovs-vsctl set Open_vSwitch . other_config:pmd-cpu-mask=2C") + + def setup_ovs_bridge(self): + self.connection.execute("ovs-vsctl del-br br0") + self.connection.execute( + "rm -rf /usr/local/var/run/openvswitch/dpdkvhostuser*") + self.connection.execute( + "ovs-vsctl add-br br0 -- set bridge br0 datapath_type=netdev") + self.connection.execute( + "ovs-vsctl add-port br0 dpdk0 -- set Interface dpdk0 type=dpdk") + self.connection.execute( + "ovs-vsctl add-port br0 dpdk1 -- set Interface dpdk1 type=dpdk") + self.connection.execute( + "ovs-vsctl add-port br0 dpdkvhostuser0 -- set Interface " + "dpdkvhostuser0 type=dpdkvhostuser") + self.connection.execute("ovs-vsctl add-port br0 dpdkvhostuser1 " + "-- set Interface dpdkvhostuser1 " + "type=dpdkvhostuser") + self.connection.execute( + "chmod 0777 {0}/var/run/" + "openvswitch/dpdkvhostuser*".format(self.ovs[0]["vpath"])) + + def add_oflows(self): + self.connection.execute("ovs-ofctl del-flows br0") + for flow in self.ovs[0]["flow"]: + self.connection.execute(flow) + self.connection.execute("ovs-ofctl dump-flows br0") + self.connection.execute( + "ovs-vsctl set Interface dpdk0 options:n_rxq=4") + self.connection.execute( + "ovs-vsctl set Interface dpdk1 options:n_rxq=4") + + def setup_ovs_context(self, pcis, nic_details, host_driver): + + ''' 1: Setup vm_ovs.xml to launch VM.''' + cfg_ovs = '/tmp/vm_ovs.xml' + vm_ovs_xml = VM_TEMPLATE.format(vm_image=self.ovs[0]["images"]) + with open(cfg_ovs, 'w') as f: + f.write(vm_ovs_xml) + + ''' 2: Create and start the VM''' + self.connection.put(cfg_ovs, cfg_ovs) + time.sleep(10) + err, out = self.check_output("virsh list --name | grep -i vm1") + if out == "vm1": + print("VM is already present") + else: + ''' FIXME: launch through libvirt''' + print("virsh create ...") + err, out, _ = self.connection.execute( + "virsh create /tmp/vm_ovs.xml") + time.sleep(10) + print("err : {0}".format(err)) + print("{0}".format(_)) + print("out : {0}".format(out)) + + ''' 3: Tuning for better performace.''' + self.pin_vcpu(pcis) + self.connection.execute( + "echo 1 > /sys/module/kvm/parameters/" + "allow_unsafe_assigned_interrupts") + self.connection.execute( + "echo never > /sys/kernel/mm/transparent_hugepage/enabled") + print("After tuning performance ...") + + ''' This is roughly compatible with check_output function in subprocess + module which is only available in python 2.7.''' + def check_output(self, cmd, stderr=None): + '''Run a command and capture its output''' + err, out, _ = self.connection.execute(cmd) + return err, out + + def read_from_file(self, filename): + data = "" + with open(filename, 'r') as the_file: + data = the_file.read() + return data + + def write_to_file(self, filename, content): + with open(filename, 'w') as the_file: + the_file.write(content) + + def pin_vcpu(self, pcis): + nodes = self.get_numa_nodes() + print("{0}".format(nodes)) + num_nodes = len(nodes) + for i in range(0, 10): + self.connection.execute( + "virsh vcpupin vm1 {0} {1}".format( + i, nodes[str(num_nodes - 1)][i])) + + def get_numa_nodes(self): + nodes_sysfs = glob.iglob("/sys/devices/system/node/node*") + nodes = {} + for node_sysfs in nodes_sysfs: + num = os.path.basename(node_sysfs).replace("node", "") + with open(os.path.join(node_sysfs, "cpulist")) as cpulist_file: + cpulist = cpulist_file.read().strip() + print("cpulist: {0}".format(cpulist)) + nodes[num] = self.split_cpu_list(cpulist) + print("nodes: {0}".format(nodes)) + return nodes + + def split_cpu_list(self, cpu_list): + if cpu_list: + ranges = cpu_list.split(',') + bounds = ([int(b) for b in r.split('-')] for r in ranges) + range_objects =\ + (range(bound[0], bound[1] + 1 if len(bound) == 2 + else bound[0] + 1) for bound in bounds) + + return sorted(itertools.chain.from_iterable(range_objects)) + else: + return [] + + def destroy_vm(self): + host_driver = self.ovs[0]['phy_driver'] + err, out = self.check_output("virsh list --name | grep -i vm1") + print("{0}".format(out)) + if err == 0: + self.connection.execute("virsh shutdown vm1") + self.connection.execute("virsh destroy vm1") + self.check_output("rmmod {0}".format(host_driver))[1].splitlines() + self.check_output("modprobe {0}".format(host_driver))[ + 1].splitlines() + else: + print("error : ", err) diff --git a/yardstick/benchmark/contexts/sriov.py b/yardstick/benchmark/contexts/sriov.py new file mode 100644 index 000000000..da143cc4b --- /dev/null +++ b/yardstick/benchmark/contexts/sriov.py @@ -0,0 +1,431 @@ +# Copyright (c) 2016-2017 Intel Corporation +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +from __future__ import absolute_import +import os +import yaml +import re +import time +import glob +import uuid +import random +import logging +import itertools +import xml.etree.ElementTree as ET +from yardstick import ssh +from yardstick.network_services.utils import get_nsb_option +from yardstick.network_services.utils import provision_tool +from yardstick.benchmark.contexts.standalone import StandaloneContext + +log = logging.getLogger(__name__) + +VM_TEMPLATE = """ +<domain type="kvm"> + <name>vm1</name> + <uuid>{random_uuid}</uuid> + <memory unit="KiB">102400</memory> + <currentMemory unit="KiB">102400</currentMemory> + <memoryBacking> + <hugepages /> + </memoryBacking> + <vcpu placement="static">20</vcpu> + <os> + <type arch="x86_64" machine="pc-i440fx-utopic">hvm</type> + <boot dev="hd" /> + </os> + <features> + <acpi /> + <apic /> + <pae /> + </features> + <cpu match="exact" mode="custom"> + <model fallback="allow">SandyBridge</model> + <topology cores="10" sockets="1" threads="2" /> + </cpu> + <clock offset="utc"> + <timer name="rtc" tickpolicy="catchup" /> + <timer name="pit" tickpolicy="delay" /> + <timer name="hpet" present="no" /> + </clock> + <on_poweroff>destroy</on_poweroff> + <on_reboot>restart</on_reboot> + <on_crash>restart</on_crash> + <devices> + <emulator>/usr/bin/kvm-spice</emulator> + <disk device="disk" type="file"> + <driver name="qemu" type="qcow2" /> + <source file="{vm_image}"/> + <target bus="virtio" dev="vda" /> + <address bus="0x00" domain="0x0000" +function="0x0" slot="0x04" type="pci" /> + </disk> + <controller index="0" model="ich9-ehci1" type="usb"> + <address bus="0x00" domain="0x0000" +function="0x7" slot="0x05" type="pci" /> + </controller> + <controller index="0" model="ich9-uhci1" type="usb"> + <master startport="0" /> + <address bus="0x00" domain="0x0000" function="0x0" +multifunction="on" slot="0x05" type="pci" /> + </controller> + <controller index="0" model="ich9-uhci2" type="usb"> + <master startport="2" /> + <address bus="0x00" domain="0x0000" +function="0x1" slot="0x05" type="pci" /> + </controller> + <controller index="0" model="ich9-uhci3" type="usb"> + <master startport="4" /> + <address bus="0x00" domain="0x0000" +function="0x2" slot="0x05" type="pci" /> + </controller> + <controller index="0" model="pci-root" type="pci" /> + <serial type="pty"> + <target port="0" /> + </serial> + <console type="pty"> + <target port="0" type="serial" /> + </console> + <input bus="usb" type="tablet" /> + <input bus="ps2" type="mouse" /> + <input bus="ps2" type="keyboard" /> + <graphics autoport="yes" listen="0.0.0.0" port="-1" type="vnc" /> + <video> + <model heads="1" type="cirrus" vram="16384" /> + <address bus="0x00" domain="0x0000" +function="0x0" slot="0x02" type="pci" /> + </video> + <memballoon model="virtio"> + <address bus="0x00" domain="0x0000" +function="0x0" slot="0x06" type="pci" /> + </memballoon> + <interface type="bridge"> + <mac address="{mac_addr}" /> + <source bridge="virbr0" /> + </interface> + </devices> +</domain> +""" + + +class Sriov(StandaloneContext): + def __init__(self): + self.name = None + self.file_path = None + self.nodes = [] + self.vm_deploy = False + self.sriov = [] + self.first_run = True + self.dpdk_nic_bind = "" + self.user = "" + self.ssh_ip = "" + self.passwd = "" + self.ssh_port = "" + self.auth_type = "" + + def init(self): + log.debug("In init") + self.parse_pod_and_get_data(self.file_path) + + def parse_pod_and_get_data(self, file_path): + self.file_path = file_path + log.debug("parsing pod file: {0}".format(self.file_path)) + try: + with open(self.file_path) as stream: + cfg = yaml.load(stream) + except IOError: + log.error("File {0} does not exist".format(self.file_path)) + raise + + self.sriov.extend([node for node in cfg["nodes"] + if node["role"] == "Sriov"]) + self.user = self.sriov[0]['user'] + self.ssh_ip = self.sriov[0]['ip'] + if self.sriov[0]['auth_type'] == "password": + self.passwd = self.sriov[0]['password'] + else: + self.ssh_port = self.sriov[0]['ssh_port'] + self.key_filename = self.sriov[0]['key_filename'] + + def ssh_remote_machine(self): + if self.sriov[0]['auth_type'] == "password": + self.connection = ssh.SSH( + self.user, + self.ssh_ip, + password=self.passwd) + self.connection.wait() + else: + if self.ssh_port is not None: + ssh_port = self.ssh_port + else: + ssh_port = ssh.DEFAULT_PORT + self.connection = ssh.SSH( + self.user, + self.ssh_ip, + port=ssh_port, + key_filename=self.key_filename) + self.connection.wait() + self.dpdk_nic_bind = provision_tool( + self.connection, + os.path.join(get_nsb_option("bin_path"), "dpdk_nic_bind.py")) + + def get_nic_details(self): + nic_details = {} + nic_details = { + 'interface': {}, + 'pci': self.sriov[0]['phy_ports'], + 'phy_driver': self.sriov[0]['phy_driver'], + 'vf_macs': self.sriov[0]['vf_macs'] + } + # Make sure that ports are bound to kernel drivers e.g. i40e/ixgbe + for i, _ in enumerate(nic_details['pci']): + err, out, _ = self.connection.execute( + "{dpdk_nic_bind} --force -b {driver} {port}".format( + dpdk_nic_bind=self.dpdk_nic_bind, + driver=self.sriov[0]['phy_driver'], + port=self.sriov[0]['phy_ports'][i])) + err, out, _ = self.connection.execute( + "lshw -c network -businfo | grep '{port}'".format( + port=self.sriov[0]['phy_ports'][i])) + a = out.split()[1] + err, out, _ = self.connection.execute( + "ip -s link show {interface}".format( + interface=out.split()[1])) + nic_details['interface'][i] = str(a) + log.info("{0}".format(nic_details)) + return nic_details + + def install_req_libs(self): + if self.first_run: + log.info("Installing required libraries...") + err, out, _ = self.connection.execute("apt-get update") + log.debug("{0}".format(out)) + err, out, _ = self.connection.execute( + "apt-get -y install qemu-kvm libvirt-bin") + log.debug("{0}".format(out)) + err, out, _ = self.connection.execute( + "apt-get -y install libvirt-dev bridge-utils numactl") + log.debug("{0}".format(out)) + self.first_run = False + + def configure_nics_for_sriov(self, host_driver, nic_details): + vf_pci = [[], []] + self.connection.execute( + "rmmod {0}".format(host_driver))[1].splitlines() + self.connection.execute( + "modprobe {0} num_vfs=1".format(host_driver))[1].splitlines() + nic_details['vf_pci'] = {} + for i in range(len(nic_details['pci'])): + self.connection.execute( + "echo 1 > /sys/bus/pci/devices/{0}/sriov_numvfs".format( + nic_details['pci'][i])) + err, out, _ = self.connection.execute( + "ip link set {interface} vf 0 mac {mac}".format( + interface=nic_details['interface'][i], + mac=nic_details['vf_macs'][i])) + time.sleep(3) + vf_pci[i] = self.get_vf_datas( + 'vf_pci', + nic_details['pci'][i], + nic_details['vf_macs'][i]) + nic_details['vf_pci'][i] = vf_pci[i] + log.debug("NIC DETAILS : {0}".format(nic_details)) + return nic_details + + def setup_sriov_context(self, pcis, nic_details, host_driver): + blacklist = "/etc/modprobe.d/blacklist.conf" + + # 1 : Blacklist the vf driver in /etc/modprobe.d/blacklist.conf + vfnic = "{0}vf".format(host_driver) + lines = self.read_from_file(blacklist) + if vfnic not in lines: + vfblacklist = "blacklist {vfnic}".format(vfnic=vfnic) + self.connection.execute( + "echo {vfblacklist} >> {blacklist}".format( + vfblacklist=vfblacklist, + blacklist=blacklist)) + + # 2 : modprobe host_driver with num_vfs + nic_details = self.configure_nics_for_sriov(host_driver, nic_details) + + # 3: Setup vm_sriov.xml to launch VM + cfg_sriov = '/tmp/vm_sriov.xml' + mac = [0x00, 0x24, 0x81, + random.randint(0x00, 0x7f), + random.randint(0x00, 0xff), + random.randint(0x00, 0xff)] + mac_address = ':'.join(map(lambda x: "%02x" % x, mac)) + vm_sriov_xml = VM_TEMPLATE.format( + random_uuid=uuid.uuid4(), + mac_addr=mac_address, + vm_image=self.sriov[0]["images"]) + with open(cfg_sriov, 'w') as f: + f.write(vm_sriov_xml) + + vf = nic_details['vf_pci'] + for index in range(len(nic_details['vf_pci'])): + self.add_sriov_interface( + index, + vf[index]['vf_pci'], + mac_address, + "/tmp/vm_sriov.xml") + self.connection.execute( + "ifconfig {interface} up".format( + interface=nic_details['interface'][index])) + + # 4: Create and start the VM + self.connection.put(cfg_sriov, cfg_sriov) + time.sleep(10) + err, out = self.check_output("virsh list --name | grep -i vm1") + try: + if out == "vm1": + log.info("VM is already present") + else: + # FIXME: launch through libvirt + log.info("virsh create ...") + err, out, _ = self.connection.execute( + "virsh create /tmp/vm_sriov.xml") + time.sleep(10) + log.error("err : {0}".format(err)) + log.error("{0}".format(_)) + log.debug("out : {0}".format(out)) + except ValueError: + raise + + # 5: Tunning for better performace + self.pin_vcpu(pcis) + self.connection.execute( + "echo 1 > /sys/module/kvm/parameters/" + "allow_unsafe_assigned_interrupts") + self.connection.execute( + "echo never > /sys/kernel/mm/transparent_hugepage/enabled") + + def add_sriov_interface(self, index, vf_pci, vfmac, xml): + root = ET.parse(xml) + pattern = "0000:(\d+):(\d+).(\d+)" + m = re.search(pattern, vf_pci, re.MULTILINE) + device = root.find('devices') + + interface = ET.SubElement(device, 'interface') + interface.set('managed', 'yes') + interface.set('type', 'hostdev') + + mac = ET.SubElement(interface, 'mac') + mac.set('address', vfmac) + source = ET.SubElement(interface, 'source') + + addr = ET.SubElement(source, "address") + addr.set('domain', "0x0") + addr.set('bus', "{0}".format(m.group(1))) + addr.set('function', "{0}".format(m.group(3))) + addr.set('slot', "{0}".format(m.group(2))) + addr.set('type', "pci") + + vf_pci = ET.SubElement(interface, 'address') + vf_pci.set('type', 'pci') + vf_pci.set('domain', '0x0000') + vf_pci.set('bus', '0x00') + vf_pci.set('slot', '0x0{0}'.format(index + 7)) + vf_pci.set('function', '0x00') + + root.write(xml) + + # This is roughly compatible with check_output function in subprocess + # module which is only available in python 2.7 + def check_output(self, cmd, stderr=None): + # Run a command and capture its output + err, out, _ = self.connection.execute(cmd) + return err, out + + def get_virtual_devices(self, pci): + pf_vfs = {} + err, extra_info = self.check_output( + "cat /sys/bus/pci/devices/{0}/virtfn0/uevent".format(pci)) + pattern = "PCI_SLOT_NAME=(?P<name>[0-9:.\s.]+)" + m = re.search(pattern, extra_info, re.MULTILINE) + + if m: + pf_vfs.update({pci: str(m.group(1).rstrip())}) + log.info("pf_vfs : {0}".format(pf_vfs)) + return pf_vfs + + def get_vf_datas(self, key, value, vfmac): + vfret = {} + pattern = "0000:(\d+):(\d+).(\d+)" + + vfret["mac"] = vfmac + vfs = self.get_virtual_devices(value) + log.info("vfs: {0}".format(vfs)) + for k, v in vfs.items(): + m = re.search(pattern, k, re.MULTILINE) + m1 = re.search(pattern, value, re.MULTILINE) + if m.group(1) == m1.group(1): + vfret["vf_pci"] = str(v) + break + + return vfret + + def read_from_file(self, filename): + data = "" + with open(filename, 'r') as the_file: + data = the_file.read() + return data + + def write_to_file(self, filename, content): + with open(filename, 'w') as the_file: + the_file.write(content) + + def pin_vcpu(self, pcis): + nodes = self.get_numa_nodes() + log.info("{0}".format(nodes)) + num_nodes = len(nodes) + for i in range(0, 10): + self.connection.execute( + "virsh vcpupin vm1 {0} {1}".format( + i, nodes[str(num_nodes - 1)][i])) + + def get_numa_nodes(self): + nodes_sysfs = glob.iglob("/sys/devices/system/node/node*") + nodes = {} + for node_sysfs in nodes_sysfs: + num = os.path.basename(node_sysfs).replace("node", "") + with open(os.path.join(node_sysfs, "cpulist")) as cpulist_file: + cpulist = cpulist_file.read().strip() + nodes[num] = self.split_cpu_list(cpulist) + log.info("nodes: {0}".format(nodes)) + return nodes + + def split_cpu_list(self, cpu_list): + if cpu_list: + ranges = cpu_list.split(',') + bounds = ([int(b) for b in r.split('-')] for r in ranges) + range_objects =\ + (range(bound[0], bound[1] + 1 if len(bound) == 2 + else bound[0] + 1) for bound in bounds) + + return sorted(itertools.chain.from_iterable(range_objects)) + else: + return [] + + def destroy_vm(self): + host_driver = self.sriov[0]["phy_driver"] + err, out = self.check_output("virsh list --name | grep -i vm1") + log.info("{0}".format(out)) + if err == 0: + self.connection.execute("virsh shutdown vm1") + self.connection.execute("virsh destroy vm1") + self.check_output("rmmod {0}".format(host_driver))[1].splitlines() + self.check_output("modprobe {0}".format(host_driver))[ + 1].splitlines() + else: + log.error("error : {0}".format(err)) diff --git a/yardstick/benchmark/contexts/standalone.py b/yardstick/benchmark/contexts/standalone.py index 8614f0cac..2bc1f3755 100644 --- a/yardstick/benchmark/contexts/standalone.py +++ b/yardstick/benchmark/contexts/standalone.py @@ -18,9 +18,11 @@ import logging import errno import collections import yaml +import time from yardstick.benchmark.contexts.base import Context from yardstick.common.constants import YARDSTICK_ROOT_PATH +from yardstick.common.utils import import_modules_from_package, itersubclasses LOG = logging.getLogger(__name__) @@ -38,7 +40,8 @@ class StandaloneContext(Context): self.nodes = [] self.networks = {} self.nfvi_node = [] - super(StandaloneContext, self).__init__() + self.nfvi_obj = None + super(self.__class__, self).__init__() def read_config_file(self): """Read from config file""" @@ -48,6 +51,14 @@ class StandaloneContext(Context): cfg = yaml.load(stream) return cfg + def get_nfvi_obj(self): + print("{0}".format(self.nfvi_node[0]['role'])) + context_type = self.get_context_impl(self.nfvi_node[0]['role']) + nfvi_obj = context_type() + nfvi_obj.__init__() + nfvi_obj.parse_pod_and_get_data(self.file_path) + return nfvi_obj + def init(self, attrs): """initializes itself from the supplied arguments""" @@ -64,11 +75,24 @@ class StandaloneContext(Context): else: raise - self.nodes.extend(cfg["nodes"]) - self.nfvi_node.extend([node for node in cfg["nodes"] - if node["role"] == "nfvi_node"]) + self.vm_deploy = attrs.get("vm_deploy", True) + self.nodes.extend([node for node in cfg["nodes"] + if str(node["role"]) != "Sriov" and + str(node["role"]) != "Ovsdpdk"]) + for node in cfg["nodes"]: + if str(node["role"]) == "Sriov": + self.nfvi_node.extend([node for node in cfg["nodes"] + if str(node["role"]) == "Sriov"]) + if str(node["role"]) == "Ovsdpdk": + self.nfvi_node.extend([node for node in cfg["nodes"] + if str(node["role"]) == "Ovsdpdk"]) + LOG.info("{0}".format(node["role"])) + else: + LOG.debug("Node role is other than SRIOV and OVS") + self.nfvi_obj = self.get_nfvi_obj() # add optional static network definition self.networks.update(cfg.get("networks", {})) + self.nfvi_obj = self.get_nfvi_obj() LOG.debug("Nodes: %r", self.nodes) LOG.debug("NFVi Node: %r", self.nfvi_node) LOG.debug("Networks: %r", self.networks) @@ -77,13 +101,44 @@ class StandaloneContext(Context): """don't need to deploy""" # Todo: NFVi deploy (sriov, vswitch, ovs etc) based on the config. - pass + if not self.vm_deploy: + return + + # Todo: NFVi deploy (sriov, vswitch, ovs etc) based on the config. + self.nfvi_obj.ssh_remote_machine() + if self.nfvi_obj.first_run is True: + self.nfvi_obj.install_req_libs() + + nic_details = self.nfvi_obj.get_nic_details() + print("{0}".format(nic_details)) + + if self.nfvi_node[0]["role"] == "Sriov": + self.nfvi_obj.setup_sriov_context( + self.nfvi_obj.sriov[0]['phy_ports'], + nic_details, + self.nfvi_obj.sriov[0]['phy_driver']) + if self.nfvi_node[0]["role"] == "Ovsdpdk": + self.nfvi_obj.setup_ovs(self.nfvi_obj.ovs[0]["phy_ports"]) + self.nfvi_obj.start_ovs_serverswitch() + time.sleep(5) + self.nfvi_obj.setup_ovs_bridge() + self.nfvi_obj.add_oflows() + self.nfvi_obj.setup_ovs_context( + self.nfvi_obj.ovs[0]['phy_ports'], + nic_details, + self.nfvi_obj.ovs[0]['phy_driver']) + pass def undeploy(self): """don't need to undeploy""" + if not self.vm_deploy: + return # Todo: NFVi undeploy (sriov, vswitch, ovs etc) based on the config. - super(StandaloneContext, self).undeploy() + # self.nfvi_obj = self.get_nfvi_obj() + self.nfvi_obj.ssh_remote_machine() + self.nfvi_obj.destroy_vm() + pass def _get_server(self, attr_name): """lookup server info by name from context @@ -91,16 +146,12 @@ class StandaloneContext(Context): Keyword arguments: attr_name -- A name for a server listed in nodes config file """ - if isinstance(attr_name, collections.Mapping): return None - - if self.name.split("-")[0] != attr_name.split(".")[1]: + if self.name != attr_name.split(".")[1]: return None - node_name = attr_name.split(".")[0] matching_nodes = (n for n in self.nodes if n["name"] == node_name) - try: # A clone is created in order to avoid affecting the # original one. @@ -115,7 +166,6 @@ class StandaloneContext(Context): else: raise ValueError("Duplicate nodes!!! Nodes: %s %s", (matching_nodes, duplicate)) - node["name"] = attr_name return node @@ -146,3 +196,19 @@ class StandaloneContext(Context): "physical_network": network.get("physical_network"), } return result + + def get_context_impl(self, nfvi_type): + """ Find the implementing class from vnf_model["vnf"]["name"] field + + :param vnf_model: dictionary containing a parsed vnfd + :return: subclass of GenericVNF + """ + import_modules_from_package( + "yardstick.benchmark.contexts") + expected_name = nfvi_type + impl = [c for c in itersubclasses(StandaloneContext) + if c.__name__ == expected_name] + try: + return next(iter(impl)) + except StopIteration: + raise ValueError("No implementation for %s", expected_name) |