X-Git-Url: https://gerrit.fd.io/r/gitweb?p=csit.git;a=blobdiff_plain;f=resources%2Flibraries%2Fpython%2FQemuUtils.py;h=35959cef260242769f9938346c85de6212dd9c8a;hp=41b5f54684ed02c3afd09c3bcaabcd7ad5c39a36;hb=22bfd741614145d86751df2d61ff77af9db63627;hpb=d7c297494fa96603a90dda7ec50e73c041ca6aa2 diff --git a/resources/libraries/python/QemuUtils.py b/resources/libraries/python/QemuUtils.py index 41b5f54684..35959cef26 100644 --- a/resources/libraries/python/QemuUtils.py +++ b/resources/libraries/python/QemuUtils.py @@ -1,4 +1,4 @@ -# Copyright (c) 2019 Cisco and/or its affiliates. +# Copyright (c) 2022 Cisco and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: @@ -13,35 +13,36 @@ """QEMU utilities library.""" -# Disable due to pylint bug -# pylint: disable=no-name-in-module,import-error -from distutils.version import StrictVersion import json + from re import match from string import Template from time import sleep from robot.api import logger + from resources.libraries.python.Constants import Constants from resources.libraries.python.DpdkUtil import DpdkUtil from resources.libraries.python.DUTSetup import DUTSetup from resources.libraries.python.OptionString import OptionString -from resources.libraries.python.VppConfigGenerator import VppConfigGenerator -from resources.libraries.python.VPPUtil import VPPUtil from resources.libraries.python.ssh import exec_cmd, exec_cmd_no_error from resources.libraries.python.topology import NodeType, Topology +from resources.libraries.python.VhostUser import VirtioFeaturesFlags +from resources.libraries.python.VhostUser import VirtioFeatureMask +from resources.libraries.python.VppConfigGenerator import VppConfigGenerator -__all__ = ["QemuUtils"] +__all__ = [u"QemuUtils"] -class QemuUtils(object): +class QemuUtils: """QEMU utilities.""" # Use one instance of class per tests. - ROBOT_LIBRARY_SCOPE = 'TEST CASE' + ROBOT_LIBRARY_SCOPE = u"TEST CASE" - def __init__(self, node, qemu_id=1, smp=1, mem=512, vnf=None, - img=Constants.QEMU_VM_IMAGE): + def __init__( + self, node, qemu_id=1, smp=1, mem=512, vnf=None, + img=Constants.QEMU_VM_IMAGE, page_size=u""): """Initialize QemuUtil class. :param node: Node to run QEMU on. @@ -50,154 +51,256 @@ class QemuUtils(object): :param mem: Amount of memory. :param vnf: Network function workload. :param img: QEMU disk image or kernel image path. + :param page_size: Hugepage Size. :type node: dict :type qemu_id: int :type smp: int :type mem: int :type vnf: str :type img: str + :type page_size: str """ - self._vhost_id = 0 + self._nic_id = 0 self._node = node self._arch = Topology.get_node_arch(self._node) - dpdk_target = 'arm64-armv8a' if self._arch == 'aarch64' \ - else 'x86_64-native' - self._testpmd_path = '{path}/{dpdk_target}-linuxapp-gcc/app'\ - .format(path=Constants.QEMU_VM_DPDK, dpdk_target=dpdk_target) + self._opt = dict() + + # Architecture specific options + if self._arch == u"aarch64": + self._opt[u"machine_args"] = \ + u"virt,accel=kvm,usb=off,mem-merge=off,gic-version=3" + self._opt[u"console"] = u"ttyAMA0" + else: + self._opt[u"machine_args"] = u"pc,accel=kvm,usb=off,mem-merge=off" + self._opt[u"console"] = u"ttyS0" + self._testpmd_path = f"{Constants.QEMU_VM_DPDK}/build/app" self._vm_info = { - 'host': node['host'], - 'type': NodeType.VM, - 'port': 10021 + qemu_id, - 'serial': 4555 + qemu_id, - 'username': 'cisco', - 'password': 'cisco', - 'interfaces': {}, + u"host": node[u"host"], + u"type": NodeType.VM, + u"port": 10021 + qemu_id, + u"serial": 4555 + qemu_id, + u"username": 'testuser', + u"password": 'Csit1234', + u"interfaces": {}, } - if node['port'] != 22: - self._vm_info['host_port'] = node['port'] - self._vm_info['host_username'] = node['username'] - self._vm_info['host_password'] = node['password'] + if node[u"port"] != 22: + self._vm_info[u"host_port"] = node[u"port"] + self._vm_info[u"host_username"] = node[u"username"] + self._vm_info[u"host_password"] = node[u"password"] # Input Options. - self._opt = dict() - self._opt['qemu_id'] = qemu_id - self._opt['mem'] = int(mem) - self._opt['smp'] = int(smp) - self._opt['img'] = img - self._opt['vnf'] = vnf + self._opt[u"qemu_id"] = qemu_id + self._opt[u"mem"] = int(mem) + self._opt[u"smp"] = int(smp) + self._opt[u"img"] = img + self._opt[u"vnf"] = vnf + self._opt[u"page_size"] = page_size + # Temporary files. self._temp = dict() - self._temp['pidfile'] = '/var/run/qemu_{id}.pid'.format(id=qemu_id) + self._temp[u"log"] = f"/tmp/serial_{qemu_id}.log" + self._temp[u"pidfile"] = f"/run/qemu_{qemu_id}.pid" if img == Constants.QEMU_VM_IMAGE: - self._opt['vm_type'] = 'nestedvm' - self._temp['qmp'] = '/var/run/qmp_{id}.sock'.format(id=qemu_id) - self._temp['qga'] = '/var/run/qga_{id}.sock'.format(id=qemu_id) + self._temp[u"qmp"] = f"/run/qmp_{qemu_id}.sock" + self._temp[u"qga"] = f"/run/qga_{qemu_id}.sock" elif img == Constants.QEMU_VM_KERNEL: - self._opt['img'], _ = exec_cmd_no_error( - node, - 'ls -1 {img}* | tail -1'.format(img=Constants.QEMU_VM_KERNEL), - message='Qemu Kernel VM image not found!') - self._opt['vm_type'] = 'kernelvm' - self._temp['log'] = '/tmp/serial_{id}.log'.format(id=qemu_id) - self._temp['ini'] = '/etc/vm_init_{id}.conf'.format(id=qemu_id) - self._opt['initrd'], _ = exec_cmd_no_error( - node, - 'ls -1 {initrd}* | tail -1'.format( - initrd=Constants.QEMU_VM_KERNEL_INITRD), - message='Qemu Kernel initrd image not found!') + self._opt[u"img"], _ = exec_cmd_no_error( + node, f"ls -1 {Constants.QEMU_VM_KERNEL}* | tail -1", + message=u"Qemu Kernel VM image not found!" + ) + self._temp[u"ini"] = f"/etc/vm_init_{qemu_id}.conf" + self._opt[u"initrd"], _ = exec_cmd_no_error( + node, f"ls -1 {Constants.QEMU_VM_KERNEL_INITRD}* | tail -1", + message=u"Qemu Kernel initrd image not found!" + ) else: - raise RuntimeError('QEMU: Unknown VM image option: {}'.format(img)) + raise RuntimeError(f"QEMU: Unknown VM image option: {img}") # Computed parameters for QEMU command line. - self._params = OptionString(prefix='-') - self.add_params() - - def add_params(self): - """Set QEMU command line parameters.""" - self.add_default_params() - if self._opt.get('vm_type', '') == 'nestedvm': - self.add_nestedvm_params() - elif self._opt.get('vm_type', '') == 'kernelvm': - self.add_kernelvm_params() - else: - raise RuntimeError('QEMU: Unsupported VM type!') + self._params = OptionString(prefix=u"-") def add_default_params(self): """Set default QEMU command line parameters.""" - self._params.add('daemonize') - self._params.add('nodefaults') - self._params.add_with_value('name', 'vnf{qemu},debug-threads=on'.format( - qemu=self._opt.get('qemu_id'))) - self._params.add('no-user-config') - self._params.add_with_value('monitor', 'none') - self._params.add_with_value('display', 'none') - self._params.add_with_value('vga', 'none') - self._params.add('enable-kvm') - self._params.add_with_value('pidfile', self._temp.get('pidfile')) - self._params.add_with_value('cpu', 'host') - - if self._arch == 'aarch64': - machine_args = 'virt,accel=kvm,usb=off,mem-merge=off,gic-version=3' - else: - machine_args = 'pc,accel=kvm,usb=off,mem-merge=off' + mem_path = f"/dev/hugepages1G" \ + if self._opt[u"page_size"] == u"1G" else u"/dev/hugepages" + + self._params.add(u"daemonize") + self._params.add(u"nodefaults") + self._params.add_with_value( + u"name", f"vnf{self._opt.get(u'qemu_id')},debug-threads=on" + ) + self._params.add(u"no-user-config") + self._params.add(u"nographic") + self._params.add(u"enable-kvm") + self._params.add_with_value(u"pidfile", self._temp.get(u"pidfile")) + self._params.add_with_value(u"cpu", u"host") + + self._params.add_with_value(u"machine", self._opt.get(u"machine_args")) self._params.add_with_value( - 'machine', machine_args) + u"smp", f"{self._opt.get(u'smp')},sockets=1," + f"cores={self._opt.get(u'smp')},threads=1" + ) self._params.add_with_value( - 'smp', '{smp},sockets=1,cores={smp},threads=1'.format( - smp=self._opt.get('smp'))) + u"object", f"memory-backend-file,id=mem," + f"size={self._opt.get(u'mem')}M," + f"mem-path={mem_path},share=on" + ) + self._params.add_with_value(u"m", f"{self._opt.get(u'mem')}M") + self._params.add_with_value(u"numa", u"node,memdev=mem") + + def add_net_user(self, net="10.0.2.0/24"): + """Set managment port forwarding.""" self._params.add_with_value( - 'object', 'memory-backend-file,id=mem,size={mem}M,' - 'mem-path=/dev/hugepages,share=on'.format(mem=self._opt.get('mem'))) + u"netdev", f"user,id=mgmt,net={net}," + f"hostfwd=tcp::{self._vm_info[u'port']}-:22" + ) self._params.add_with_value( - 'm', '{mem}M'.format(mem=self._opt.get('mem'))) - self._params.add_with_value('numa', 'node,memdev=mem') - self._params.add_with_value('balloon', 'none') + u"device", f"virtio-net,netdev=mgmt" + ) - def add_nestedvm_params(self): - """Set NestedVM QEMU parameters.""" + def add_qmp_qga(self): + """Set QMP, QGA management.""" self._params.add_with_value( - 'net', 'nic,macaddr=52:54:00:00:{qemu:02x}:ff'.format( - qemu=self._opt.get('qemu_id'))) + u"chardev", f"socket,path={self._temp.get(u'qga')}," + f"server,nowait,id=qga0" + ) self._params.add_with_value( - 'net', 'user,hostfwd=tcp::{info[port]}-:22'.format( - info=self._vm_info)) - # TODO: Remove try except after fully migrated to Bionic or - # qemu_set_node is removed. - try: - locking = ',file.locking=off'\ - if self.qemu_version(version='2.10') else '' - except AttributeError: - locking = '' + u"device", u"isa-serial,chardev=qga0" + ) + self._params.add_with_value( + u"qmp", f"unix:{self._temp.get(u'qmp')},server,nowait" + ) + + def add_serial(self): + """Set serial to file redirect.""" + self._params.add_with_value( + u"chardev", f"socket,host=127.0.0.1," + f"port={self._vm_info[u'serial']},id=gnc0,server,nowait" + ) self._params.add_with_value( - 'drive', 'file={img},format=raw,cache=none,if=virtio{locking}'. - format(img=self._opt.get('img'), locking=locking)) + u"device", u"isa-serial,chardev=gnc0" + ) self._params.add_with_value( - 'qmp', 'unix:{qmp},server,nowait'.format(qmp=self._temp.get('qmp'))) + u"serial", f"file:{self._temp.get(u'log')}" + ) + + def add_drive_cdrom(self, drive_file, index=None): + """Set CD-ROM drive. + + :param drive_file: Path to drive image. + :param index: Drive index. + :type drive_file: str + :type index: int + """ + index = f"index={index}," if index else u"" self._params.add_with_value( - 'chardev', 'socket,host=127.0.0.1,port={info[serial]},' - 'id=gnc0,server,nowait'.format(info=self._vm_info)) - self._params.add_with_value('device', 'isa-serial,chardev=gnc0') + u"drive", f"file={drive_file},{index}media=cdrom" + ) + + def add_drive(self, drive_file, drive_format): + """Set drive with custom format. + + :param drive_file: Path to drive image. + :param drive_format: Drive image format. + :type drive_file: str + :type drive_format: str + """ self._params.add_with_value( - 'chardev', 'socket,path={qga},server,nowait,id=qga0'.format( - qga=self._temp.get('qga'))) - self._params.add_with_value('device', 'isa-serial,chardev=qga0') + u"drive", f"file={drive_file},format={drive_format}," + u"cache=none,if=virtio,file.locking=off" + ) def add_kernelvm_params(self): """Set KernelVM QEMU parameters.""" - console = 'ttyAMA0' if self._arch == 'aarch64' else 'ttyS0' - self._params.add_with_value('serial', 'file:{log}'.format( - log=self._temp.get('log'))) + hugepages = 3 if self._opt[u"page_size"] == u"1G" else 512 + self._params.add_with_value( - 'fsdev', 'local,id=root9p,path=/,security_model=none') + u"serial", f"file:{self._temp.get(u'log')}" + ) self._params.add_with_value( - 'device', 'virtio-9p-pci,fsdev=root9p,mount_tag=virtioroot') + u"fsdev", u"local,id=root9p,path=/,security_model=none" + ) self._params.add_with_value( - 'kernel', '{img}'.format(img=self._opt.get('img'))) + u"device", u"virtio-9p-pci,fsdev=root9p,mount_tag=virtioroot" + ) self._params.add_with_value( - 'initrd', '{initrd}'.format(initrd=self._opt.get('initrd'))) + u"kernel", f"{self._opt.get(u'img')}" + ) self._params.add_with_value( - 'append', '"ro rootfstype=9p rootflags=trans=virtio ' - 'root=virtioroot console={console} tsc=reliable ' - 'hugepages=256 init={init} fastboot"'.format( - console=console, init=self._temp.get('ini'))) + u"initrd", f"{self._opt.get(u'initrd')}" + ) + self._params.add_with_value( + u"append", f"'ro rootfstype=9p rootflags=trans=virtio " + f"root=virtioroot console={self._opt.get(u'console')} " + f"tsc=reliable hugepages={hugepages} " + f"hugepagesz={self._opt.get(u'page_size')} " + f"init={self._temp.get(u'ini')} fastboot'" + ) + + def add_vhost_user_if( + self, socket, server=True, jumbo_frames=False, queue_size=None, + queues=1, virtio_feature_mask=None): + """Add Vhost-user interface. + + :param socket: Path of the unix socket. + :param server: If True the socket shall be a listening socket. + :param jumbo_frames: Set True if jumbo frames are used in the test. + :param queue_size: Vring queue size. + :param queues: Number of queues. + :param virtio_feature_mask: Mask of virtio features to be enabled. + :type socket: str + :type server: bool + :type jumbo_frames: bool + :type queue_size: int + :type queues: int + :type virtio_feature_mask: int + """ + self._nic_id += 1 + if jumbo_frames: + logger.debug(u"Jumbo frames temporarily disabled!") + self._params.add_with_value( + u"chardev", f"socket,id=char{self._nic_id}," + f"path={socket}{u',server' if server is True else u''}" + ) + self._params.add_with_value( + u"netdev", f"vhost-user,id=vhost{self._nic_id}," + f"chardev=char{self._nic_id},queues={queues}" + ) + mac = f"52:54:00:00:{self._opt.get(u'qemu_id'):02x}:" \ + f"{self._nic_id:02x}" + queue_size = f"rx_queue_size={queue_size},tx_queue_size={queue_size}" \ + if queue_size else u"" + gso = VirtioFeatureMask.is_feature_enabled( + virtio_feature_mask, VirtioFeaturesFlags.VIRTIO_NET_F_API_GSO) + csum = VirtioFeatureMask.is_feature_enabled( + virtio_feature_mask, VirtioFeaturesFlags.VIRTIO_NET_F_API_CSUM) + + self._params.add_with_value( + u"device", f"virtio-net-pci,netdev=vhost{self._nic_id},mac={mac}," + f"addr={self._nic_id+5}.0,mq=on,vectors={2 * queues + 2}," + f"csum={u'on' if csum else u'off'}," + f"gso={u'on' if gso else u'off'}," + f"guest_tso4={u'on' if gso else u'off'}," + f"guest_tso6={u'on' if gso else u'off'}," + f"guest_ecn={u'on' if gso else u'off'}," + f"{queue_size}" + ) + + # Add interface MAC and socket to the node dict. + if_data = {u"mac_address": mac, u"socket": socket} + if_name = f"vhost{self._nic_id}" + self._vm_info[u"interfaces"][if_name] = if_data + # Add socket to temporary file list. + self._temp[if_name] = socket + + def add_vfio_pci_if(self, pci): + """Add VFIO PCI interface. + + :param pci: PCI address of interface. + :type pci: str + """ + self._nic_id += 1 + self._params.add_with_value( + u"device", f"vfio-pci,host={pci},addr={self._nic_id+5}.0" + ) def create_kernelvm_config_vpp(self, **kwargs): """Create QEMU VPP config files. @@ -206,15 +309,12 @@ class QemuUtils(object): file. :type kwargs: dict """ - startup = ('/etc/vpp/vm_startup_{id}.conf'. - format(id=self._opt.get('qemu_id'))) - running = ('/etc/vpp/vm_running_{id}.exec'. - format(id=self._opt.get('qemu_id'))) + startup = f"/etc/vpp/vm_startup_{self._opt.get(u'qemu_id')}.conf" + running = f"/etc/vpp/vm_running_{self._opt.get(u'qemu_id')}.exec" - self._temp['startup'] = startup - self._temp['running'] = running - self._opt['vnf_bin'] = ('/usr/bin/vpp -c {startup}'. - format(startup=startup)) + self._temp[u"startup"] = startup + self._temp[u"running"] = running + self._opt[u"vnf_bin"] = f"/usr/bin/vpp -c {startup}" # Create VPP startup configuration. vpp_config = VppConfigGenerator() @@ -223,32 +323,47 @@ class QemuUtils(object): vpp_config.add_unix_cli_listen() vpp_config.add_unix_exec(running) vpp_config.add_socksvr() - vpp_config.add_cpu_main_core('0') - if self._opt.get('smp') > 1: - vpp_config.add_cpu_corelist_workers('1-{smp}'.format( - smp=self._opt.get('smp')-1)) - vpp_config.add_dpdk_dev('0000:00:06.0', '0000:00:07.0') - vpp_config.add_dpdk_dev_default_rxq(kwargs['queues']) - vpp_config.add_dpdk_log_level('debug') - if not kwargs['jumbo_frames']: - vpp_config.add_dpdk_no_multi_seg() - vpp_config.add_dpdk_no_tx_checksum_offload() - vpp_config.add_plugin('disable', 'default') - vpp_config.add_plugin('enable', 'dpdk_plugin.so') - vpp_config.add_plugin('enable', 'memif_plugin.so') + vpp_config.add_main_heap_size(u"512M") + vpp_config.add_main_heap_page_size(self._opt[u"page_size"]) + vpp_config.add_default_hugepage_size(self._opt[u"page_size"]) + vpp_config.add_statseg_size(u"512M") + vpp_config.add_statseg_page_size(self._opt[u"page_size"]) + vpp_config.add_statseg_per_node_counters(u"on") + vpp_config.add_buffers_per_numa(107520) + vpp_config.add_cpu_main_core(u"0") + if self._opt.get(u"smp") > 1: + vpp_config.add_cpu_corelist_workers(f"1-{self._opt.get(u'smp')-1}") + vpp_config.add_plugin(u"disable", u"default") + vpp_config.add_plugin(u"enable", u"ping_plugin.so") + if "2vfpt" in self._opt.get(u'vnf'): + vpp_config.add_plugin(u"enable", u"avf_plugin.so") + if "vhost" in self._opt.get(u'vnf'): + vpp_config.add_plugin(u"enable", u"dpdk_plugin.so") + vpp_config.add_dpdk_dev(u"0000:00:06.0", u"0000:00:07.0") + vpp_config.add_dpdk_dev_default_rxq(kwargs[u"queues"]) + vpp_config.add_dpdk_log_level(u".*,debug") + if not kwargs[u"jumbo_frames"]: + vpp_config.add_dpdk_no_multi_seg() + vpp_config.add_dpdk_no_tx_checksum_offload() + if "ipsec" in self._opt.get(u'vnf'): + vpp_config.add_plugin(u"enable", u"crypto_native_plugin.so") + vpp_config.add_plugin(u"enable", u"crypto_ipsecmb_plugin.so") + vpp_config.add_plugin(u"enable", u"crypto_openssl_plugin.so") + if "nat" in self._opt.get(u'vnf'): + vpp_config.add_nat(value=u"endpoint-dependent") + vpp_config.add_plugin(u"enable", u"nat_plugin.so") vpp_config.write_config(startup) # Create VPP running configuration. - template = '{res}/{tpl}.exec'.format(res=Constants.RESOURCES_TPL_VM, - tpl=self._opt.get('vnf')) - exec_cmd_no_error(self._node, 'rm -f {running}'.format(running=running), - sudo=True) + template = f"{Constants.RESOURCES_TPL}/vm/{self._opt.get(u'vnf')}.exec" + exec_cmd_no_error(self._node, f"rm -f {running}", sudo=True) - with open(template, 'r') as src_file: + with open(template, u"rt") as src_file: src = Template(src_file.read()) exec_cmd_no_error( - self._node, "echo '{out}' | sudo tee {running}".format( - out=src.safe_substitute(**kwargs), running=running)) + self._node, f"echo '{src.safe_substitute(**kwargs)}' | " + f"sudo tee {running}" + ) def create_kernelvm_config_testpmd_io(self, **kwargs): """Create QEMU testpmd-io command line. @@ -256,20 +371,26 @@ class QemuUtils(object): :param kwargs: Key-value pairs to construct command line parameters. :type kwargs: dict """ + pmd_max_pkt_len = u"9200" if kwargs[u"jumbo_frames"] else u"1518" testpmd_cmd = DpdkUtil.get_testpmd_cmdline( - eal_corelist='0-{smp}'.format(smp=self._opt.get('smp') - 1), + eal_corelist=f"0-{self._opt.get(u'smp') - 1}", eal_driver=False, + eal_pci_whitelist0=u"0000:00:06.0", + eal_pci_whitelist1=u"0000:00:07.0", eal_in_memory=True, - pmd_num_mbufs=16384, - pmd_rxq=kwargs['queues'], - pmd_txq=kwargs['queues'], + pmd_num_mbufs=32768, + pmd_fwd_mode=u"io", + pmd_nb_ports=u"2", + pmd_portmask=u"0x3", + pmd_max_pkt_len=pmd_max_pkt_len, + pmd_mbuf_size=u"16384", + pmd_rxq=kwargs[u"queues"], + pmd_txq=kwargs[u"queues"], pmd_tx_offloads='0x0', - pmd_disable_hw_vlan=False, - pmd_nb_cores=str(self._opt.get('smp') - 1)) + pmd_nb_cores=str(self._opt.get(u"smp") - 1) + ) - self._opt['vnf_bin'] = ('{testpmd_path}/{testpmd_cmd}'. - format(testpmd_path=self._testpmd_path, - testpmd_cmd=testpmd_cmd)) + self._opt[u"vnf_bin"] = f"{self._testpmd_path}/{testpmd_cmd}" def create_kernelvm_config_testpmd_mac(self, **kwargs): """Create QEMU testpmd-mac command line. @@ -277,23 +398,32 @@ class QemuUtils(object): :param kwargs: Key-value pairs to construct command line parameters. :type kwargs: dict """ + pmd_max_pkt_len = u"9200" if kwargs[u"jumbo_frames"] else u"1518" testpmd_cmd = DpdkUtil.get_testpmd_cmdline( - eal_corelist='0-{smp}'.format(smp=self._opt.get('smp') - 1), + eal_corelist=f"0-{self._opt.get(u'smp') - 1}", eal_driver=False, + eal_pci_whitelist0=u"0000:00:06.0", + eal_pci_whitelist1=u"0000:00:07.0", eal_in_memory=True, - pmd_num_mbufs=16384, - pmd_fwd_mode='mac', - pmd_eth_peer_0='0,{mac}'.format(mac=kwargs['vif1_mac']), - pmd_eth_peer_1='1,{mac}'.format(mac=kwargs['vif2_mac']), - pmd_rxq=kwargs['queues'], - pmd_txq=kwargs['queues'], - pmd_tx_offloads='0x0', - pmd_disable_hw_vlan=False, - pmd_nb_cores=str(self._opt.get('smp') - 1)) - - self._opt['vnf_bin'] = ('{testpmd_path}/{testpmd_cmd}'. - format(testpmd_path=self._testpmd_path, - testpmd_cmd=testpmd_cmd)) + pmd_num_mbufs=32768, + pmd_fwd_mode=u"mac", + pmd_nb_ports=u"2", + pmd_portmask=u"0x3", + pmd_max_pkt_len=pmd_max_pkt_len, + pmd_mbuf_size=u"16384", + pmd_eth_peer_0=f"0,{kwargs[u'vif1_mac']}", + pmd_eth_peer_1=f"1,{kwargs[u'vif2_mac']}", + pmd_rxq=kwargs[u"queues"], + pmd_txq=kwargs[u"queues"], + pmd_tx_offloads=u"0x0", + pmd_nb_cores=str(self._opt.get(u"smp") - 1) + ) + + self._opt[u"vnf_bin"] = f"{self._testpmd_path}/{testpmd_cmd}" + + def create_kernelvm_config_iperf3(self): + """Create QEMU iperf3 command line.""" + self._opt[u"vnf_bin"] = f"mkdir /run/sshd; /usr/sbin/sshd -D -d" def create_kernelvm_init(self, **kwargs): """Create QEMU init script. @@ -301,18 +431,16 @@ class QemuUtils(object): :param kwargs: Key-value pairs to replace content of init startup file. :type kwargs: dict """ - template = '{res}/init.sh'.format(res=Constants.RESOURCES_TPL_VM) - init = self._temp.get('ini') - exec_cmd_no_error( - self._node, 'rm -f {init}'.format(init=init), sudo=True) + init = self._temp.get(u"ini") + exec_cmd_no_error(self._node, f"rm -f {init}", sudo=True) - with open(template, 'r') as src_file: + with open(kwargs[u"template"], u"rt") as src_file: src = Template(src_file.read()) exec_cmd_no_error( - self._node, "echo '{out}' | sudo tee {init}".format( - out=src.safe_substitute(**kwargs), init=init)) - exec_cmd_no_error( - self._node, "chmod +x {init}".format(init=init), sudo=True) + self._node, f"echo '{src.safe_substitute(**kwargs)}' | " + f"sudo tee {init}" + ) + exec_cmd_no_error(self._node, f"chmod +x {init}", sudo=True) def configure_kernelvm_vnf(self, **kwargs): """Create KernelVM VNF configurations. @@ -320,15 +448,36 @@ class QemuUtils(object): :param kwargs: Key-value pairs for templating configs. :type kwargs: dict """ - if 'vpp' in self._opt.get('vnf'): + if u"vpp" in self._opt.get(u"vnf"): self.create_kernelvm_config_vpp(**kwargs) - elif 'testpmd_io' in self._opt.get('vnf'): + self.create_kernelvm_init( + template=f"{Constants.RESOURCES_TPL}/vm/init.sh", + vnf_bin=self._opt.get(u"vnf_bin") + ) + elif u"testpmd_io" in self._opt.get(u"vnf"): self.create_kernelvm_config_testpmd_io(**kwargs) - elif 'testpmd_mac' in self._opt.get('vnf'): + self.create_kernelvm_init( + template=f"{Constants.RESOURCES_TPL}/vm/init.sh", + vnf_bin=self._opt.get(u"vnf_bin") + ) + elif u"testpmd_mac" in self._opt.get(u"vnf"): self.create_kernelvm_config_testpmd_mac(**kwargs) + self.create_kernelvm_init( + template=f"{Constants.RESOURCES_TPL}/vm/init.sh", + vnf_bin=self._opt.get(u"vnf_bin") + ) + elif u"iperf3" in self._opt.get(u"vnf"): + qemu_id = self._opt.get(u'qemu_id') % 2 + self.create_kernelvm_config_iperf3() + self.create_kernelvm_init( + template=f"{Constants.RESOURCES_TPL}/vm/init_iperf3.sh", + vnf_bin=self._opt.get(u"vnf_bin"), + ip_address_l=u"2.2.2.2/30" if qemu_id else u"1.1.1.1/30", + ip_address_r=u"2.2.2.1" if qemu_id else u"1.1.1.2", + ip_route_r=u"1.1.1.0/30" if qemu_id else u"2.2.2.0/30" + ) else: - raise RuntimeError('QEMU: Unsupported VNF!') - self.create_kernelvm_init(vnf_bin=self._opt['vnf_bin']) + raise RuntimeError(u"QEMU: Unsupported VNF!") def get_qemu_pids(self): """Get QEMU CPU pids. @@ -336,9 +485,9 @@ class QemuUtils(object): :returns: List of QEMU CPU pids. :rtype: list of str """ - command = ("grep -rwl 'CPU' /proc/$(sudo cat {pidfile})/task/*/comm ". - format(pidfile=self._temp.get('pidfile'))) - command += (r"| xargs dirname | sed -e 's/\/.*\///g' | uniq") + command = f"grep -rwl 'CPU' /proc/$(sudo cat " \ + f"{self._temp.get(u'pidfile')})/task/*/comm " + command += r"| xargs dirname | sed -e 's/\/.*\///g' | uniq" stdout, _ = exec_cmd_no_error(self._node, command) return stdout.splitlines() @@ -359,19 +508,19 @@ class QemuUtils(object): sleep(1) continue for qemu_cpu, host_cpu in zip(qemu_cpus, host_cpus): - command = ('taskset -pc {host_cpu} {thread}'. - format(host_cpu=host_cpu, thread=qemu_cpu)) - message = ('QEMU: Set affinity failed on {host}!'. - format(host=self._node['host'])) - exec_cmd_no_error(self._node, command, sudo=True, - message=message) + command = f"taskset -pc {host_cpu} {qemu_cpu}" + message = f"QEMU: Set affinity failed " \ + f"on {self._node[u'host']}!" + exec_cmd_no_error( + self._node, command, sudo=True, message=message + ) break except (RuntimeError, ValueError): self.qemu_kill_all() raise else: self.qemu_kill_all() - raise RuntimeError('Failed to set Qemu threads affinity!') + raise RuntimeError(u"Failed to set Qemu threads affinity!") def qemu_set_scheduler_policy(self): """Set scheduler policy to SCHED_RR with priority 1 for all Qemu CPU @@ -383,60 +532,15 @@ class QemuUtils(object): qemu_cpus = self.get_qemu_pids() for qemu_cpu in qemu_cpus: - command = ('chrt -r -p 1 {thread}'. - format(thread=qemu_cpu)) - message = ('QEMU: Set SCHED_RR failed on {host}'. - format(host=self._node['host'])) - exec_cmd_no_error(self._node, command, sudo=True, - message=message) + command = f"chrt -r -p 1 {qemu_cpu}" + message = f"QEMU: Set SCHED_RR failed on {self._node[u'host']}" + exec_cmd_no_error( + self._node, command, sudo=True, message=message + ) except (RuntimeError, ValueError): self.qemu_kill_all() raise - def qemu_add_vhost_user_if(self, socket, server=True, jumbo_frames=False, - queue_size=None, queues=1): - """Add Vhost-user interface. - - :param socket: Path of the unix socket. - :param server: If True the socket shall be a listening socket. - :param jumbo_frames: Set True if jumbo frames are used in the test. - :param queue_size: Vring queue size. - :param queues: Number of queues. - :type socket: str - :type server: bool - :type jumbo_frames: bool - :type queue_size: int - :type queues: int - """ - self._vhost_id += 1 - self._params.add_with_value( - 'chardev', 'socket,id=char{vhost},path={socket}{server}'.format( - vhost=self._vhost_id, socket=socket, - server=',server' if server is True else '')) - self._params.add_with_value( - 'netdev', 'vhost-user,id=vhost{vhost},chardev=char{vhost},' - 'queues={queues}'.format(vhost=self._vhost_id, queues=queues)) - mac = ('52:54:00:00:{qemu:02x}:{vhost:02x}'. - format(qemu=self._opt.get('qemu_id'), vhost=self._vhost_id)) - queue_size = ('rx_queue_size={queue_size},tx_queue_size={queue_size}'. - format(queue_size=queue_size)) if queue_size else '' - mbuf = 'on,host_mtu=9200' - self._params.add_with_value( - 'device', 'virtio-net-pci,netdev=vhost{vhost},mac={mac},' - 'addr={addr}.0,mq=on,vectors={vectors},csum=off,gso=off,' - 'guest_tso4=off,guest_tso6=off,guest_ecn=off,mrg_rxbuf={mbuf},' - '{queue_size}'.format( - addr=self._vhost_id+5, vhost=self._vhost_id, mac=mac, - mbuf=mbuf if jumbo_frames else 'off', queue_size=queue_size, - vectors=(2 * queues + 2))) - - # Add interface MAC and socket to the node dict. - if_data = {'mac_address': mac, 'socket': socket} - if_name = 'vhost{vhost}'.format(vhost=self._vhost_id) - self._vm_info['interfaces'][if_name] = if_data - # Add socket to temporary file list. - self._temp[if_name] = socket - def _qemu_qmp_exec(self, cmd): """Execute QMP command. @@ -449,32 +553,32 @@ class QemuUtils(object): response will contain the "error" keyword instead of "return". """ # To enter command mode, the qmp_capabilities command must be issued. - command = ('echo "{{ \\"execute\\": \\"qmp_capabilities\\" }}' - '{{ \\"execute\\": \\"{cmd}\\" }}" | ' - 'sudo -S socat - UNIX-CONNECT:{qmp}'. - format(cmd=cmd, qmp=self._temp.get('qmp'))) - message = ('QMP execute "{cmd}" failed on {host}'. - format(cmd=cmd, host=self._node['host'])) + command = f"echo \"{{{{ \\\"execute\\\": " \ + f"\\\"qmp_capabilities\\\" }}}}" \ + f"{{{{ \\\"execute\\\": \\\"{cmd}\\\" }}}}\" | " \ + f"sudo -S socat - UNIX-CONNECT:{self._temp.get(u'qmp')}" + message = f"QMP execute '{cmd}' failed on {self._node[u'host']}" + stdout, _ = exec_cmd_no_error( - self._node, command, sudo=False, message=message) + self._node, command, sudo=False, message=message + ) # Skip capabilities negotiation messages. out_list = stdout.splitlines() if len(out_list) < 3: - raise RuntimeError( - 'Invalid QMP output on {host}'.format(host=self._node['host'])) + raise RuntimeError(f"Invalid QMP output on {self._node[u'host']}") return json.loads(out_list[2]) def _qemu_qga_flush(self): """Flush the QGA parser state.""" - command = ('(printf "\xFF"; sleep 1) | ' - 'sudo -S socat - UNIX-CONNECT:{qga}'. - format(qga=self._temp.get('qga'))) - message = ('QGA flush failed on {host}'.format(host=self._node['host'])) + command = f"(printf \"\xFF\"; sleep 1) | sudo -S socat " \ + f"- UNIX-CONNECT:{self._temp.get(u'qga')}" + message = f"QGA flush failed on {self._node[u'host']}" stdout, _ = exec_cmd_no_error( - self._node, command, sudo=False, message=message) + self._node, command, sudo=False, message=message + ) - return json.loads(stdout.split('\n', 1)[0]) if stdout else dict() + return json.loads(stdout.split(u"\n", 1)[0]) if stdout else dict() def _qemu_qga_exec(self, cmd): """Execute QGA command. @@ -484,27 +588,51 @@ class QemuUtils(object): :param cmd: QGA command to execute. :type cmd: str """ - command = ('(echo "{{ \\"execute\\": \\"{cmd}\\" }}"; sleep 1) | ' - 'sudo -S socat - UNIX-CONNECT:{qga}'. - format(cmd=cmd, qga=self._temp.get('qga'))) - message = ('QGA execute "{cmd}" failed on {host}'. - format(cmd=cmd, host=self._node['host'])) + command = f"(echo \"{{{{ \\\"execute\\\": " \ + f"\\\"{cmd}\\\" }}}}\"; sleep 1) | " \ + f"sudo -S socat - UNIX-CONNECT:{self._temp.get(u'qga')}" + message = f"QGA execute '{cmd}' failed on {self._node[u'host']}" stdout, _ = exec_cmd_no_error( - self._node, command, sudo=False, message=message) + self._node, command, sudo=False, message=message + ) - return json.loads(stdout.split('\n', 1)[0]) if stdout else dict() + return json.loads(stdout.split(u"\n", 1)[0]) if stdout else dict() def _wait_until_vm_boot(self): - """Wait until QEMU with NestedVM is booted.""" - if self._opt.get('vm_type') == 'nestedvm': - self._wait_until_nestedvm_boot() - self._update_vm_interfaces() - elif self._opt.get('vm_type') == 'kernelvm': - self._wait_until_kernelvm_boot() + """Wait until QEMU VM is booted.""" + try: + getattr(self, f'_wait_{self._opt["vnf"]}')() + except AttributeError: + self._wait_default() + + def _wait_default(self, retries=60): + """Wait until QEMU with VPP is booted. + + :param retries: Number of retries. + :type retries: int + """ + for _ in range(retries): + command = f"tail -1 {self._temp.get(u'log')}" + stdout = None + try: + stdout, _ = exec_cmd_no_error(self._node, command, sudo=True) + sleep(1) + except RuntimeError: + pass + if "vpp " in stdout and "built by" in stdout: + break + if u"Press enter to exit" in stdout: + break + if u"reboot: Power down" in stdout: + raise RuntimeError( + f"QEMU: NF failed to run on {self._node[u'host']}!" + ) else: - raise RuntimeError('QEMU: Unsupported VM type!') + raise RuntimeError( + f"QEMU: Timeout, VM not booted on {self._node[u'host']}!" + ) - def _wait_until_nestedvm_boot(self, retries=12): + def _wait_nestedvm(self, retries=12): """Wait until QEMU with NestedVM is booted. First try to flush qga until there is output. @@ -518,91 +646,79 @@ class QemuUtils(object): try: out = self._qemu_qga_flush() except ValueError: - logger.trace('QGA qga flush unexpected output {out}'. - format(out=out)) + logger.trace(f"QGA qga flush unexpected output {out}") # Empty output - VM not booted yet if not out: sleep(5) else: break else: - raise RuntimeError('QEMU: Timeout, VM not booted on {host}!'. - format(host=self._node['host'])) + raise RuntimeError( + f"QEMU: Timeout, VM not booted on {self._node[u'host']}!" + ) for _ in range(retries): out = None try: - out = self._qemu_qga_exec('guest-ping') + out = self._qemu_qga_exec(u"guest-ping") except ValueError: - logger.trace('QGA guest-ping unexpected output {out}'. - format(out=out)) + logger.trace(f"QGA guest-ping unexpected output {out}") # Empty output - VM not booted yet. if not out: sleep(5) # Non-error return - VM booted. - elif out.get('return') is not None: + elif out.get(u"return") is not None: break # Skip error and wait. - elif out.get('error') is not None: + elif out.get(u"error") is not None: sleep(5) else: # If there is an unexpected output from QGA guest-info, try # again until timeout. - logger.trace('QGA guest-ping unexpected output {out}'. - format(out=out)) + logger.trace(f"QGA guest-ping unexpected output {out}") else: - raise RuntimeError('QEMU: Timeout, VM not booted on {host}!'. - format(host=self._node['host'])) + raise RuntimeError( + f"QEMU: Timeout, VM not booted on {self._node[u'host']}!" + ) - def _wait_until_kernelvm_boot(self, retries=60): - """Wait until QEMU KernelVM is booted. + def _wait_iperf3(self, retries=60): + """Wait until QEMU with iPerf3 is booted. :param retries: Number of retries. :type retries: int """ - vpp_ver = VPPUtil.vpp_show_version(self._node) - - for _ in range(retries): - command = ('tail -1 {log}'.format(log=self._temp.get('log'))) - stdout = None - try: - stdout, _ = exec_cmd_no_error(self._node, command, sudo=True) - sleep(1) - except RuntimeError: - pass - if vpp_ver in stdout or 'Press enter to exit' in stdout: - break - if 'reboot: Power down' in stdout: - raise RuntimeError('QEMU: NF failed to run on {host}!'. - format(host=self._node['host'])) - else: - raise RuntimeError('QEMU: Timeout, VM not booted on {host}!'. - format(host=self._node['host'])) + grep = u"Server listening on 0.0.0.0 port 22." + cmd = f"fgrep '{grep}' {self._temp.get(u'log')}" + message = f"QEMU: Timeout, VM not booted on {self._node[u'host']}!" + exec_cmd_no_error( + self._node, cmd=cmd, sudo=True, message=message, retries=retries, + include_reason=True + ) def _update_vm_interfaces(self): """Update interface names in VM node dict.""" # Send guest-network-get-interfaces command via QGA, output example: # {"return": [{"name": "eth0", "hardware-address": "52:54:00:00:04:01"}, # {"name": "eth1", "hardware-address": "52:54:00:00:04:02"}]}. - out = self._qemu_qga_exec('guest-network-get-interfaces') - interfaces = out.get('return') + out = self._qemu_qga_exec(u"guest-network-get-interfaces") + interfaces = out.get(u"return") mac_name = {} if not interfaces: - raise RuntimeError('Get VM interface list failed on {host}'. - format(host=self._node['host'])) + raise RuntimeError( + f"Get VM interface list failed on {self._node[u'host']}" + ) # Create MAC-name dict. for interface in interfaces: - if 'hardware-address' not in interface: + if u"hardware-address" not in interface: continue - mac_name[interface['hardware-address']] = interface['name'] + mac_name[interface[u"hardware-address"]] = interface[u"name"] # Match interface by MAC and save interface name. - for interface in self._vm_info['interfaces'].values(): - mac = interface.get('mac_address') + for interface in self._vm_info[u"interfaces"].values(): + mac = interface.get(u"mac_address") if_name = mac_name.get(mac) if if_name is None: - logger.trace( - 'Interface name for MAC {mac} not found'.format(mac=mac)) + logger.trace(f"Interface name for MAC {mac} not found") else: - interface['name'] = if_name + interface[u"name"] = if_name def qemu_start(self): """Start QEMU and wait until VM boot. @@ -611,17 +727,18 @@ class QemuUtils(object): :rtype: dict """ cmd_opts = OptionString() - cmd_opts.add('{bin_path}/qemu-system-{arch}'.format( - bin_path=Constants.QEMU_BIN_PATH, arch=self._arch)) + cmd_opts.add(f"{Constants.QEMU_BIN_PATH}/qemu-system-{self._arch}") cmd_opts.extend(self._params) - message = ('QEMU: Start failed on {host}!'. - format(host=self._node['host'])) + message = f"QEMU: Start failed on {self._node[u'host']}!" try: DUTSetup.check_huge_page( - self._node, '/dev/hugepages', self._opt.get('mem')) + self._node, self._opt.get(u"mem-path"), + int(self._opt.get(u"mem")) + ) exec_cmd_no_error( - self._node, cmd_opts, timeout=300, sudo=True, message=message) + self._node, cmd_opts, timeout=300, sudo=True, message=message + ) self._wait_until_vm_boot() except RuntimeError: self.qemu_kill_all() @@ -630,39 +747,37 @@ class QemuUtils(object): def qemu_kill(self): """Kill qemu process.""" - exec_cmd(self._node, 'chmod +r {pidfile}'. - format(pidfile=self._temp.get('pidfile')), sudo=True) - exec_cmd(self._node, 'kill -SIGKILL $(cat {pidfile})'. - format(pidfile=self._temp.get('pidfile')), sudo=True) + exec_cmd( + self._node, f"chmod +r {self._temp.get(u'pidfile')}", sudo=True + ) + exec_cmd( + self._node, f"kill -SIGKILL $(cat {self._temp.get(u'pidfile')})", + sudo=True + ) for value in self._temp.values(): - exec_cmd(self._node, 'cat {value}'.format(value=value), sudo=True) - exec_cmd(self._node, 'rm -f {value}'.format(value=value), sudo=True) + exec_cmd(self._node, f"cat {value}", sudo=True) + exec_cmd(self._node, f"rm -f {value}", sudo=True) def qemu_kill_all(self): """Kill all qemu processes on DUT node if specified.""" - exec_cmd(self._node, 'pkill -SIGKILL qemu', sudo=True) + exec_cmd(self._node, u"pkill -SIGKILL qemu", sudo=True) for value in self._temp.values(): - exec_cmd(self._node, 'cat {value}'.format(value=value), sudo=True) - exec_cmd(self._node, 'rm -f {value}'.format(value=value), sudo=True) + exec_cmd(self._node, f"cat {value}", sudo=True) + exec_cmd(self._node, f"rm -f {value}", sudo=True) - def qemu_version(self, version=None): - """Return Qemu version or compare if version is higher than parameter. + def qemu_version(self): + """Return Qemu version. - :param version: Version to compare. - :type version: str - :returns: Qemu version or Boolean if version is higher than parameter. - :rtype: str or bool + :returns: Qemu version. + :rtype: str """ - command = ('{bin_path}/qemu-system-{arch} --version'.format( - bin_path=Constants.QEMU_BIN_PATH, - arch=self._arch)) + command = f"{Constants.QEMU_BIN_PATH}/qemu-system-{self._arch} " \ + f"--version" try: stdout, _ = exec_cmd_no_error(self._node, command, sudo=True) - ver = match(r'QEMU emulator version ([\d.]*)', stdout).group(1) - return StrictVersion(ver) > StrictVersion(version) \ - if version else ver + return match(r"QEMU emulator version ([\d.]*)", stdout).group(1) except RuntimeError: self.qemu_kill_all() raise