From ce12c26a285093f8c297d167c28eaf984df85cb7 Mon Sep 17 00:00:00 2001 From: pmikus Date: Wed, 22 Jul 2020 12:35:54 +0000 Subject: [PATCH] Framework: Profiles load Signed-off-by: pmikus Change-Id: Iab84aff31a23bb9d8e1165f5314004803fd8a501 --- resources/libraries/python/QemuManager.py | 89 +++++--- resources/libraries/python/QemuUtils.py | 250 +++++++++++---------- .../libraries/robot/shared/test_teardown.robot | 25 ++- resources/libraries/robot/shared/vm.robot | 40 ++++ 4 files changed, 264 insertions(+), 140 deletions(-) diff --git a/resources/libraries/python/QemuManager.py b/resources/libraries/python/QemuManager.py index ce55be0a51..66a21aa115 100644 --- a/resources/libraries/python/QemuManager.py +++ b/resources/libraries/python/QemuManager.py @@ -56,15 +56,12 @@ class QemuManager: nf_dtcr = kwargs[u"nf_dtcr"] \ if isinstance(kwargs[u"nf_dtcr"], int) else 2 - img = Constants.QEMU_VM_KERNEL - for nf_chain in range(1, nf_chains + 1): for nf_node in range(1, nf_nodes + 1): qemu_id = (nf_chain - 1) * nf_nodes + nf_node name = f"{node}_{qemu_id}" - sock1 = f"/run/vpp/sock-{qemu_id}-1" - sock2 = f"/run/vpp/sock-{qemu_id}-2" idx1 = (nf_chain - 1) * nf_nodes * 2 + nf_node * 2 - 1 + vif1_mac = Topology.get_interface_mac( self.nodes[node], f"vhost{idx1}" ) if kwargs[u"vnf"] == u"testpmd_mac" \ @@ -83,27 +80,15 @@ class QemuManager: vs_dtc=vs_dtc, nf_dtc=nf_dtc, nf_dtcr=nf_dtcr ) - self.machines[name] = QemuUtils( - node=self.nodes[node], qemu_id=qemu_id, - smp=len(self.machines_affinity[name]), mem=4096, - vnf=kwargs[u"vnf"], img=img - ) - self.machines[name].configure_kernelvm_vnf( - mac1=f"52:54:00:00:{qemu_id:02x}:01", - mac2=f"52:54:00:00:{qemu_id:02x}:02", - vif1_mac=vif1_mac, vif2_mac=vif2_mac, queues=queues, - jumbo_frames=kwargs[u"jumbo"] - ) - self.machines[name].qemu_add_vhost_user_if( - sock1, jumbo_frames=kwargs[u"jumbo"], queues=queues, - queue_size=kwargs[u"perf_qemu_qsz"], - csum=kwargs[u"enable_csum"], gso=kwargs[u"enable_gso"] - ) - self.machines[name].qemu_add_vhost_user_if( - sock2, jumbo_frames=kwargs[u"jumbo"], queues=queues, - queue_size=kwargs[u"perf_qemu_qsz"], - csum=kwargs[u"enable_csum"], gso=kwargs[u"enable_gso"] - ) + try: + getattr(self, f'_c_{kwargs["vnf"]}')( + qemu_id=qemu_id, name=name, queues=queues, **kwargs + ) + except AttributeError: + self._c_default( + qemu_id=qemu_id, name=name, queues=queues, + vif1_mac=vif1_mac, vif2_mac=vif2_mac, **kwargs + ) def construct_vms_on_all_nodes(self, **kwargs): """Construct 1..Mx1..N VMs(s) with specified name on all nodes. @@ -124,7 +109,9 @@ class QemuManager: """ for machine, machine_affinity in \ zip(self.machines.values(), self.machines_affinity.values()): - machine.qemu_start() + index = list(self.machines.values()).index(machine) + name = list(self.machines.keys())[index] + self.nodes[name] = machine.qemu_start() if pinning: machine.qemu_set_affinity(*machine_affinity) @@ -134,8 +121,58 @@ class QemuManager: :param force: Force kill all Qemu instances by pkill qemu if True. :type force: bool """ + for node in list(self.nodes.values()): + if node["type"] == NodeType.VM: + try: + self.nodes.popitem(node) + except TypeError: + pass for machine in self.machines.values(): if force: machine.qemu_kill_all() else: machine.qemu_kill() + + def _c_default(self, **kwargs): + """Instantiate one VM with default configuration. + + :param kwargs: Named parameters. + :type kwargs: dict + """ + qemu_id = kwargs[u"qemu_id"] + name = kwargs[u"name"] + + self.machines[name] = QemuUtils( + node=self.nodes[kwargs[u"node"]], + qemu_id=qemu_id, + smp=len(self.machines_affinity[name]), + mem=4096, + vnf=kwargs[u"vnf"], + img=Constants.QEMU_VM_KERNEL + ) + self.machines[name].add_default_params() + self.machines[name].add_kernelvm_params() + self.machines[name].configure_kernelvm_vnf( + mac1=f"52:54:00:00:{qemu_id:02x}:01", + mac2=f"52:54:00:00:{qemu_id:02x}:02", + vif1_mac=kwargs[u"vif1_mac"], + vif2_mac=kwargs[u"vif2_mac"], + queues=kwargs[u"queues"], + jumbo_frames=kwargs[u"jumbo"] + ) + self.machines[name].add_vhost_user_if( + f"/run/vpp/sock-{qemu_id}-1", + jumbo_frames=kwargs[u"jumbo"], + queues=kwargs[u"queues"], + queue_size=kwargs[u"perf_qemu_qsz"], + csum=kwargs[u"enable_csum"], + gso=kwargs[u"enable_gso"] + ) + self.machines[name].add_vhost_user_if( + f"/run/vpp/sock-{qemu_id}-2", + jumbo_frames=kwargs[u"jumbo"], + queues=kwargs[u"queues"], + queue_size=kwargs[u"perf_qemu_qsz"], + csum=kwargs[u"enable_csum"], + gso=kwargs[u"enable_gso"] + ) diff --git a/resources/libraries/python/QemuUtils.py b/resources/libraries/python/QemuUtils.py index 3ba64f0dbf..bb0c20c554 100644 --- a/resources/libraries/python/QemuUtils.py +++ b/resources/libraries/python/QemuUtils.py @@ -28,7 +28,6 @@ from resources.libraries.python.OptionString import OptionString from resources.libraries.python.ssh import exec_cmd, exec_cmd_no_error from resources.libraries.python.topology import NodeType, Topology from resources.libraries.python.VppConfigGenerator import VppConfigGenerator -from resources.libraries.python.VPPUtil import VPPUtil __all__ = [u"QemuUtils"] @@ -95,9 +94,9 @@ class QemuUtils: self._opt[u"vnf"] = vnf # Temporary files. self._temp = dict() + self._temp[u"log"] = f"/tmp/serial_{qemu_id}.log" self._temp[u"pidfile"] = f"/run/qemu_{qemu_id}.pid" if img == Constants.QEMU_VM_IMAGE: - self._opt[u"vm_type"] = u"nestedvm" self._temp[u"qmp"] = f"/run/qmp_{qemu_id}.sock" self._temp[u"qga"] = f"/run/qga_{qemu_id}.sock" elif img == Constants.QEMU_VM_KERNEL: @@ -105,8 +104,6 @@ class QemuUtils: node, f"ls -1 {Constants.QEMU_VM_KERNEL}* | tail -1", message=u"Qemu Kernel VM image not found!" ) - self._opt[u"vm_type"] = u"kernelvm" - self._temp[u"log"] = f"/tmp/serial_{qemu_id}.log" self._temp[u"ini"] = f"/etc/vm_init_{qemu_id}.conf" self._opt[u"initrd"], _ = exec_cmd_no_error( node, f"ls -1 {Constants.QEMU_VM_KERNEL_INITRD}* | tail -1", @@ -116,17 +113,6 @@ class QemuUtils: raise RuntimeError(f"QEMU: Unknown VM image option: {img}") # Computed parameters for QEMU command line. self._params = OptionString(prefix=u"-") - self.add_params() - - def add_params(self): - """Set QEMU command line parameters.""" - self.add_default_params() - if self._opt.get(u"vm_type", u"") == u"nestedvm": - self.add_nestedvm_params() - elif self._opt.get(u"vm_type", u"") == u"kernelvm": - self.add_kernelvm_params() - else: - raise RuntimeError(u"QEMU: Unsupported VM type!") def add_default_params(self): """Set default QEMU command line parameters.""" @@ -136,9 +122,7 @@ class QemuUtils: u"name", f"vnf{self._opt.get(u'qemu_id')},debug-threads=on" ) self._params.add(u"no-user-config") - self._params.add_with_value(u"monitor", u"none") - self._params.add_with_value(u"display", u"none") - self._params.add_with_value(u"vga", u"none") + self._params.add(u"nographic") self._params.add(u"enable-kvm") self._params.add_with_value(u"pidfile", self._temp.get(u"pidfile")) self._params.add_with_value(u"cpu", u"host") @@ -156,32 +140,59 @@ class QemuUtils: self._params.add_with_value(u"numa", u"node,memdev=mem") self._params.add_with_value(u"balloon", u"none") - def add_nestedvm_params(self): - """Set NestedVM QEMU parameters.""" + def add_net_user(self): + """Set managment port forwarding.""" self._params.add_with_value( - u"net", - f"nic,macaddr=52:54:00:00:{self._opt.get(u'qemu_id'):02x}:ff" + u"netdev", f"user,id=mgmt,net=192.168.76.0/24," + f"hostfwd=tcp::{self._vm_info[u'port']}-:22" ) self._params.add_with_value( - u"net", f"user,hostfwd=tcp::{self._vm_info[u'port']}-:22" + u"device", f"virtio-net,netdev=mgmt" ) - locking = u",file.locking=off" + + def add_qmp_qga(self): + """Set QMP, QGA management.""" self._params.add_with_value( - u"drive", f"file={self._opt.get(u'img')}," - f"format=raw,cache=none,if=virtio{locking}" + u"chardev", f"socket,path={self._temp.get(u'qga')}," + f"server,nowait,id=qga0" ) + self._params.add_with_value(u"device", u"isa-serial,chardev=qga0") self._params.add_with_value( u"qmp", f"unix:{self._temp.get(u'qmp')},server,nowait" ) + + def add_serial(self): + """Set serial to file redirect.""" self._params.add_with_value( u"chardev", f"socket,host=127.0.0.1," f"port={self._vm_info[u'serial']},id=gnc0,server,nowait") self._params.add_with_value(u"device", u"isa-serial,chardev=gnc0") self._params.add_with_value( - u"chardev", f"socket,path={self._temp.get(u'qga')}," - f"server,nowait,id=qga0" + u"serial", f"file:{self._temp.get(u'log')}" + ) + + def add_drive_cdrom(self, drive_file): + """Set CD-ROM drive. + + :param drive_file: Path to drive image. + :type drive_file: str + """ + self._params.add_with_value( + u"drive", f"file={drive_file},media=cdrom" + ) + + def add_drive(self, drive_file, drive_format): + """Set drive with custom format. + + :param drive_file: Path to drive image. + :param drive_format: Drive image format. + :type drive_file: str + :type drive_format: str + """ + self._params.add_with_value( + u"drive", f"file={drive_file},format={drive_format}," + u"cache=none,if=virtio,file.locking=off" ) - self._params.add_with_value(u"device", u"isa-serial,chardev=qga0") def add_kernelvm_params(self): """Set KernelVM QEMU parameters.""" @@ -203,6 +214,62 @@ class QemuUtils: f"init={self._temp.get(u'ini')} fastboot'" ) + def add_vhost_user_if( + self, socket, server=True, jumbo_frames=False, queue_size=None, + queues=1, csum=False, gso=False): + """Add Vhost-user interface. + + :param socket: Path of the unix socket. + :param server: If True the socket shall be a listening socket. + :param jumbo_frames: Set True if jumbo frames are used in the test. + :param queue_size: Vring queue size. + :param queues: Number of queues. + :param csum: Checksum offloading. + :param gso: Generic segmentation offloading. + :type socket: str + :type server: bool + :type jumbo_frames: bool + :type queue_size: int + :type queues: int + :type csum: bool + :type gso: bool + """ + self._vhost_id += 1 + self._params.add_with_value( + u"chardev", f"socket,id=char{self._vhost_id}," + f"path={socket}{u',server' if server is True else u''}" + ) + self._params.add_with_value( + u"netdev", f"vhost-user,id=vhost{self._vhost_id}," + f"chardev=char{self._vhost_id},queues={queues}" + ) + mac = f"52:54:00:00:{self._opt.get(u'qemu_id'):02x}:" \ + f"{self._vhost_id:02x}" + queue_size = f"rx_queue_size={queue_size},tx_queue_size={queue_size}" \ + if queue_size else u"" + self._params.add_with_value( + u"device", f"virtio-net-pci,netdev=vhost{self._vhost_id},mac={mac}," + f"addr={self._vhost_id+5}.0,mq=on,vectors={2 * queues + 2}," + f"csum={u'on' if csum else u'off'},gso={u'on' if gso else u'off'}," + f"guest_tso4=off,guest_tso6=off,guest_ecn=off," + f"{queue_size}" + ) + + # Add interface MAC and socket to the node dict. + if_data = {u"mac_address": mac, u"socket": socket} + if_name = f"vhost{self._vhost_id}" + self._vm_info[u"interfaces"][if_name] = if_data + # Add socket to temporary file list. + self._temp[if_name] = socket + + def add_vfio_pci_if(self, pci): + """Add VFIO PCI interface. + + :param pci: PCI address of interface. + :type pci: str + """ + self._params.add_with_value(u"device", f"vfio-pci,host={pci}") + def create_kernelvm_config_vpp(self, **kwargs): """Create QEMU VPP config files. @@ -224,11 +291,19 @@ class QemuUtils: vpp_config.add_unix_cli_listen() vpp_config.add_unix_exec(running) vpp_config.add_socksvr() + vpp_config.add_buffers_per_numa(107520) vpp_config.add_cpu_main_core(u"0") if self._opt.get(u"smp") > 1: vpp_config.add_cpu_corelist_workers(f"1-{self._opt.get(u'smp')-1}") vpp_config.add_plugin(u"disable", u"default") - if "virtio" not in self._opt.get(u'vnf'): + vpp_config.add_plugin(u"enable", u"ping_plugin.so") + if "ipsec" in self._opt.get(u'vnf'): + vpp_config.add_plugin(u"enable", u"crypto_native_plugin.so") + vpp_config.add_plugin(u"enable", u"crypto_ipsecmb_plugin.so") + vpp_config.add_plugin(u"enable", u"crypto_openssl_plugin.so") + if "2vfpt" in self._opt.get(u'vnf'): + vpp_config.add_plugin(u"enable", u"avf_plugin.so") + if "vhost" in self._opt.get(u'vnf'): vpp_config.add_plugin(u"enable", u"dpdk_plugin.so") vpp_config.add_dpdk_dev(u"0000:00:06.0", u"0000:00:07.0") vpp_config.add_dpdk_dev_default_rxq(kwargs[u"queues"]) @@ -401,54 +476,6 @@ class QemuUtils: self.qemu_kill_all() raise - def qemu_add_vhost_user_if( - self, socket, server=True, jumbo_frames=False, queue_size=None, - queues=1, csum=False, gso=False): - """Add Vhost-user interface. - - :param socket: Path of the unix socket. - :param server: If True the socket shall be a listening socket. - :param jumbo_frames: Set True if jumbo frames are used in the test. - :param queue_size: Vring queue size. - :param queues: Number of queues. - :param csum: Checksum offloading. - :param gso: Generic segmentation offloading. - :type socket: str - :type server: bool - :type jumbo_frames: bool - :type queue_size: int - :type queues: int - :type csum: bool - :type gso: bool - """ - self._vhost_id += 1 - self._params.add_with_value( - u"chardev", f"socket,id=char{self._vhost_id}," - f"path={socket}{u',server' if server is True else u''}" - ) - self._params.add_with_value( - u"netdev", f"vhost-user,id=vhost{self._vhost_id}," - f"chardev=char{self._vhost_id},queues={queues}" - ) - mac = f"52:54:00:00:{self._opt.get(u'qemu_id'):02x}:" \ - f"{self._vhost_id:02x}" - queue_size = f"rx_queue_size={queue_size},tx_queue_size={queue_size}" \ - if queue_size else u"" - self._params.add_with_value( - u"device", f"virtio-net-pci,netdev=vhost{self._vhost_id},mac={mac}," - f"addr={self._vhost_id+5}.0,mq=on,vectors={2 * queues + 2}," - f"csum={u'on' if csum else u'off'},gso={u'on' if gso else u'off'}," - f"guest_tso4=off,guest_tso6=off,guest_ecn=off," - f"{queue_size}" - ) - - # Add interface MAC and socket to the node dict. - if_data = {u"mac_address": mac, u"socket": socket} - if_name = f"vhost{self._vhost_id}" - self._vm_info[u"interfaces"][if_name] = if_data - # Add socket to temporary file list. - self._temp[if_name] = socket - def _qemu_qmp_exec(self, cmd): """Execute QMP command. @@ -507,16 +534,40 @@ class QemuUtils: return json.loads(stdout.split(u"\n", 1)[0]) if stdout else dict() def _wait_until_vm_boot(self): - """Wait until QEMU with NestedVM is booted.""" - if self._opt.get(u"vm_type") == u"nestedvm": - self._wait_until_nestedvm_boot() - self._update_vm_interfaces() - elif self._opt.get(u"vm_type") == u"kernelvm": - self._wait_until_kernelvm_boot() + """Wait until QEMU VM is booted.""" + try: + getattr(self, f'_wait_{self._opt["vnf"]}')() + except AttributeError: + self._wait_default() + + def _wait_default(self, retries=60): + """Wait until QEMU with VPP is booted. + + :param retries: Number of retries. + :type retries: int + """ + for _ in range(retries): + command = f"tail -1 {self._temp.get(u'log')}" + stdout = None + try: + stdout, _ = exec_cmd_no_error(self._node, command, sudo=True) + sleep(1) + except RuntimeError: + pass + if "vpp " in stdout and "built by" in stdout: + break + if u"Press enter to exit" in stdout: + break + if u"reboot: Power down" in stdout: + raise RuntimeError( + f"QEMU: NF failed to run on {self._node[u'host']}!" + ) else: - raise RuntimeError(u"QEMU: Unsupported VM type!") + raise RuntimeError( + f"QEMU: Timeout, VM not booted on {self._node[u'host']}!" + ) - def _wait_until_nestedvm_boot(self, retries=12): + def _wait_nestedvm(self, retries=12): """Wait until QEMU with NestedVM is booted. First try to flush qga until there is output. @@ -564,33 +615,6 @@ class QemuUtils: f"QEMU: Timeout, VM not booted on {self._node[u'host']}!" ) - def _wait_until_kernelvm_boot(self, retries=60): - """Wait until QEMU KernelVM is booted. - - :param retries: Number of retries. - :type retries: int - """ - vpp_ver = VPPUtil.vpp_show_version(self._node) - - for _ in range(retries): - command = f"tail -1 {self._temp.get(u'log')}" - stdout = None - try: - stdout, _ = exec_cmd_no_error(self._node, command, sudo=True) - sleep(1) - except RuntimeError: - pass - if vpp_ver in stdout or u"Press enter to exit" in stdout: - break - if u"reboot: Power down" in stdout: - raise RuntimeError( - f"QEMU: NF failed to run on {self._node[u'host']}!" - ) - else: - raise RuntimeError( - f"QEMU: Timeout, VM not booted on {self._node[u'host']}!" - ) - def _update_vm_interfaces(self): """Update interface names in VM node dict.""" # Send guest-network-get-interfaces command via QGA, output example: diff --git a/resources/libraries/robot/shared/test_teardown.robot b/resources/libraries/robot/shared/test_teardown.robot index 68bbe716f2..b5cdab0664 100644 --- a/resources/libraries/robot/shared/test_teardown.robot +++ b/resources/libraries/robot/shared/test_teardown.robot @@ -24,7 +24,7 @@ *** Keywords *** | Tear down test | | [Documentation] -| | ... | Common test teardown for tests. +| | ... | Common test teardown for VPP tests. | | | | ... | *Arguments:* | | ... | - ${actions} - Additional teardown action. Type: list @@ -46,6 +46,21 @@ | | END | | Clean Sockets On All Nodes | ${nodes} +| Tear down test raw +| | [Documentation] +| | ... | Common test teardown for raw tests. +| | +| | ... | *Arguments:* +| | ... | - ${actions} - Additional teardown action. Type: list +| | +| | [Arguments] | @{actions} +| | +| | Remove All Added Ports On All DUTs From Topology | ${nodes} +| | FOR | ${action} | IN | @{actions} +| | | Run Keyword | Additional Test Tear Down Action For ${action} +| | END +| | Clean Sockets On All Nodes | ${nodes} + | Additional Test Tear Down Action For performance | | [Documentation] | | ... | Additional teardown for tests which uses performance measurement. @@ -78,6 +93,14 @@ | | ... | Keyword Should Exist | vnf_manager.Kill All VMs | | Run Keyword If | '${vnf_status}' == 'PASS' | vnf_manager.Kill All VMs +| Additional Test Tear Down Action For vhost-pt +| | [Documentation] +| | ... | Additional teardown for tests which uses pci-passtrough and VM(s). +| | +| | ${vnf_status} | ${value}= | Run Keyword And Ignore Error +| | ... | Keyword Should Exist | vnf_manager.Kill All VMs +| | Run Keyword If | '${vnf_status}' == 'PASS' | vnf_manager.Kill All VMs + | Additional Test Tear Down Action For nat | | [Documentation] | | ... | Additional teardown for tests which uses NAT feature. diff --git a/resources/libraries/robot/shared/vm.robot b/resources/libraries/robot/shared/vm.robot index 21fce77683..2f15853779 100644 --- a/resources/libraries/robot/shared/vm.robot +++ b/resources/libraries/robot/shared/vm.robot @@ -103,3 +103,43 @@ | | Run Keyword | vnf_manager.Start All VMs | pinning=${pinning} | | All VPP Interfaces Ready Wait | ${nodes} | retries=${300} | | VPP round robin RX placement on all DUTs | ${nodes} | prefix=Virtual + +| Configure chains of NFs connected via passtrough +| | [Documentation] +| | ... | Start 1..N chains of 1..N QEMU guests (VNFs) with two pci passtrough\ +| | ... | interfaces and interconnecting NF. +| | +| | ... | *Arguments:* +| | ... | - nf_chains - Number of chains of NFs. Type: integer +| | ... | - nf_nodes - Number of NFs nodes per chain. Type: integer +| | ... | - jumbo - Jumbo frames are used (True) or are not used (False) +| | ... | in the test. Type: boolean +| | ... | - perf_qemu_qsz - Virtio Queue Size. Type: integer +| | ... | - use_tuned_cfs - Set True if CFS RR should be used for Qemu SMP. +| | ... | Type: boolean +| | ... | - auto_scale - Whether to use same amount of RXQs for memif interface +| | ... | in containers as vswitch, otherwise use single RXQ. Type: boolean +| | ... | - vnf - Network function as a payload. Type: string +| | ... | - pinning - Whether to pin QEMU VMs to specific cores +| | +| | ... | *Example:* +| | +| | ... | \| Configure chains of VMs connected via passtrough +| | ... | \| 1 \| 1 \| False \| 1024 \| False \| False \| vpp \| True \| +| | +| | [Arguments] | ${nf_chains}=${1} | ${nf_nodes}=${1} | ${jumbo}=${False} +| | ... | ${perf_qemu_qsz}=${1024} | ${use_tuned_cfs}=${False} +| | ... | ${auto_scale}=${True} | ${vnf}=vpp | ${pinning}=${True} +| | +| | Import Library | resources.libraries.python.QemuManager | ${nodes} +| | ... | WITH NAME | vnf_manager +| | Run Keyword | vnf_manager.Construct VMs on all nodes +| | ... | nf_chains=${nf_chains} | nf_nodes=${nf_nodes} | jumbo=${jumbo} +| | ... | perf_qemu_qsz=${perf_qemu_qsz} | use_tuned_cfs=${use_tuned_cfs} +| | ... | auto_scale=${auto_scale} | vnf=${vnf} +| | ... | tg_pf1_mac=${TG_pf1_mac}[0] | tg_pf2_mac=${TG_pf2_mac}[0] +| | ... | vs_dtc=${cpu_count_int} | nf_dtc=${nf_dtc} | nf_dtcr=${nf_dtcr} +| | ... | rxq_count_int=${rxq_count_int} | enable_csum=${False} +| | ... | enable_gso=${False} +| | ... | if1=${DUT1_${int}1}[0] | if2=${DUT1_${int}2}[0] +| | Run Keyword | vnf_manager.Start All VMs | pinning=${pinning} -- 2.16.6