Dpdk in VM: Increase num_mbufs
[csit.git] / resources / libraries / python / QemuUtils.py
index c723878..682def1 100644 (file)
@@ -1,4 +1,4 @@
-# Copyright (c) 2016 Cisco and/or its affiliates.
+# Copyright (c) 2021 Cisco and/or its affiliates.
 # Licensed under the Apache License, Version 2.0 (the "License");
 # you may not use this file except in compliance with the License.
 # You may obtain a copy of the License at:
 
 """QEMU utilities library."""
 
+import json
+
+from re import match
+from string import Template
+from time import sleep
+
 from robot.api import logger
-from ssh import SSH
-from constants import Constants
 
+from resources.libraries.python.Constants import Constants
+from resources.libraries.python.DpdkUtil import DpdkUtil
+from resources.libraries.python.DUTSetup import DUTSetup
+from resources.libraries.python.OptionString import OptionString
+from resources.libraries.python.ssh import exec_cmd, exec_cmd_no_error
+from resources.libraries.python.topology import NodeType, Topology
+from resources.libraries.python.VhostUser import VirtioFeaturesFlags
+from resources.libraries.python.VhostUser import VirtioFeatureMask
+from resources.libraries.python.VppConfigGenerator import VppConfigGenerator
+
+__all__ = [u"QemuUtils"]
 
-class QemuUtils(object):
+
+class QemuUtils:
     """QEMU utilities."""
 
-    @staticmethod
-    def build_qemu(node):
-        """Build QEMU from sources.
+    # Use one instance of class per tests.
+    ROBOT_LIBRARY_SCOPE = u"TEST CASE"
+
+    def __init__(
+            self, node, qemu_id=1, smp=1, mem=512, vnf=None,
+            img=Constants.QEMU_VM_IMAGE):
+        """Initialize QemuUtil class.
 
-        :param node: Node to build QEMU on.
+        :param node: Node to run QEMU on.
+        :param qemu_id: QEMU identifier.
+        :param smp: Number of virtual SMP units (cores).
+        :param mem: Amount of memory.
+        :param vnf: Network function workload.
+        :param img: QEMU disk image or kernel image path.
         :type node: dict
+        :type qemu_id: int
+        :type smp: int
+        :type mem: int
+        :type vnf: str
+        :type img: str
+        """
+        self._nic_id = 0
+        self._node = node
+        self._arch = Topology.get_node_arch(self._node)
+        self._opt = dict()
+
+        # Architecture specific options
+        if self._arch == u"aarch64":
+            self._opt[u"machine_args"] = \
+                u"virt,accel=kvm,usb=off,mem-merge=off,gic-version=3"
+            self._opt[u"console"] = u"ttyAMA0"
+        else:
+            self._opt[u"machine_args"] = u"pc,accel=kvm,usb=off,mem-merge=off"
+            self._opt[u"console"] = u"ttyS0"
+        self._testpmd_path = f"{Constants.QEMU_VM_DPDK}/build/app"
+        self._vm_info = {
+            u"host": node[u"host"],
+            u"type": NodeType.VM,
+            u"port": 10021 + qemu_id,
+            u"serial": 4555 + qemu_id,
+            u"username": 'testuser',
+            u"password": 'Csit1234',
+            u"interfaces": {},
+        }
+        if node[u"port"] != 22:
+            self._vm_info[u"host_port"] = node[u"port"]
+            self._vm_info[u"host_username"] = node[u"username"]
+            self._vm_info[u"host_password"] = node[u"password"]
+        # Input Options.
+        self._opt[u"qemu_id"] = qemu_id
+        self._opt[u"mem"] = int(mem)
+        self._opt[u"smp"] = int(smp)
+        self._opt[u"img"] = img
+        self._opt[u"vnf"] = vnf
+        # Temporary files.
+        self._temp = dict()
+        self._temp[u"log"] = f"/tmp/serial_{qemu_id}.log"
+        self._temp[u"pidfile"] = f"/run/qemu_{qemu_id}.pid"
+        if img == Constants.QEMU_VM_IMAGE:
+            self._temp[u"qmp"] = f"/run/qmp_{qemu_id}.sock"
+            self._temp[u"qga"] = f"/run/qga_{qemu_id}.sock"
+        elif img == Constants.QEMU_VM_KERNEL:
+            self._opt[u"img"], _ = exec_cmd_no_error(
+                node, f"ls -1 {Constants.QEMU_VM_KERNEL}* | tail -1",
+                message=u"Qemu Kernel VM image not found!"
+            )
+            self._temp[u"ini"] = f"/etc/vm_init_{qemu_id}.conf"
+            self._opt[u"initrd"], _ = exec_cmd_no_error(
+                node, f"ls -1 {Constants.QEMU_VM_KERNEL_INITRD}* | tail -1",
+                message=u"Qemu Kernel initrd image not found!"
+            )
+        else:
+            raise RuntimeError(f"QEMU: Unknown VM image option: {img}")
+        # Computed parameters for QEMU command line.
+        self._params = OptionString(prefix=u"-")
+
+    def add_default_params(self):
+        """Set default QEMU command line parameters."""
+        self._params.add(u"daemonize")
+        self._params.add(u"nodefaults")
+        self._params.add_with_value(
+            u"name", f"vnf{self._opt.get(u'qemu_id')},debug-threads=on"
+        )
+        self._params.add(u"no-user-config")
+        self._params.add(u"nographic")
+        self._params.add(u"enable-kvm")
+        self._params.add_with_value(u"pidfile", self._temp.get(u"pidfile"))
+        self._params.add_with_value(u"cpu", u"host")
+
+        self._params.add_with_value(u"machine", self._opt.get(u"machine_args"))
+        self._params.add_with_value(
+            u"smp", f"{self._opt.get(u'smp')},sockets=1,"
+            f"cores={self._opt.get(u'smp')},threads=1"
+        )
+        self._params.add_with_value(
+            u"object", f"memory-backend-file,id=mem,"
+            f"size={self._opt.get(u'mem')}M,mem-path=/dev/hugepages,share=on"
+        )
+        self._params.add_with_value(u"m", f"{self._opt.get(u'mem')}M")
+        self._params.add_with_value(u"numa", u"node,memdev=mem")
+
+    def add_net_user(self, net="10.0.2.0/24"):
+        """Set managment port forwarding."""
+        self._params.add_with_value(
+            u"netdev", f"user,id=mgmt,net={net},"
+            f"hostfwd=tcp::{self._vm_info[u'port']}-:22"
+        )
+        self._params.add_with_value(
+            u"device", f"virtio-net,netdev=mgmt"
+        )
+
+    def add_qmp_qga(self):
+        """Set QMP, QGA management."""
+        self._params.add_with_value(
+            u"chardev", f"socket,path={self._temp.get(u'qga')},"
+            f"server,nowait,id=qga0"
+        )
+        self._params.add_with_value(
+            u"device", u"isa-serial,chardev=qga0"
+        )
+        self._params.add_with_value(
+            u"qmp", f"unix:{self._temp.get(u'qmp')},server,nowait"
+        )
+
+    def add_serial(self):
+        """Set serial to file redirect."""
+        self._params.add_with_value(
+            u"chardev", f"socket,host=127.0.0.1,"
+            f"port={self._vm_info[u'serial']},id=gnc0,server,nowait"
+        )
+        self._params.add_with_value(
+            u"device", u"isa-serial,chardev=gnc0"
+        )
+        self._params.add_with_value(
+            u"serial", f"file:{self._temp.get(u'log')}"
+        )
+
+    def add_drive_cdrom(self, drive_file, index=None):
+        """Set CD-ROM drive.
+
+        :param drive_file: Path to drive image.
+        :param index: Drive index.
+        :type drive_file: str
+        :type index: int
+        """
+        index = f"index={index}," if index else u""
+        self._params.add_with_value(
+            u"drive", f"file={drive_file},{index}media=cdrom"
+        )
+
+    def add_drive(self, drive_file, drive_format):
+        """Set drive with custom format.
+
+        :param drive_file: Path to drive image.
+        :param drive_format: Drive image format.
+        :type drive_file: str
+        :type drive_format: str
+        """
+        self._params.add_with_value(
+            u"drive", f"file={drive_file},format={drive_format},"
+            u"cache=none,if=virtio,file.locking=off"
+        )
+
+    def add_kernelvm_params(self):
+        """Set KernelVM QEMU parameters."""
+        self._params.add_with_value(
+            u"serial", f"file:{self._temp.get(u'log')}"
+        )
+        self._params.add_with_value(
+            u"fsdev", u"local,id=root9p,path=/,security_model=none"
+        )
+        self._params.add_with_value(
+            u"device", u"virtio-9p-pci,fsdev=root9p,mount_tag=virtioroot"
+        )
+        self._params.add_with_value(
+            u"kernel", f"{self._opt.get(u'img')}"
+        )
+        self._params.add_with_value(
+            u"initrd", f"{self._opt.get(u'initrd')}"
+        )
+        self._params.add_with_value(
+            u"append", f"'ro rootfstype=9p rootflags=trans=virtio "
+            f"root=virtioroot console={self._opt.get(u'console')} "
+            f"tsc=reliable hugepages=512 "
+            f"init={self._temp.get(u'ini')} fastboot'"
+        )
+
+    def add_vhost_user_if(
+            self, socket, server=True, jumbo_frames=False, queue_size=None,
+            queues=1, virtio_feature_mask=None):
+        """Add Vhost-user interface.
+
+        :param socket: Path of the unix socket.
+        :param server: If True the socket shall be a listening socket.
+        :param jumbo_frames: Set True if jumbo frames are used in the test.
+        :param queue_size: Vring queue size.
+        :param queues: Number of queues.
+        :param virtio_feature_mask: Mask of virtio features to be enabled.
+        :type socket: str
+        :type server: bool
+        :type jumbo_frames: bool
+        :type queue_size: int
+        :type queues: int
+        :type virtio_feature_mask: int
+        """
+        self._nic_id += 1
+        self._params.add_with_value(
+            u"chardev", f"socket,id=char{self._nic_id},"
+            f"path={socket}{u',server' if server is True else u''}"
+        )
+        self._params.add_with_value(
+            u"netdev", f"vhost-user,id=vhost{self._nic_id},"
+            f"chardev=char{self._nic_id},queues={queues}"
+        )
+        mac = f"52:54:00:00:{self._opt.get(u'qemu_id'):02x}:" \
+            f"{self._nic_id:02x}"
+        queue_size = f"rx_queue_size={queue_size},tx_queue_size={queue_size}" \
+            if queue_size else u""
+        gso = VirtioFeatureMask.is_feature_enabled(
+            virtio_feature_mask, VirtioFeaturesFlags.VIRTIO_NET_F_API_GSO)
+        csum = VirtioFeatureMask.is_feature_enabled(
+            virtio_feature_mask, VirtioFeaturesFlags.VIRTIO_NET_F_API_CSUM)
+
+        self._params.add_with_value(
+            u"device", f"virtio-net-pci,netdev=vhost{self._nic_id},mac={mac},"
+            f"addr={self._nic_id+5}.0,mq=on,vectors={2 * queues + 2},"
+            f"csum={u'on' if csum else u'off'},"
+            f"gso={u'on' if gso else u'off'},"
+            f"guest_tso4={u'on' if gso else u'off'},"
+            f"guest_tso6={u'on' if gso else u'off'},"
+            f"guest_ecn={u'on' if gso else u'off'},"
+            f"{queue_size}"
+        )
+
+        # Add interface MAC and socket to the node dict.
+        if_data = {u"mac_address": mac, u"socket": socket}
+        if_name = f"vhost{self._nic_id}"
+        self._vm_info[u"interfaces"][if_name] = if_data
+        # Add socket to temporary file list.
+        self._temp[if_name] = socket
+
+    def add_vfio_pci_if(self, pci):
+        """Add VFIO PCI interface.
+
+        :param pci: PCI address of interface.
+        :type pci: str
+        """
+        self._nic_id += 1
+        self._params.add_with_value(
+            u"device", f"vfio-pci,host={pci},addr={self._nic_id+5}.0"
+        )
+
+    def create_kernelvm_config_vpp(self, **kwargs):
+        """Create QEMU VPP config files.
+
+        :param kwargs: Key-value pairs to replace content of VPP configuration
+            file.
+        :type kwargs: dict
+        """
+        startup = f"/etc/vpp/vm_startup_{self._opt.get(u'qemu_id')}.conf"
+        running = f"/etc/vpp/vm_running_{self._opt.get(u'qemu_id')}.exec"
+
+        self._temp[u"startup"] = startup
+        self._temp[u"running"] = running
+        self._opt[u"vnf_bin"] = f"/usr/bin/vpp -c {startup}"
+
+        # Create VPP startup configuration.
+        vpp_config = VppConfigGenerator()
+        vpp_config.set_node(self._node)
+        vpp_config.add_unix_nodaemon()
+        vpp_config.add_unix_cli_listen()
+        vpp_config.add_unix_exec(running)
+        vpp_config.add_socksvr()
+        vpp_config.add_main_heap_size(u"512M")
+        vpp_config.add_main_heap_page_size(u"2M")
+        vpp_config.add_statseg_size(u"512M")
+        vpp_config.add_statseg_page_size(u"2M")
+        vpp_config.add_statseg_per_node_counters(u"on")
+        vpp_config.add_buffers_per_numa(107520)
+        vpp_config.add_cpu_main_core(u"0")
+        if self._opt.get(u"smp") > 1:
+            vpp_config.add_cpu_corelist_workers(f"1-{self._opt.get(u'smp')-1}")
+        vpp_config.add_plugin(u"disable", u"default")
+        vpp_config.add_plugin(u"enable", u"ping_plugin.so")
+        if "2vfpt" in self._opt.get(u'vnf'):
+            vpp_config.add_plugin(u"enable", u"avf_plugin.so")
+        if "vhost" in self._opt.get(u'vnf'):
+            vpp_config.add_plugin(u"enable", u"dpdk_plugin.so")
+            vpp_config.add_dpdk_dev(u"0000:00:06.0", u"0000:00:07.0")
+            vpp_config.add_dpdk_dev_default_rxq(kwargs[u"queues"])
+            vpp_config.add_dpdk_log_level(u"debug")
+            if not kwargs[u"jumbo_frames"]:
+                vpp_config.add_dpdk_no_multi_seg()
+                vpp_config.add_dpdk_no_tx_checksum_offload()
+        if "ipsec" in self._opt.get(u'vnf'):
+            vpp_config.add_plugin(u"enable", u"crypto_native_plugin.so")
+            vpp_config.add_plugin(u"enable", u"crypto_ipsecmb_plugin.so")
+            vpp_config.add_plugin(u"enable", u"crypto_openssl_plugin.so")
+        if "nat" in self._opt.get(u'vnf'):
+            vpp_config.add_nat(value=u"endpoint-dependent")
+            vpp_config.add_plugin(u"enable", u"nat_plugin.so")
+        vpp_config.write_config(startup)
+
+        # Create VPP running configuration.
+        template = f"{Constants.RESOURCES_TPL}/vm/{self._opt.get(u'vnf')}.exec"
+        exec_cmd_no_error(self._node, f"rm -f {running}", sudo=True)
+
+        with open(template, u"rt") as src_file:
+            src = Template(src_file.read())
+            exec_cmd_no_error(
+                self._node, f"echo '{src.safe_substitute(**kwargs)}' | "
+                f"sudo tee {running}"
+            )
+
+    def create_kernelvm_config_testpmd_io(self, **kwargs):
+        """Create QEMU testpmd-io command line.
+
+        :param kwargs: Key-value pairs to construct command line parameters.
+        :type kwargs: dict
+        """
+        pmd_max_pkt_len = u"9200" if kwargs[u"jumbo_frames"] else u"1518"
+        testpmd_cmd = DpdkUtil.get_testpmd_cmdline(
+            eal_corelist=f"0-{self._opt.get(u'smp') - 1}",
+            eal_driver=False,
+            eal_pci_whitelist0=u"0000:00:06.0",
+            eal_pci_whitelist1=u"0000:00:07.0",
+            eal_in_memory=True,
+            pmd_num_mbufs=32768,
+            pmd_fwd_mode=u"io",
+            pmd_nb_ports=u"2",
+            pmd_portmask=u"0x3",
+            pmd_max_pkt_len=pmd_max_pkt_len,
+            pmd_mbuf_size=u"16384",
+            pmd_rxq=kwargs[u"queues"],
+            pmd_txq=kwargs[u"queues"],
+            pmd_tx_offloads='0x0',
+            pmd_nb_cores=str(self._opt.get(u"smp") - 1)
+        )
+
+        self._opt[u"vnf_bin"] = f"{self._testpmd_path}/{testpmd_cmd}"
+
+    def create_kernelvm_config_testpmd_mac(self, **kwargs):
+        """Create QEMU testpmd-mac command line.
+
+        :param kwargs: Key-value pairs to construct command line parameters.
+        :type kwargs: dict
+        """
+        pmd_max_pkt_len = u"9200" if kwargs[u"jumbo_frames"] else u"1518"
+        testpmd_cmd = DpdkUtil.get_testpmd_cmdline(
+            eal_corelist=f"0-{self._opt.get(u'smp') - 1}",
+            eal_driver=False,
+            eal_pci_whitelist0=u"0000:00:06.0",
+            eal_pci_whitelist1=u"0000:00:07.0",
+            eal_in_memory=True,
+            pmd_num_mbufs=32768,
+            pmd_fwd_mode=u"mac",
+            pmd_nb_ports=u"2",
+            pmd_portmask=u"0x3",
+            pmd_max_pkt_len=pmd_max_pkt_len,
+            pmd_mbuf_size=u"16384",
+            pmd_eth_peer_0=f"0,{kwargs[u'vif1_mac']}",
+            pmd_eth_peer_1=f"1,{kwargs[u'vif2_mac']}",
+            pmd_rxq=kwargs[u"queues"],
+            pmd_txq=kwargs[u"queues"],
+            pmd_tx_offloads=u"0x0",
+            pmd_nb_cores=str(self._opt.get(u"smp") - 1)
+        )
+
+        self._opt[u"vnf_bin"] = f"{self._testpmd_path}/{testpmd_cmd}"
+
+    def create_kernelvm_config_iperf3(self):
+        """Create QEMU iperf3 command line."""
+        self._opt[u"vnf_bin"] = f"mkdir /run/sshd; /usr/sbin/sshd -D -d"
+
+    def create_kernelvm_init(self, **kwargs):
+        """Create QEMU init script.
+
+        :param kwargs: Key-value pairs to replace content of init startup file.
+        :type kwargs: dict
+        """
+        init = self._temp.get(u"ini")
+        exec_cmd_no_error(self._node, f"rm -f {init}", sudo=True)
+
+        with open(kwargs[u"template"], u"rt") as src_file:
+            src = Template(src_file.read())
+            exec_cmd_no_error(
+                self._node, f"echo '{src.safe_substitute(**kwargs)}' | "
+                f"sudo tee {init}"
+            )
+            exec_cmd_no_error(self._node, f"chmod +x {init}", sudo=True)
+
+    def configure_kernelvm_vnf(self, **kwargs):
+        """Create KernelVM VNF configurations.
+
+        :param kwargs: Key-value pairs for templating configs.
+        :type kwargs: dict
+        """
+        if u"vpp" in self._opt.get(u"vnf"):
+            self.create_kernelvm_config_vpp(**kwargs)
+            self.create_kernelvm_init(
+                template=f"{Constants.RESOURCES_TPL}/vm/init.sh",
+                vnf_bin=self._opt.get(u"vnf_bin")
+            )
+        elif u"testpmd_io" in self._opt.get(u"vnf"):
+            self.create_kernelvm_config_testpmd_io(**kwargs)
+            self.create_kernelvm_init(
+                template=f"{Constants.RESOURCES_TPL}/vm/init.sh",
+                vnf_bin=self._opt.get(u"vnf_bin")
+            )
+        elif u"testpmd_mac" in self._opt.get(u"vnf"):
+            self.create_kernelvm_config_testpmd_mac(**kwargs)
+            self.create_kernelvm_init(
+                template=f"{Constants.RESOURCES_TPL}/vm/init.sh",
+                vnf_bin=self._opt.get(u"vnf_bin")
+            )
+        elif u"iperf3" in self._opt.get(u"vnf"):
+            qemu_id = self._opt.get(u'qemu_id') % 2
+            self.create_kernelvm_config_iperf3()
+            self.create_kernelvm_init(
+                template=f"{Constants.RESOURCES_TPL}/vm/init_iperf3.sh",
+                vnf_bin=self._opt.get(u"vnf_bin"),
+                ip_address_l=u"2.2.2.2/30" if qemu_id else u"1.1.1.1/30",
+                ip_address_r=u"2.2.2.1" if qemu_id else u"1.1.1.2",
+                ip_route_r=u"1.1.1.0/30" if qemu_id else u"2.2.2.0/30"
+            )
+        else:
+            raise RuntimeError(u"QEMU: Unsupported VNF!")
+
+    def get_qemu_pids(self):
+        """Get QEMU CPU pids.
+
+        :returns: List of QEMU CPU pids.
+        :rtype: list of str
+        """
+        command = f"grep -rwl 'CPU' /proc/$(sudo cat " \
+            f"{self._temp.get(u'pidfile')})/task/*/comm "
+        command += r"| xargs dirname | sed -e 's/\/.*\///g' | uniq"
+
+        stdout, _ = exec_cmd_no_error(self._node, command)
+        return stdout.splitlines()
+
+    def qemu_set_affinity(self, *host_cpus):
+        """Set qemu affinity by getting thread PIDs via QMP and taskset to list
+        of CPU cores. Function tries to execute 3 times to avoid race condition
+        in getting thread PIDs.
+
+        :param host_cpus: List of CPU cores.
+        :type host_cpus: list
+        """
+        for _ in range(3):
+            try:
+                qemu_cpus = self.get_qemu_pids()
+
+                if len(qemu_cpus) != len(host_cpus):
+                    sleep(1)
+                    continue
+                for qemu_cpu, host_cpu in zip(qemu_cpus, host_cpus):
+                    command = f"taskset -pc {host_cpu} {qemu_cpu}"
+                    message = f"QEMU: Set affinity failed " \
+                        f"on {self._node[u'host']}!"
+                    exec_cmd_no_error(
+                        self._node, command, sudo=True, message=message
+                    )
+                break
+            except (RuntimeError, ValueError):
+                self.qemu_kill_all()
+                raise
+        else:
+            self.qemu_kill_all()
+            raise RuntimeError(u"Failed to set Qemu threads affinity!")
+
+    def qemu_set_scheduler_policy(self):
+        """Set scheduler policy to SCHED_RR with priority 1 for all Qemu CPU
+        processes.
+
+        :raises RuntimeError: Set scheduler policy failed.
+        """
+        try:
+            qemu_cpus = self.get_qemu_pids()
+
+            for qemu_cpu in qemu_cpus:
+                command = f"chrt -r -p 1 {qemu_cpu}"
+                message = f"QEMU: Set SCHED_RR failed on {self._node[u'host']}"
+                exec_cmd_no_error(
+                    self._node, command, sudo=True, message=message
+                )
+        except (RuntimeError, ValueError):
+            self.qemu_kill_all()
+            raise
+
+    def _qemu_qmp_exec(self, cmd):
+        """Execute QMP command.
+
+        QMP is JSON based protocol which allows to control QEMU instance.
+
+        :param cmd: QMP command to execute.
+        :type cmd: str
+        :returns: Command output in python representation of JSON format. The
+            { "return": {} } response is QMP's success response. An error
+            response will contain the "error" keyword instead of "return".
+        """
+        # To enter command mode, the qmp_capabilities command must be issued.
+        command = f"echo \"{{{{ \\\"execute\\\": " \
+            f"\\\"qmp_capabilities\\\" }}}}" \
+            f"{{{{ \\\"execute\\\": \\\"{cmd}\\\" }}}}\" | " \
+            f"sudo -S socat - UNIX-CONNECT:{self._temp.get(u'qmp')}"
+        message = f"QMP execute '{cmd}' failed on {self._node[u'host']}"
+
+        stdout, _ = exec_cmd_no_error(
+            self._node, command, sudo=False, message=message
+        )
+
+        # Skip capabilities negotiation messages.
+        out_list = stdout.splitlines()
+        if len(out_list) < 3:
+            raise RuntimeError(f"Invalid QMP output on {self._node[u'host']}")
+        return json.loads(out_list[2])
+
+    def _qemu_qga_flush(self):
+        """Flush the QGA parser state."""
+        command = f"(printf \"\xFF\"; sleep 1) | sudo -S socat " \
+            f"- UNIX-CONNECT:{self._temp.get(u'qga')}"
+        message = f"QGA flush failed on {self._node[u'host']}"
+        stdout, _ = exec_cmd_no_error(
+            self._node, command, sudo=False, message=message
+        )
+
+        return json.loads(stdout.split(u"\n", 1)[0]) if stdout else dict()
+
+    def _qemu_qga_exec(self, cmd):
+        """Execute QGA command.
+
+        QGA provide access to a system-level agent via standard QMP commands.
+
+        :param cmd: QGA command to execute.
+        :type cmd: str
+        """
+        command = f"(echo \"{{{{ \\\"execute\\\": " \
+            f"\\\"{cmd}\\\" }}}}\"; sleep 1) | " \
+            f"sudo -S socat - UNIX-CONNECT:{self._temp.get(u'qga')}"
+        message = f"QGA execute '{cmd}' failed on {self._node[u'host']}"
+        stdout, _ = exec_cmd_no_error(
+            self._node, command, sudo=False, message=message
+        )
+
+        return json.loads(stdout.split(u"\n", 1)[0]) if stdout else dict()
+
+    def _wait_until_vm_boot(self):
+        """Wait until QEMU VM is booted."""
+        try:
+            getattr(self, f'_wait_{self._opt["vnf"]}')()
+        except AttributeError:
+            self._wait_default()
+
+    def _wait_default(self, retries=60):
+        """Wait until QEMU with VPP is booted.
+
+        :param retries: Number of retries.
+        :type retries: int
+        """
+        for _ in range(retries):
+            command = f"tail -1 {self._temp.get(u'log')}"
+            stdout = None
+            try:
+                stdout, _ = exec_cmd_no_error(self._node, command, sudo=True)
+                sleep(1)
+            except RuntimeError:
+                pass
+            if "vpp " in stdout and "built by" in stdout:
+                break
+            if u"Press enter to exit" in stdout:
+                break
+            if u"reboot: Power down" in stdout:
+                raise RuntimeError(
+                    f"QEMU: NF failed to run on {self._node[u'host']}!"
+                )
+        else:
+            raise RuntimeError(
+                f"QEMU: Timeout, VM not booted on {self._node[u'host']}!"
+            )
+
+    def _wait_nestedvm(self, retries=12):
+        """Wait until QEMU with NestedVM is booted.
+
+        First try to flush qga until there is output.
+        Then ping QEMU guest agent each 5s until VM booted or timeout.
+
+        :param retries: Number of retries with 5s between trials.
+        :type retries: int
+        """
+        for _ in range(retries):
+            out = None
+            try:
+                out = self._qemu_qga_flush()
+            except ValueError:
+                logger.trace(f"QGA qga flush unexpected output {out}")
+            # Empty output - VM not booted yet
+            if not out:
+                sleep(5)
+            else:
+                break
+        else:
+            raise RuntimeError(
+                f"QEMU: Timeout, VM not booted on {self._node[u'host']}!"
+            )
+        for _ in range(retries):
+            out = None
+            try:
+                out = self._qemu_qga_exec(u"guest-ping")
+            except ValueError:
+                logger.trace(f"QGA guest-ping unexpected output {out}")
+            # Empty output - VM not booted yet.
+            if not out:
+                sleep(5)
+            # Non-error return - VM booted.
+            elif out.get(u"return") is not None:
+                break
+            # Skip error and wait.
+            elif out.get(u"error") is not None:
+                sleep(5)
+            else:
+                # If there is an unexpected output from QGA guest-info, try
+                # again until timeout.
+                logger.trace(f"QGA guest-ping unexpected output {out}")
+        else:
+            raise RuntimeError(
+                f"QEMU: Timeout, VM not booted on {self._node[u'host']}!"
+            )
+
+    def _wait_iperf3(self, retries=60):
+        """Wait until QEMU with iPerf3 is booted.
+
+        :param retries: Number of retries.
+        :type retries: int
+        """
+        grep = u"Server listening on 0.0.0.0 port 22."
+        cmd = f"fgrep '{grep}' {self._temp.get(u'log')}"
+        message = f"QEMU: Timeout, VM not booted on {self._node[u'host']}!"
+        exec_cmd_no_error(
+            self._node, cmd=cmd, sudo=True, message=message, retries=retries,
+            include_reason=True
+        )
+
+    def _update_vm_interfaces(self):
+        """Update interface names in VM node dict."""
+        # Send guest-network-get-interfaces command via QGA, output example:
+        # {"return": [{"name": "eth0", "hardware-address": "52:54:00:00:04:01"},
+        # {"name": "eth1", "hardware-address": "52:54:00:00:04:02"}]}.
+        out = self._qemu_qga_exec(u"guest-network-get-interfaces")
+        interfaces = out.get(u"return")
+        mac_name = {}
+        if not interfaces:
+            raise RuntimeError(
+                f"Get VM interface list failed on {self._node[u'host']}"
+            )
+        # Create MAC-name dict.
+        for interface in interfaces:
+            if u"hardware-address" not in interface:
+                continue
+            mac_name[interface[u"hardware-address"]] = interface[u"name"]
+        # Match interface by MAC and save interface name.
+        for interface in self._vm_info[u"interfaces"].values():
+            mac = interface.get(u"mac_address")
+            if_name = mac_name.get(mac)
+            if if_name is None:
+                logger.trace(f"Interface name for MAC {mac} not found")
+            else:
+                interface[u"name"] = if_name
+
+    def qemu_start(self):
+        """Start QEMU and wait until VM boot.
+
+        :returns: VM node info.
+        :rtype: dict
         """
-        ssh = SSH()
-        ssh.connect(node)
+        cmd_opts = OptionString()
+        cmd_opts.add(f"{Constants.QEMU_BIN_PATH}/qemu-system-{self._arch}")
+        cmd_opts.extend(self._params)
+        message = f"QEMU: Start failed on {self._node[u'host']}!"
+        try:
+            DUTSetup.check_huge_page(
+                self._node, u"/dev/hugepages", int(self._opt.get(u"mem")))
+
+            exec_cmd_no_error(
+                self._node, cmd_opts, timeout=300, sudo=True, message=message
+            )
+            self._wait_until_vm_boot()
+        except RuntimeError:
+            self.qemu_kill_all()
+            raise
+        return self._vm_info
 
-        (ret_code, stdout, stderr) = \
-            ssh.exec_command('sudo -Sn bash {0}/{1}/qemu_build.sh'.format(
-                Constants.REMOTE_FW_DIR, Constants.RESOURCES_LIB_SH), 1000)
-        logger.trace(stdout)
-        if 0 != int(ret_code):
-            logger.debug('QEMU build failed {0}'.format(stderr))
-            raise RuntimeError('QEMU build failed on {0}'.format(node['host']))
+    def qemu_kill(self):
+        """Kill qemu process."""
+        exec_cmd(
+            self._node, f"chmod +r {self._temp.get(u'pidfile')}", sudo=True
+        )
+        exec_cmd(
+            self._node, f"kill -SIGKILL $(cat {self._temp.get(u'pidfile')})",
+            sudo=True
+        )
+
+        for value in self._temp.values():
+            exec_cmd(self._node, f"cat {value}", sudo=True)
+            exec_cmd(self._node, f"rm -f {value}", sudo=True)
+
+    def qemu_kill_all(self):
+        """Kill all qemu processes on DUT node if specified."""
+        exec_cmd(self._node, u"pkill -SIGKILL qemu", sudo=True)
+
+        for value in self._temp.values():
+            exec_cmd(self._node, f"cat {value}", sudo=True)
+            exec_cmd(self._node, f"rm -f {value}", sudo=True)
+
+    def qemu_version(self):
+        """Return Qemu version.
+
+        :returns: Qemu version.
+        :rtype: str
+        """
+        command = f"{Constants.QEMU_BIN_PATH}/qemu-system-{self._arch} " \
+            f"--version"
+        try:
+            stdout, _ = exec_cmd_no_error(self._node, command, sudo=True)
+            return match(r"QEMU emulator version ([\d.]*)", stdout).group(1)
+        except RuntimeError:
+            self.qemu_kill_all()
+            raise