X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=resources%2Flibraries%2Fpython%2FQemuManager.py;h=ce55be0a51356f3626e90fc9efe5c3b77f555d70;hb=79f5ba9bf7656972dd988508eff9465562dde42c;hp=e81f0a84ceb7e00ca9d18e87b2a1700185384bfd;hpb=33bd6f2425ddff8e4167d36ea61b64ac81da3a65;p=csit.git diff --git a/resources/libraries/python/QemuManager.py b/resources/libraries/python/QemuManager.py index e81f0a84ce..ce55be0a51 100644 --- a/resources/libraries/python/QemuManager.py +++ b/resources/libraries/python/QemuManager.py @@ -1,4 +1,4 @@ -# Copyright (c) 2019 Cisco and/or its affiliates. +# Copyright (c) 2020 Cisco and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: @@ -15,64 +15,19 @@ from collections import OrderedDict -from robot.libraries.BuiltIn import BuiltIn - from resources.libraries.python.Constants import Constants from resources.libraries.python.CpuUtils import CpuUtils from resources.libraries.python.QemuUtils import QemuUtils from resources.libraries.python.topology import NodeType, Topology -__all__ = ["QemuManager"] - - -def get_affinity_vm(nodes, node, nf_chains=1, nf_nodes=1, nf_chain=1, nf_node=1, - cpu_count_int=1): - """Get affinity of VM. Result will be used to compute the amount of - CPUs and also affinity. - - :param node: SUT nodes. - :param node: DUT node. - :param nf_chains: Number of NF chains. - :param nf_nodes: Number of NF nodes in chain. - :param nf_chain: Chain ID. - :param nf_node: Node ID. - :param cpu_count_int: Amount of Dataplane threads of vswitch. - :type nodes: dict - :type node: dict - :type nf_chains: int - :type nf_nodes: int - :type nf_chain: int - :type nf_node: int - :type cpu_count_int: int - :returns: List of CPUs allocated to VM. - :rtype: list - """ - sut_sc = 1 - dut_mc = 1 - dut_dc = cpu_count_int - skip_cnt = sut_sc + dut_mc + dut_dc - dtc = cpu_count_int - - interface_list = [] - interface_list.append( - BuiltIn().get_variable_value('${{{node}_if1}}'.format(node=node))) - interface_list.append( - BuiltIn().get_variable_value('${{{node}_if2}}'.format(node=node))) - - cpu_node = Topology.get_interfaces_numa_node(nodes[node], *interface_list) - - nf_cpus = CpuUtils.cpu_slice_of_list_for_nf( - node=nodes[node], cpu_node=cpu_node, chains=nf_chains, - nodeness=nf_nodes, chain_id=nf_chain, node_id=nf_node, mtcr=2, dtcr=1, - dtc=dtc, skip_cnt=skip_cnt) - - return nf_cpus - -class QemuManager(object): +__all__ = [u"QemuManager"] + + +class QemuManager: """QEMU lifecycle management class""" # Use one instance of class per tests. - ROBOT_LIBRARY_SCOPE = 'TEST CASE' + ROBOT_LIBRARY_SCOPE = u"TEST CASE" def __init__(self, nodes): """Init QemuManager object.""" @@ -91,45 +46,64 @@ class QemuManager(object): :param kwargs: Named parameters. :type kwargs: dict """ - node = kwargs['node'] - nf_chains = int(kwargs['nf_chains']) - nf_nodes = int(kwargs['nf_nodes']) - queues = kwargs['rxq_count_int'] if kwargs['auto_scale'] else 1 - cpu_count_int = kwargs['cpu_count_int'] if kwargs['auto_scale'] else 1 - - img = Constants.QEMU_PERF_VM_KERNEL + node = kwargs[u"node"] + nf_chains = int(kwargs[u"nf_chains"]) + nf_nodes = int(kwargs[u"nf_nodes"]) + queues = kwargs[u"rxq_count_int"] if kwargs[u"auto_scale"] else 1 + vs_dtc = kwargs[u"vs_dtc"] + nf_dtc = kwargs[u"vs_dtc"] if kwargs[u"auto_scale"] \ + else kwargs[u"nf_dtc"] + nf_dtcr = kwargs[u"nf_dtcr"] \ + if isinstance(kwargs[u"nf_dtcr"], int) else 2 + + img = Constants.QEMU_VM_KERNEL for nf_chain in range(1, nf_chains + 1): for nf_node in range(1, nf_nodes + 1): qemu_id = (nf_chain - 1) * nf_nodes + nf_node - name = '{node}_{qemu_id}'.format(node=node, qemu_id=qemu_id) - sock1 = '/var/run/vpp/sock-{qemu_id}-1'.format(qemu_id=qemu_id) - sock2 = '/var/run/vpp/sock-{qemu_id}-2'.format(qemu_id=qemu_id) - vif1_mac = kwargs['tg_if1_mac'] if nf_node == 1 \ - else '52:54:00:00:{id:02x}:02'.format(id=qemu_id - 1) - vif2_mac = kwargs['tg_if2_mac'] if nf_node == nf_nodes \ - else '52:54:00:00:{id:02x}:01'.format(id=qemu_id + 1) - - self.machines_affinity[name] = get_affinity_vm( + name = f"{node}_{qemu_id}" + sock1 = f"/run/vpp/sock-{qemu_id}-1" + sock2 = f"/run/vpp/sock-{qemu_id}-2" + idx1 = (nf_chain - 1) * nf_nodes * 2 + nf_node * 2 - 1 + vif1_mac = Topology.get_interface_mac( + self.nodes[node], f"vhost{idx1}" + ) if kwargs[u"vnf"] == u"testpmd_mac" \ + else kwargs[u"tg_pf1_mac"] if nf_node == 1 \ + else f"52:54:00:00:{(qemu_id - 1):02x}:02" + idx2 = (nf_chain - 1) * nf_nodes * 2 + nf_node * 2 + vif2_mac = Topology.get_interface_mac( + self.nodes[node], f"vhost{idx2}" + ) if kwargs[u"vnf"] == u"testpmd_mac" \ + else kwargs[u"tg_pf2_mac"] if nf_node == nf_nodes \ + else f"52:54:00:00:{(qemu_id + 1):02x}:01" + + self.machines_affinity[name] = CpuUtils.get_affinity_nf( nodes=self.nodes, node=node, nf_chains=nf_chains, nf_nodes=nf_nodes, nf_chain=nf_chain, nf_node=nf_node, - cpu_count_int=cpu_count_int) + vs_dtc=vs_dtc, nf_dtc=nf_dtc, nf_dtcr=nf_dtcr + ) self.machines[name] = QemuUtils( node=self.nodes[node], qemu_id=qemu_id, smp=len(self.machines_affinity[name]), mem=4096, - vnf=kwargs['vnf'], img=img) + vnf=kwargs[u"vnf"], img=img + ) self.machines[name].configure_kernelvm_vnf( - mac1='52:54:00:00:{id:02x}:01'.format(id=qemu_id), - mac2='52:54:00:00:{id:02x}:02'.format(id=qemu_id), - vif1_mac=vif1_mac, - vif2_mac=vif2_mac) + mac1=f"52:54:00:00:{qemu_id:02x}:01", + mac2=f"52:54:00:00:{qemu_id:02x}:02", + vif1_mac=vif1_mac, vif2_mac=vif2_mac, queues=queues, + jumbo_frames=kwargs[u"jumbo"] + ) self.machines[name].qemu_add_vhost_user_if( - sock1, jumbo_frames=kwargs['jumbo'], queues=queues, - queue_size=1024) + sock1, jumbo_frames=kwargs[u"jumbo"], queues=queues, + queue_size=kwargs[u"perf_qemu_qsz"], + csum=kwargs[u"enable_csum"], gso=kwargs[u"enable_gso"] + ) self.machines[name].qemu_add_vhost_user_if( - sock2, jumbo_frames=kwargs['jumbo'], queues=queues, - queue_size=1024) + sock2, jumbo_frames=kwargs[u"jumbo"], queues=queues, + queue_size=kwargs[u"perf_qemu_qsz"], + csum=kwargs[u"enable_csum"], gso=kwargs[u"enable_gso"] + ) def construct_vms_on_all_nodes(self, **kwargs): """Construct 1..Mx1..N VMs(s) with specified name on all nodes. @@ -139,7 +113,7 @@ class QemuManager(object): """ self.initialize() for node in self.nodes: - if self.nodes[node]['type'] == NodeType.DUT: + if self.nodes[node][u"type"] == NodeType.DUT: self.construct_vms_on_node(node=node, **kwargs) def start_all_vms(self, pinning=False): @@ -148,17 +122,12 @@ class QemuManager(object): :param pinning: If True, then do also QEMU process pinning. :type pinning: bool """ - for machine, machine_affinity in zip(self.machines.values(), - self.machines_affinity.values()): + for machine, machine_affinity in \ + zip(self.machines.values(), self.machines_affinity.values()): machine.qemu_start() if pinning: machine.qemu_set_affinity(*machine_affinity) - def set_scheduler_all_vms(self): - """Set CFS scheduler policy on all VMs in manager.""" - for machine in self.machines.values(): - machine.qemu_set_scheduler_policy() - def kill_all_vms(self, force=False): """Kill all added VMs in manager.