X-Git-Url: https://gerrit.fd.io/r/gitweb?p=csit.git;a=blobdiff_plain;f=resources%2Flibraries%2Fpython%2FInterfaceUtil.py;h=94c78a1bef9379d80355451033b7261c393c4584;hp=1a97130a8e6c0d65ace81db21816672eaa2ab064;hb=75fe508bc9e03291977c1a28ca0adc31c3149df9;hpb=669d320bc64e41f879c047a39a66b8ec99bf4b1e diff --git a/resources/libraries/python/InterfaceUtil.py b/resources/libraries/python/InterfaceUtil.py index 1a97130a8e..94c78a1bef 100644 --- a/resources/libraries/python/InterfaceUtil.py +++ b/resources/libraries/python/InterfaceUtil.py @@ -1,4 +1,4 @@ -# Copyright (c) 2019 Cisco and/or its affiliates. +# Copyright (c) 2021 Cisco and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: @@ -14,42 +14,112 @@ """Interface util library.""" from time import sleep - from enum import IntEnum + from ipaddress import ip_address from robot.api import logger from resources.libraries.python.Constants import Constants from resources.libraries.python.CpuUtils import CpuUtils from resources.libraries.python.DUTSetup import DUTSetup +from resources.libraries.python.IPAddress import IPAddress from resources.libraries.python.L2Util import L2Util from resources.libraries.python.PapiExecutor import PapiSocketExecutor from resources.libraries.python.parsers.JsonParser import JsonParser -from resources.libraries.python.ssh import SSH, exec_cmd_no_error +from resources.libraries.python.ssh import SSH, exec_cmd, exec_cmd_no_error from resources.libraries.python.topology import NodeType, Topology from resources.libraries.python.VPPUtil import VPPUtil -class LinkBondLoadBalance(IntEnum): - """Link bonding load balance.""" - L2 = 0 # pylint: disable=invalid-name - L34 = 1 - L23 = 2 +class InterfaceStatusFlags(IntEnum): + """Interface status flags.""" + IF_STATUS_API_FLAG_ADMIN_UP = 1 + IF_STATUS_API_FLAG_LINK_UP = 2 + + +class MtuProto(IntEnum): + """MTU protocol.""" + MTU_PROTO_API_L3 = 0 + MTU_PROTO_API_IP4 = 1 + MTU_PROTO_API_IP6 = 2 + MTU_PROTO_API_MPLS = 3 + MTU_PROTO_API_N = 4 + + +class LinkDuplex(IntEnum): + """Link duplex""" + LINK_DUPLEX_API_UNKNOWN = 0 + LINK_DUPLEX_API_HALF = 1 + LINK_DUPLEX_API_FULL = 2 + + +class SubInterfaceFlags(IntEnum): + """Sub-interface flags.""" + SUB_IF_API_FLAG_NO_TAGS = 1 + SUB_IF_API_FLAG_ONE_TAG = 2 + SUB_IF_API_FLAG_TWO_TAGS = 4 + SUB_IF_API_FLAG_DOT1AD = 8 + SUB_IF_API_FLAG_EXACT_MATCH = 16 + SUB_IF_API_FLAG_DEFAULT = 32 + SUB_IF_API_FLAG_OUTER_VLAN_ID_ANY = 64 + SUB_IF_API_FLAG_INNER_VLAN_ID_ANY = 128 + SUB_IF_API_FLAG_DOT1AH = 256 + + +class RxMode(IntEnum): + """RX mode""" + RX_MODE_API_UNKNOWN = 0 + RX_MODE_API_POLLING = 1 + RX_MODE_API_INTERRUPT = 2 + RX_MODE_API_ADAPTIVE = 3 + RX_MODE_API_DEFAULT = 4 + + +class IfType(IntEnum): + """Interface type""" + # A hw interface + IF_API_TYPE_HARDWARE = 0 + # A sub-interface + IF_API_TYPE_SUB = 1 + IF_API_TYPE_P2P = 2 + IF_API_TYPE_PIPE = 3 + + +class LinkBondLoadBalanceAlgo(IntEnum): + """Link bonding load balance algorithm.""" + BOND_API_LB_ALGO_L2 = 0 + BOND_API_LB_ALGO_L34 = 1 + BOND_API_LB_ALGO_L23 = 2 + BOND_API_LB_ALGO_RR = 3 + BOND_API_LB_ALGO_BC = 4 + BOND_API_LB_ALGO_AB = 5 class LinkBondMode(IntEnum): - """Link bonding load balance.""" - ROUND_ROBIN = 1 - ACTIVE_BACKUP = 2 - XOR = 3 - BROADCAST = 4 - LACP = 5 + """Link bonding mode.""" + BOND_API_MODE_ROUND_ROBIN = 1 + BOND_API_MODE_ACTIVE_BACKUP = 2 + BOND_API_MODE_XOR = 3 + BOND_API_MODE_BROADCAST = 4 + BOND_API_MODE_LACP = 5 -class InterfaceUtil(object): - """General utilities for managing interfaces""" +class RdmaMode(IntEnum): + """RDMA interface mode.""" + RDMA_API_MODE_AUTO = 0 + RDMA_API_MODE_IBV = 1 + RDMA_API_MODE_DV = 2 + - __UDEV_IF_RULES_FILE = '/etc/udev/rules.d/10-network.rules' +class AfXdpMode(IntEnum): + """AF_XDP interface mode.""" + AF_XDP_API_MODE_AUTO = 0 + AF_XDP_API_MODE_COPY = 1 + AF_XDP_API_MODE_ZERO_COPY = 2 + + +class InterfaceUtil: + """General utilities for managing interfaces""" @staticmethod def pci_to_int(pci_str): @@ -61,12 +131,31 @@ class InterfaceUtil(object): :returns: Integer representation of PCI address. :rtype: int """ - pci = list(pci_str.split(':')[0:2]) - pci.extend(pci_str.split(':')[2].split('.')) + pci = list(pci_str.split(u":")[0:2]) + pci.extend(pci_str.split(u":")[2].split(u".")) return (int(pci[0], 16) | int(pci[1], 16) << 16 | int(pci[2], 16) << 24 | int(pci[3], 16) << 29) + @staticmethod + def pci_to_eth(node, pci_str): + """Convert PCI address on DUT to Linux ethernet name. + + :param node: DUT node + :param pci_str: PCI address. + :type node: dict + :type pci_str: str + :returns: Ethernet name. + :rtype: str + """ + cmd = f"basename /sys/bus/pci/devices/{pci_str}/net/*" + try: + stdout, _ = exec_cmd_no_error(node, cmd) + except RuntimeError: + raise RuntimeError(f"Cannot convert {pci_str} to ethernet name!") + + return stdout.strip() + @staticmethod def get_interface_index(node, interface): """Get interface sw_if_index from topology file. @@ -86,13 +175,12 @@ class InterfaceUtil(object): sw_if_index = \ Topology.get_interface_sw_index_by_name(node, interface) except TypeError as err: - raise TypeError('Wrong interface format {ifc}: {err}'.format( - ifc=interface, err=err.message)) + raise TypeError(f"Wrong interface format {interface}") from err return sw_if_index @staticmethod - def set_interface_state(node, interface, state, if_type='key'): + def set_interface_state(node, interface, state, if_type=u"key"): """Set interface state on a node. Function can be used for DUTs as well as for TGs. @@ -110,82 +198,135 @@ class InterfaceUtil(object): :raises ValueError: If the state of interface is unexpected. :raises ValueError: If the node has an unknown node type. """ - if if_type == 'key': - if isinstance(interface, basestring): + if if_type == u"key": + if isinstance(interface, str): sw_if_index = Topology.get_interface_sw_index(node, interface) iface_name = Topology.get_interface_name(node, interface) else: sw_if_index = interface - elif if_type == 'name': + elif if_type == u"name": iface_key = Topology.get_interface_by_name(node, interface) if iface_key is not None: sw_if_index = Topology.get_interface_sw_index(node, iface_key) iface_name = interface else: - raise ValueError('Unknown if_type: {type}'.format(type=if_type)) + raise ValueError(f"Unknown if_type: {if_type}") - if node['type'] == NodeType.DUT: - if state == 'up': - admin_up_down = 1 - elif state == 'down': - admin_up_down = 0 + if node[u"type"] == NodeType.DUT: + if state == u"up": + flags = InterfaceStatusFlags.IF_STATUS_API_FLAG_ADMIN_UP.value + elif state == u"down": + flags = 0 else: - raise ValueError('Unexpected interface state: {state}'.format( - state=state)) - cmd = 'sw_interface_set_flags' - err_msg = 'Failed to set interface state on host {host}'.format( - host=node['host']) - args = dict(sw_if_index=sw_if_index, - admin_up_down=admin_up_down) + raise ValueError(f"Unexpected interface state: {state}") + cmd = u"sw_interface_set_flags" + err_msg = f"Failed to set interface state on host {node[u'host']}" + args = dict( + sw_if_index=int(sw_if_index), + flags=flags + ) with PapiSocketExecutor(node) as papi_exec: papi_exec.add(cmd, **args).get_reply(err_msg) - elif node['type'] == NodeType.TG or node['type'] == NodeType.VM: - cmd = 'ip link set {ifc} {state}'.format( - ifc=iface_name, state=state) + elif node[u"type"] == NodeType.TG or node[u"type"] == NodeType.VM: + cmd = f"ip link set {iface_name} {state}" exec_cmd_no_error(node, cmd, sudo=True) else: - raise ValueError('Node {} has unknown NodeType: "{}"' - .format(node['host'], node['type'])) + raise ValueError( + f"Node {node[u'host']} has unknown NodeType: {node[u'type']}" + ) @staticmethod - def set_interface_ethernet_mtu(node, iface_key, mtu): - """Set Ethernet MTU for specified interface. + def set_interface_state_pci( + node, pf_pcis, namespace=None, state=u"up"): + """Set operational state for interface specified by PCI address. + + :param node: Topology node. + :param pf_pcis: List of node's interfaces PCI addresses. + :param namespace: Exec command in namespace. (Optional, Default: none) + :param state: Up/Down. (Optional, default: up) + :type nodes: dict + :type pf_pcis: list + :type namespace: str + :type state: str + """ + for pf_pci in pf_pcis: + pf_eth = InterfaceUtil.pci_to_eth(node, pf_pci) + InterfaceUtil.set_linux_interface_state( + node, pf_eth, namespace=namespace, state=state + ) - Function can be used only for TGs. + @staticmethod + def set_interface_mtu(node, pf_pcis, mtu=9200): + """Set Ethernet MTU for specified interfaces. - :param node: Node where the interface is. - :param iface_key: Interface key from topology file. - :param mtu: MTU to set. - :type node: dict - :type iface_key: str + :param node: Topology node. + :param pf_pcis: List of node's interfaces PCI addresses. + :param mtu: MTU to set. Default: 9200. + :type nodes: dict + :type pf_pcis: list :type mtu: int - :returns: Nothing. - :raises ValueError: If the node type is "DUT". - :raises ValueError: If the node has an unknown node type. + :raises RuntimeError: If failed to set MTU on interface. """ - if node['type'] == NodeType.DUT: - raise ValueError('Node {}: Setting Ethernet MTU for interface ' - 'on DUT nodes not supported', node['host']) - elif node['type'] == NodeType.TG: - iface_name = Topology.get_interface_name(node, iface_key) - cmd = 'ip link set {} mtu {}'.format(iface_name, mtu) + for pf_pci in pf_pcis: + pf_eth = InterfaceUtil.pci_to_eth(node, pf_pci) + cmd = f"ip link set {pf_eth} mtu {mtu}" exec_cmd_no_error(node, cmd, sudo=True) - else: - raise ValueError('Node {} has unknown NodeType: "{}"' - .format(node['host'], node['type'])) @staticmethod - def set_default_ethernet_mtu_on_all_interfaces_on_node(node): - """Set default Ethernet MTU on all interfaces on node. + def set_interface_channels( + node, pf_pcis, num_queues=1, channel=u"combined"): + """Set interface channels for specified interfaces. + + :param node: Topology node. + :param pf_pcis: List of node's interfaces PCI addresses. + :param num_queues: Number of channels. (Optional, Default: 1) + :param channel: Channel type. (Optional, Default: combined) + :type nodes: dict + :type pf_pcis: list + :type num_queues: int + :type channel: str + """ + for pf_pci in pf_pcis: + pf_eth = InterfaceUtil.pci_to_eth(node, pf_pci) + cmd = f"ethtool --set-channels {pf_eth} {channel} {num_queues}" + exec_cmd_no_error(node, cmd, sudo=True) - Function can be used only for TGs. + @staticmethod + def set_interface_flow_control(node, pf_pcis, rxf=u"off", txf=u"off"): + """Set Ethernet flow control for specified interfaces. - :param node: Node where to set default MTU. - :type node: dict - :returns: Nothing. + :param node: Topology node. + :param pf_pcis: List of node's interfaces PCI addresses. + :param rxf: RX flow. (Optional, Default: off). + :param txf: TX flow. (Optional, Default: off). + :type nodes: dict + :type pf_pcis: list + :type rxf: str + :type txf: str """ - for ifc in node['interfaces']: - InterfaceUtil.set_interface_ethernet_mtu(node, ifc, 1500) + for pf_pci in pf_pcis: + pf_eth = InterfaceUtil.pci_to_eth(node, pf_pci) + cmd = f"ethtool -A {pf_eth} rx {rxf} tx {txf}" + ret_code, _, _ = exec_cmd(node, cmd, sudo=True) + if int(ret_code) not in (0, 78): + raise RuntimeError("Failed to set flow control on {pf_eth}!") + + @staticmethod + def set_pci_parameter(node, pf_pcis, key, value): + """Set PCI parameter for specified interfaces. + + :param node: Topology node. + :param pf_pcis: List of node's interfaces PCI addresses. + :param key: Key to set. + :param value: Value to set. + :type nodes: dict + :type pf_pcis: list + :type key: str + :type value: str + """ + for pf_pci in pf_pcis: + cmd = f"setpci -s {pf_pci} {key}={value}" + exec_cmd_no_error(node, cmd, sudo=True) @staticmethod def vpp_set_interface_mtu(node, interface, mtu=9200): @@ -198,23 +339,22 @@ class InterfaceUtil(object): :type interface: str or int :type mtu: int """ - if isinstance(interface, basestring): + if isinstance(interface, str): sw_if_index = Topology.get_interface_sw_index(node, interface) else: sw_if_index = interface - cmd = 'hw_interface_set_mtu' - err_msg = 'Failed to set interface MTU on host {host}'.format( - host=node['host']) - args = dict(sw_if_index=sw_if_index, - mtu=int(mtu)) + cmd = u"hw_interface_set_mtu" + err_msg = f"Failed to set interface MTU on host {node[u'host']}" + args = dict( + sw_if_index=sw_if_index, + mtu=int(mtu) + ) try: with PapiSocketExecutor(node) as papi_exec: papi_exec.add(cmd, **args).get_reply(err_msg) except AssertionError as err: - # TODO: Make failure tolerance optional. - logger.debug("Setting MTU failed. Expected?\n{err}".format( - err=err)) + logger.debug(f"Setting MTU failed.\n{err}") @staticmethod def vpp_set_interfaces_mtu_on_node(node, mtu=9200): @@ -225,7 +365,7 @@ class InterfaceUtil(object): :type node: dict :type mtu: int """ - for interface in node['interfaces']: + for interface in node[u"interfaces"]: InterfaceUtil.vpp_set_interface_mtu(node, interface, mtu) @staticmethod @@ -238,7 +378,7 @@ class InterfaceUtil(object): :type mtu: int """ for node in nodes.values(): - if node['type'] == NodeType.DUT: + if node[u"type"] == NodeType.DUT: InterfaceUtil.vpp_set_interfaces_mtu_on_node(node, mtu) @staticmethod @@ -254,22 +394,22 @@ class InterfaceUtil(object): :raises RuntimeError: If any interface is not in link-up state after defined number of retries. """ - for _ in xrange(0, retries): + for _ in range(0, retries): not_ready = list() out = InterfaceUtil.vpp_get_interface_data(node) for interface in out: - if interface.get('admin_up_down') == 1: - if interface.get('link_up_down') != 1: - not_ready.append(interface.get('interface_name')) - if not not_ready: - break - else: - logger.debug('Interfaces still in link-down state:\n{ifs} ' - '\nWaiting...'.format(ifs=not_ready)) + if interface.get(u"flags") == 1: + not_ready.append(interface.get(u"interface_name")) + if not_ready: + logger.debug( + f"Interfaces still not in link-up state:\n{not_ready}" + ) sleep(1) + else: + break else: - err = 'Timeout, interfaces not up:\n{ifs}'.format(ifs=not_ready) \ - if 'not_ready' in locals() else 'No check executed!' + err = f"Timeout, interfaces not up:\n{not_ready}" \ + if u"not_ready" in locals() else u"No check executed!" raise RuntimeError(err) @staticmethod @@ -285,7 +425,7 @@ class InterfaceUtil(object): :returns: Nothing. """ for node in nodes.values(): - if node['type'] == NodeType.DUT: + if node[u"type"] == NodeType.DUT: InterfaceUtil.vpp_node_interfaces_ready_wait(node, retries) @staticmethod @@ -312,41 +452,48 @@ class InterfaceUtil(object): :returns: Processed interface dump. :rtype: dict """ - if_dump['interface_name'] = if_dump['interface_name'].rstrip('\x00') - if_dump['tag'] = if_dump['tag'].rstrip('\x00') - if_dump['l2_address'] = L2Util.bin_to_mac(if_dump['l2_address']) - if_dump['b_dmac'] = L2Util.bin_to_mac(if_dump['b_dmac']) - if_dump['b_smac'] = L2Util.bin_to_mac(if_dump['b_smac']) + if_dump[u"l2_address"] = str(if_dump[u"l2_address"]) + if_dump[u"b_dmac"] = str(if_dump[u"b_dmac"]) + if_dump[u"b_smac"] = str(if_dump[u"b_smac"]) + if_dump[u"flags"] = if_dump[u"flags"].value + if_dump[u"type"] = if_dump[u"type"].value + if_dump[u"link_duplex"] = if_dump[u"link_duplex"].value + if_dump[u"sub_if_flags"] = if_dump[u"sub_if_flags"].value \ + if hasattr(if_dump[u"sub_if_flags"], u"value") \ + else int(if_dump[u"sub_if_flags"]) + return if_dump if interface is not None: - if isinstance(interface, basestring): - param = 'interface_name' + if isinstance(interface, str): + param = u"interface_name" elif isinstance(interface, int): - param = 'sw_if_index' + param = u"sw_if_index" else: - raise TypeError('Wrong interface format {ifc}'.format( - ifc=interface)) + raise TypeError(f"Wrong interface format {interface}") else: - param = '' + param = u"" + + cmd = u"sw_interface_dump" + args = dict( + name_filter_valid=False, + name_filter=u"" + ) + err_msg = f"Failed to get interface dump on host {node[u'host']}" - cmd = 'sw_interface_dump' - args = dict(name_filter_valid=0, - name_filter='') - err_msg = 'Failed to get interface dump on host {host}'.format( - host=node['host']) with PapiSocketExecutor(node) as papi_exec: details = papi_exec.add(cmd, **args).get_details(err_msg) + logger.debug(f"Received data:\n{details!r}") data = list() if interface is None else dict() for dump in details: if interface is None: data.append(process_if_dump(dump)) - elif str(dump.get(param)).rstrip('\x00') == str(interface): + elif str(dump.get(param)).rstrip(u"\x00") == str(interface): data = process_if_dump(dump) break - logger.debug('Interface data:\n{if_data}'.format(if_data=data)) + logger.debug(f"Interface data:\n{data}") return data @staticmethod @@ -362,11 +509,12 @@ class InterfaceUtil(object): :rtype: str """ if_data = InterfaceUtil.vpp_get_interface_data(node, sw_if_index) - if if_data['sup_sw_if_index'] != if_data['sw_if_index']: + if if_data[u"sup_sw_if_index"] != if_data[u"sw_if_index"]: if_data = InterfaceUtil.vpp_get_interface_data( - node, if_data['sup_sw_if_index']) + node, if_data[u"sup_sw_if_index"] + ) - return if_data.get('interface_name') + return if_data.get(u"interface_name") @staticmethod def vpp_get_interface_sw_index(node, interface_name): @@ -382,7 +530,7 @@ class InterfaceUtil(object): """ if_data = InterfaceUtil.vpp_get_interface_data(node, interface_name) - return if_data.get('sw_if_index') + return if_data.get(u"sw_if_index") @staticmethod def vpp_get_interface_mac(node, interface): @@ -396,11 +544,32 @@ class InterfaceUtil(object): :rtype: str """ if_data = InterfaceUtil.vpp_get_interface_data(node, interface) - if if_data['sup_sw_if_index'] != if_data['sw_if_index']: + if if_data[u"sup_sw_if_index"] != if_data[u"sw_if_index"]: if_data = InterfaceUtil.vpp_get_interface_data( - node, if_data['sup_sw_if_index']) + node, if_data[u"sup_sw_if_index"]) + + return if_data.get(u"l2_address") - return if_data.get('l2_address') + @staticmethod + def vpp_set_interface_mac(node, interface, mac): + """Set MAC address for the given interface. + + :param node: VPP node to set interface MAC. + :param interface: Numeric index or name string of a specific interface. + :param mac: Required MAC address. + :type node: dict + :type interface: int or str + :type mac: str + """ + cmd = u"sw_interface_set_mac_address" + args = dict( + sw_if_index=InterfaceUtil.get_interface_index(node, interface), + mac_address=L2Util.mac_to_bin(mac) + ) + err_msg = f"Failed to set MAC address of interface {interface}" \ + f"on host {node[u'host']}" + with PapiSocketExecutor(node) as papi_exec: + papi_exec.add(cmd, **args).get_reply(err_msg) @staticmethod def tg_set_interface_driver(node, pci_addr, driver): @@ -424,20 +593,17 @@ class InterfaceUtil(object): # Unbind from current driver if old_driver is not None: - cmd = 'sh -c "echo {0} > /sys/bus/pci/drivers/{1}/unbind"'\ - .format(pci_addr, old_driver) - (ret_code, _, _) = ssh.exec_command_sudo(cmd) + cmd = f"sh -c \"echo {pci_addr} > " \ + f"/sys/bus/pci/drivers/{old_driver}/unbind\"" + ret_code, _, _ = ssh.exec_command_sudo(cmd) if int(ret_code) != 0: - raise RuntimeError("'{0}' failed on '{1}'" - .format(cmd, node['host'])) + raise RuntimeError(f"'{cmd}' failed on '{node[u'host']}'") # Bind to the new driver - cmd = 'sh -c "echo {0} > /sys/bus/pci/drivers/{1}/bind"'\ - .format(pci_addr, driver) - (ret_code, _, _) = ssh.exec_command_sudo(cmd) + cmd = f"sh -c \"echo {pci_addr} > /sys/bus/pci/drivers/{driver}/bind\"" + ret_code, _, _ = ssh.exec_command_sudo(cmd) if int(ret_code) != 0: - raise RuntimeError("'{0}' failed on '{1}'" - .format(cmd, node['host'])) + raise RuntimeError(f"'{cmd}' failed on '{node[u'host']}'") @staticmethod def tg_get_interface_driver(node, pci_addr): @@ -453,46 +619,6 @@ class InterfaceUtil(object): """ return DUTSetup.get_pci_dev_driver(node, pci_addr) - @staticmethod - def tg_set_interfaces_udev_rules(node): - """Set udev rules for interfaces. - - Create udev rules file in /etc/udev/rules.d where are rules for each - interface used by TG node, based on MAC interface has specific name. - So after unbind and bind again to kernel driver interface has same - name as before. This must be called after TG has set name for each - port in topology dictionary. - udev rule example - SUBSYSTEM=="net", ACTION=="add", ATTR{address}=="52:54:00:e1:8a:0f", - NAME="eth1" - - :param node: Node to set udev rules on (must be TG node). - :type node: dict - :raises RuntimeError: If setting of udev rules fails. - """ - ssh = SSH() - ssh.connect(node) - - cmd = 'rm -f {0}'.format(InterfaceUtil.__UDEV_IF_RULES_FILE) - (ret_code, _, _) = ssh.exec_command_sudo(cmd) - if int(ret_code) != 0: - raise RuntimeError("'{0}' failed on '{1}'" - .format(cmd, node['host'])) - - for interface in node['interfaces'].values(): - rule = 'SUBSYSTEM==\\"net\\", ACTION==\\"add\\", ATTR{address}' + \ - '==\\"' + interface['mac_address'] + '\\", NAME=\\"' + \ - interface['name'] + '\\"' - cmd = 'sh -c "echo \'{0}\' >> {1}"'.format( - rule, InterfaceUtil.__UDEV_IF_RULES_FILE) - (ret_code, _, _) = ssh.exec_command_sudo(cmd) - if int(ret_code) != 0: - raise RuntimeError("'{0}' failed on '{1}'" - .format(cmd, node['host'])) - - cmd = '/etc/init.d/udev restart' - ssh.exec_command_sudo(cmd) - @staticmethod def tg_set_interfaces_default_driver(node): """Set interfaces default driver specified in topology yaml file. @@ -500,10 +626,10 @@ class InterfaceUtil(object): :param node: Node to setup interfaces driver on (must be TG node). :type node: dict """ - for interface in node['interfaces'].values(): - InterfaceUtil.tg_set_interface_driver(node, - interface['pci_address'], - interface['driver']) + for interface in node[u"interfaces"].values(): + InterfaceUtil.tg_set_interface_driver( + node, interface[u"pci_address"], interface[u"driver"] + ) @staticmethod def update_vpp_interface_data_on_node(node): @@ -521,20 +647,24 @@ class InterfaceUtil(object): interface_list = InterfaceUtil.vpp_get_interface_data(node) interface_dict = dict() for ifc in interface_list: - interface_dict[ifc['l2_address']] = ifc + interface_dict[ifc[u"l2_address"]] = ifc - for if_name, if_data in node['interfaces'].items(): - ifc_dict = interface_dict.get(if_data['mac_address']) + for if_name, if_data in node[u"interfaces"].items(): + ifc_dict = interface_dict.get(if_data[u"mac_address"]) if ifc_dict is not None: - if_data['name'] = ifc_dict['interface_name'] - if_data['vpp_sw_index'] = ifc_dict['sw_if_index'] - if_data['mtu'] = ifc_dict['mtu'][0] - logger.trace('Interface {ifc} found by MAC {mac}'.format( - ifc=if_name, mac=if_data['mac_address'])) + if_data[u"name"] = ifc_dict[u"interface_name"] + if_data[u"vpp_sw_index"] = ifc_dict[u"sw_if_index"] + if_data[u"mtu"] = ifc_dict[u"mtu"][0] + logger.trace( + f"Interface {if_name} found by MAC " + f"{if_data[u'mac_address']}" + ) else: - logger.trace('Interface {ifc} not found by MAC {mac}'.format( - ifc=if_name, mac=if_data['mac_address'])) - if_data['vpp_sw_index'] = None + logger.trace( + f"Interface {if_name} not found by MAC " + f"{if_data[u'mac_address']}" + ) + if_data[u"vpp_sw_index"] = None @staticmethod def update_nic_interface_names(node): @@ -545,24 +675,22 @@ class InterfaceUtil(object): :param node: Node dictionary. :type node: dict """ - for ifc in node['interfaces'].values(): - if_pci = ifc['pci_address'].replace('.', ':').split(':') - bus = '{:x}'.format(int(if_pci[1], 16)) - dev = '{:x}'.format(int(if_pci[2], 16)) - fun = '{:x}'.format(int(if_pci[3], 16)) - loc = '{bus}/{dev}/{fun}'.format(bus=bus, dev=dev, fun=fun) - if ifc['model'] == 'Intel-XL710': - ifc['name'] = 'FortyGigabitEthernet{loc}'.format(loc=loc) - elif ifc['model'] == 'Intel-X710': - ifc['name'] = 'TenGigabitEthernet{loc}'.format(loc=loc) - elif ifc['model'] == 'Intel-X520-DA2': - ifc['name'] = 'TenGigabitEthernet{loc}'.format(loc=loc) - elif ifc['model'] == 'Cisco-VIC-1385': - ifc['name'] = 'FortyGigabitEthernet{loc}'.format(loc=loc) - elif ifc['model'] == 'Cisco-VIC-1227': - ifc['name'] = 'TenGigabitEthernet{loc}'.format(loc=loc) + for ifc in node[u"interfaces"].values(): + if_pci = ifc[u"pci_address"].replace(u".", u":").split(u":") + loc = f"{int(if_pci[1], 16):x}/{int(if_pci[2], 16):x}/" \ + f"{int(if_pci[3], 16):x}" + if ifc[u"model"] == u"Intel-XL710": + ifc[u"name"] = f"FortyGigabitEthernet{loc}" + elif ifc[u"model"] == u"Intel-X710": + ifc[u"name"] = f"TenGigabitEthernet{loc}" + elif ifc[u"model"] == u"Intel-X520-DA2": + ifc[u"name"] = f"TenGigabitEthernet{loc}" + elif ifc[u"model"] == u"Cisco-VIC-1385": + ifc[u"name"] = f"FortyGigabitEthernet{loc}" + elif ifc[u"model"] == u"Cisco-VIC-1227": + ifc[u"name"] = f"TenGigabitEthernet{loc}" else: - ifc['name'] = 'UnknownEthernet{loc}'.format(loc=loc) + ifc[u"name"] = f"UnknownEthernet{loc}" @staticmethod def update_nic_interface_names_on_all_duts(nodes): @@ -574,11 +702,11 @@ class InterfaceUtil(object): :type nodes: dict """ for node in nodes.values(): - if node['type'] == NodeType.DUT: + if node[u"type"] == NodeType.DUT: InterfaceUtil.update_nic_interface_names(node) @staticmethod - def update_tg_interface_data_on_node(node, skip_tg_udev=False): + def update_tg_interface_data_on_node(node): """Update interface name for TG/linux node in DICT__nodes. .. note:: @@ -590,9 +718,7 @@ class InterfaceUtil(object): "00:00:00:00:00:00": "lo" :param node: Node selected from DICT__nodes. - :param skip_tg_udev: Skip udev rename on TG node. :type node: dict - :type skip_tg_udev: bool :raises RuntimeError: If getting of interface name and MAC fails. """ # First setup interface driver specified in yaml file @@ -602,23 +728,20 @@ class InterfaceUtil(object): ssh = SSH() ssh.connect(node) - cmd = ('for dev in `ls /sys/class/net/`; do echo "\\"`cat ' - '/sys/class/net/$dev/address`\\": \\"$dev\\""; done;') + cmd = u'for dev in `ls /sys/class/net/`; do echo "\\"`cat ' \ + u'/sys/class/net/$dev/address`\\": \\"$dev\\""; done;' - (ret_code, stdout, _) = ssh.exec_command(cmd) + ret_code, stdout, _ = ssh.exec_command(cmd) if int(ret_code) != 0: - raise RuntimeError('Get interface name and MAC failed') - tmp = "{" + stdout.rstrip().replace('\n', ',') + "}" + raise RuntimeError(u"Get interface name and MAC failed") + tmp = u"{" + stdout.rstrip().replace(u"\n", u",") + u"}" + interfaces = JsonParser().parse_data(tmp) - for interface in node['interfaces'].values(): - name = interfaces.get(interface['mac_address']) + for interface in node[u"interfaces"].values(): + name = interfaces.get(interface[u"mac_address"]) if name is None: continue - interface['name'] = name - - # Set udev rules for interfaces - if not skip_tg_udev: - InterfaceUtil.tg_set_interfaces_udev_rules(node) + interface[u"name"] = name @staticmethod def iface_update_numa_node(node): @@ -629,55 +752,33 @@ class InterfaceUtil(object): :type node: dict :returns: Nothing. :raises ValueError: If numa node ia less than 0. - :raises RuntimeError: If update of numa node failes. + :raises RuntimeError: If update of numa node failed. """ ssh = SSH() for if_key in Topology.get_node_interfaces(node): if_pci = Topology.get_interface_pci_addr(node, if_key) ssh.connect(node) - cmd = "cat /sys/bus/pci/devices/{}/numa_node".format(if_pci) + cmd = f"cat /sys/bus/pci/devices/{if_pci}/numa_node" for _ in range(3): - (ret, out, _) = ssh.exec_command(cmd) + ret, out, _ = ssh.exec_command(cmd) if ret == 0: try: - numa_node = int(out) - if numa_node < 0: - if CpuUtils.cpu_node_count(node) == 1: - numa_node = 0 - else: - raise ValueError + numa_node = 0 if int(out) < 0 else int(out) except ValueError: - logger.trace('Reading numa location failed for: {0}' - .format(if_pci)) + logger.trace( + f"Reading numa location failed for: {if_pci}" + ) else: - Topology.set_interface_numa_node(node, if_key, - numa_node) + Topology.set_interface_numa_node( + node, if_key, numa_node + ) break else: - raise RuntimeError('Update numa node failed for: {0}' - .format(if_pci)) + raise RuntimeError(f"Update numa node failed for: {if_pci}") @staticmethod - def update_all_numa_nodes(nodes, skip_tg=False): - """For all nodes and all their interfaces from topology file update numa - node information based on information from the node. - - :param nodes: Nodes in the topology. - :param skip_tg: Skip TG node - :type nodes: dict - :type skip_tg: bool - :returns: Nothing. - """ - for node in nodes.values(): - if node['type'] == NodeType.DUT: - InterfaceUtil.iface_update_numa_node(node) - elif node['type'] == NodeType.TG and not skip_tg: - InterfaceUtil.iface_update_numa_node(node) - - @staticmethod - def update_all_interface_data_on_all_nodes(nodes, skip_tg=False, - skip_tg_udev=False, - numa_node=False): + def update_all_interface_data_on_all_nodes( + nodes, skip_tg=False, skip_vpp=False): """Update interface names on all nodes in DICT__nodes. This method updates the topology dictionary by querying interface lists @@ -685,25 +786,17 @@ class InterfaceUtil(object): :param nodes: Nodes in the topology. :param skip_tg: Skip TG node. - :param skip_tg_udev: Skip udev rename on TG node. - :param numa_node: Retrieve numa_node location. + :param skip_vpp: Skip VPP node. :type nodes: dict :type skip_tg: bool - :type skip_tg_udev: bool - :type numa_node: bool + :type skip_vpp: bool """ - for node_data in nodes.values(): - if node_data['type'] == NodeType.DUT: - InterfaceUtil.update_vpp_interface_data_on_node(node_data) - elif node_data['type'] == NodeType.TG and not skip_tg: - InterfaceUtil.update_tg_interface_data_on_node( - node_data, skip_tg_udev) - - if numa_node: - if node_data['type'] == NodeType.DUT: - InterfaceUtil.iface_update_numa_node(node_data) - elif node_data['type'] == NodeType.TG and not skip_tg: - InterfaceUtil.iface_update_numa_node(node_data) + for node in nodes.values(): + if node[u"type"] == NodeType.DUT and not skip_vpp: + InterfaceUtil.update_vpp_interface_data_on_node(node) + elif node[u"type"] == NodeType.TG and not skip_tg: + InterfaceUtil.update_tg_interface_data_on_node(node) + InterfaceUtil.iface_update_numa_node(node) @staticmethod def create_vlan_subinterface(node, interface, vlan): @@ -723,20 +816,22 @@ class InterfaceUtil(object): """ sw_if_index = InterfaceUtil.get_interface_index(node, interface) - cmd = 'create_vlan_subif' - args = dict(sw_if_index=sw_if_index, - vlan_id=int(vlan)) - err_msg = 'Failed to create VLAN sub-interface on host {host}'.format( - host=node['host']) + cmd = u"create_vlan_subif" + args = dict( + sw_if_index=sw_if_index, + vlan_id=int(vlan) + ) + err_msg = f"Failed to create VLAN sub-interface on host {node[u'host']}" + with PapiSocketExecutor(node) as papi_exec: sw_if_index = papi_exec.add(cmd, **args).get_sw_if_index(err_msg) - if_key = Topology.add_new_port(node, 'vlan_subif') + if_key = Topology.add_new_port(node, u"vlan_subif") Topology.update_interface_sw_if_index(node, if_key, sw_if_index) ifc_name = InterfaceUtil.vpp_get_interface_name(node, sw_if_index) Topology.update_interface_name(node, if_key, ifc_name) - return '{ifc}.{vlan}'.format(ifc=interface, vlan=vlan), sw_if_index + return f"{interface}.{vlan}", sw_if_index @staticmethod def create_vxlan_interface(node, vni, source_ip, destination_ip): @@ -755,25 +850,27 @@ class InterfaceUtil(object): :raises RuntimeError: if it is unable to create VxLAN interface on the node. """ - src_address = ip_address(unicode(source_ip)) - dst_address = ip_address(unicode(destination_ip)) - - cmd = 'vxlan_add_del_tunnel' - args = dict(is_add=1, - is_ipv6=1 if src_address.version == 6 else 0, - instance=Constants.BITWISE_NON_ZERO, - src_address=src_address.packed, - dst_address=dst_address.packed, - mcast_sw_if_index=Constants.BITWISE_NON_ZERO, - encap_vrf_id=0, - decap_next_index=Constants.BITWISE_NON_ZERO, - vni=int(vni)) - err_msg = 'Failed to create VXLAN tunnel interface on host {host}'.\ - format(host=node['host']) + cmd = u"vxlan_add_del_tunnel" + args = dict( + is_add=True, + instance=Constants.BITWISE_NON_ZERO, + src_address=IPAddress.create_ip_address_object( + ip_address(source_ip) + ), + dst_address=IPAddress.create_ip_address_object( + ip_address(destination_ip) + ), + mcast_sw_if_index=Constants.BITWISE_NON_ZERO, + encap_vrf_id=0, + decap_next_index=Constants.BITWISE_NON_ZERO, + vni=int(vni) + ) + err_msg = f"Failed to create VXLAN tunnel interface " \ + f"on host {node[u'host']}" with PapiSocketExecutor(node) as papi_exec: sw_if_index = papi_exec.add(cmd, **args).get_sw_if_index(err_msg) - if_key = Topology.add_new_port(node, 'vxlan_tunnel') + if_key = Topology.add_new_port(node, u"vxlan_tunnel") Topology.update_interface_sw_if_index(node, if_key, sw_if_index) ifc_name = InterfaceUtil.vpp_get_interface_name(node, sw_if_index) Topology.update_interface_name(node, if_key, ifc_name) @@ -798,12 +895,14 @@ class InterfaceUtil(object): """ sw_if_index = InterfaceUtil.get_interface_index(node, interface) - cmd = 'sw_interface_set_vxlan_bypass' - args = dict(is_ipv6=0, - sw_if_index=sw_if_index, - enable=1) - err_msg = 'Failed to set VXLAN bypass on interface on host {host}'.\ - format(host=node['host']) + cmd = u"sw_interface_set_vxlan_bypass" + args = dict( + is_ipv6=False, + sw_if_index=sw_if_index, + enable=True + ) + err_msg = f"Failed to set VXLAN bypass on interface " \ + f"on host {node[u'host']}" with PapiSocketExecutor(node) as papi_exec: papi_exec.add(cmd, **args).get_replies(err_msg) @@ -830,16 +929,8 @@ class InterfaceUtil(object): :returns: Processed vxlan interface dump. :rtype: dict """ - if vxlan_dump['is_ipv6']: - vxlan_dump['src_address'] = \ - ip_address(unicode(vxlan_dump['src_address'])) - vxlan_dump['dst_address'] = \ - ip_address(unicode(vxlan_dump['dst_address'])) - else: - vxlan_dump['src_address'] = \ - ip_address(unicode(vxlan_dump['src_address'][0:4])) - vxlan_dump['dst_address'] = \ - ip_address(unicode(vxlan_dump['dst_address'][0:4])) + vxlan_dump[u"src_address"] = str(vxlan_dump[u"src_address"]) + vxlan_dump[u"dst_address"] = str(vxlan_dump[u"dst_address"]) return vxlan_dump if interface is not None: @@ -847,10 +938,12 @@ class InterfaceUtil(object): else: sw_if_index = int(Constants.BITWISE_NON_ZERO) - cmd = 'vxlan_tunnel_dump' - args = dict(sw_if_index=sw_if_index) - err_msg = 'Failed to get VXLAN dump on host {host}'.format( - host=node['host']) + cmd = u"vxlan_tunnel_dump" + args = dict( + sw_if_index=sw_if_index + ) + err_msg = f"Failed to get VXLAN dump on host {node[u'host']}" + with PapiSocketExecutor(node) as papi_exec: details = papi_exec.add(cmd, **args).get_details(err_msg) @@ -858,55 +951,17 @@ class InterfaceUtil(object): for dump in details: if interface is None: data.append(process_vxlan_dump(dump)) - elif dump['sw_if_index'] == sw_if_index: + elif dump[u"sw_if_index"] == sw_if_index: data = process_vxlan_dump(dump) break - logger.debug('VXLAN data:\n{vxlan_data}'.format(vxlan_data=data)) + logger.debug(f"VXLAN data:\n{data}") return data @staticmethod - def vhost_user_dump(node): - """Get vhost-user data for the given node. - - TODO: Move to VhostUser.py - - :param node: VPP node to get interface data from. - :type node: dict - :returns: List of dictionaries with all vhost-user interfaces. - :rtype: list - """ - def process_vhost_dump(vhost_dump): - """Process vhost dump. - - :param vhost_dump: Vhost interface dump. - :type vhost_dump: dict - :returns: Processed vhost interface dump. - :rtype: dict - """ - vhost_dump['interface_name'] = \ - vhost_dump['interface_name'].rstrip('\x00') - vhost_dump['sock_filename'] = \ - vhost_dump['sock_filename'].rstrip('\x00') - return vhost_dump - - cmd = 'sw_interface_vhost_user_dump' - err_msg = 'Failed to get vhost-user dump on host {host}'.format( - host=node['host']) - with PapiSocketExecutor(node) as papi_exec: - details = papi_exec.add(cmd).get_details(err_msg) - - for dump in details: - # In-place edits. - process_vhost_dump(dump) - - logger.debug('Vhost-user details:\n{vhost_details}'.format( - vhost_details=details)) - return details - - @staticmethod - def create_subinterface(node, interface, sub_id, outer_vlan_id=None, - inner_vlan_id=None, type_subif=None): + def create_subinterface( + node, interface, sub_id, outer_vlan_id=None, inner_vlan_id=None, + type_subif=None): """Create sub-interface on node. It is possible to set required sub-interface type and VLAN tag(s). @@ -930,31 +985,42 @@ class InterfaceUtil(object): """ subif_types = type_subif.split() - cmd = 'create_subif' + flags = 0 + if u"no_tags" in subif_types: + flags = flags | SubInterfaceFlags.SUB_IF_API_FLAG_NO_TAGS + if u"one_tag" in subif_types: + flags = flags | SubInterfaceFlags.SUB_IF_API_FLAG_ONE_TAG + if u"two_tags" in subif_types: + flags = flags | SubInterfaceFlags.SUB_IF_API_FLAG_TWO_TAGS + if u"dot1ad" in subif_types: + flags = flags | SubInterfaceFlags.SUB_IF_API_FLAG_DOT1AD + if u"exact_match" in subif_types: + flags = flags | SubInterfaceFlags.SUB_IF_API_FLAG_EXACT_MATCH + if u"default_sub" in subif_types: + flags = flags | SubInterfaceFlags.SUB_IF_API_FLAG_DEFAULT + if type_subif == u"default_sub": + flags = flags | SubInterfaceFlags.SUB_IF_API_FLAG_INNER_VLAN_ID_ANY\ + | SubInterfaceFlags.SUB_IF_API_FLAG_OUTER_VLAN_ID_ANY + + cmd = u"create_subif" args = dict( sw_if_index=InterfaceUtil.get_interface_index(node, interface), sub_id=int(sub_id), - no_tags=1 if 'no_tags' in subif_types else 0, - one_tag=1 if 'one_tag' in subif_types else 0, - two_tags=1 if 'two_tags' in subif_types else 0, - dot1ad=1 if 'dot1ad' in subif_types else 0, - exact_match=1 if 'exact_match' in subif_types else 0, - default_sub=1 if 'default_sub' in subif_types else 0, - outer_vlan_id_any=1 if type_subif == 'default_sub' else 0, - inner_vlan_id_any=1 if type_subif == 'default_sub' else 0, + sub_if_flags=flags.value if hasattr(flags, u"value") + else int(flags), outer_vlan_id=int(outer_vlan_id) if outer_vlan_id else 0, - inner_vlan_id=int(inner_vlan_id) if inner_vlan_id else 0) - err_msg = 'Failed to create sub-interface on host {host}'.format( - host=node['host']) + inner_vlan_id=int(inner_vlan_id) if inner_vlan_id else 0 + ) + err_msg = f"Failed to create sub-interface on host {node[u'host']}" with PapiSocketExecutor(node) as papi_exec: sw_if_index = papi_exec.add(cmd, **args).get_sw_if_index(err_msg) - if_key = Topology.add_new_port(node, 'subinterface') + if_key = Topology.add_new_port(node, u"subinterface") Topology.update_interface_sw_if_index(node, if_key, sw_if_index) ifc_name = InterfaceUtil.vpp_get_interface_name(node, sw_if_index) Topology.update_interface_name(node, if_key, ifc_name) - return '{ifc}.{s_id}'.format(ifc=interface, s_id=sub_id), sw_if_index + return f"{interface}.{sub_id}", sw_if_index @staticmethod def create_gre_tunnel_interface(node, source_ip, destination_ip): @@ -970,33 +1036,80 @@ class InterfaceUtil(object): :rtype: tuple :raises RuntimeError: If unable to create GRE tunnel interface. """ - cmd = 'gre_tunnel_add_del' - tunnel = dict(type=0, - instance=Constants.BITWISE_NON_ZERO, - src=str(source_ip), - dst=str(destination_ip), - outer_fib_id=0, - session_id=0) - args = dict(is_add=1, - tunnel=tunnel) - err_msg = 'Failed to create GRE tunnel interface on host {host}'.format( - host=node['host']) + cmd = u"gre_tunnel_add_del" + tunnel = dict( + type=0, + instance=Constants.BITWISE_NON_ZERO, + src=str(source_ip), + dst=str(destination_ip), + outer_fib_id=0, + session_id=0 + ) + args = dict( + is_add=1, + tunnel=tunnel + ) + err_msg = f"Failed to create GRE tunnel interface " \ + f"on host {node[u'host']}" with PapiSocketExecutor(node) as papi_exec: sw_if_index = papi_exec.add(cmd, **args).get_sw_if_index(err_msg) - if_key = Topology.add_new_port(node, 'gre_tunnel') + if_key = Topology.add_new_port(node, u"gre_tunnel") Topology.update_interface_sw_if_index(node, if_key, sw_if_index) ifc_name = InterfaceUtil.vpp_get_interface_name(node, sw_if_index) Topology.update_interface_name(node, if_key, ifc_name) return ifc_name, sw_if_index + @staticmethod + def create_gtpu_tunnel_interface(node, teid, source_ip, destination_ip): + """Create GTPU interface and return sw if index of created interface. + + :param node: Node where to create GTPU interface. + :param teid: GTPU Tunnel Endpoint Identifier. + :param source_ip: Source IP of a GTPU Tunnel End Point. + :param destination_ip: Destination IP of a GTPU Tunnel End Point. + :type node: dict + :type teid: int + :type source_ip: str + :type destination_ip: str + :returns: SW IF INDEX of created interface. + :rtype: int + :raises RuntimeError: if it is unable to create GTPU interface on the + node. + """ + cmd = u"gtpu_add_del_tunnel" + args = dict( + is_add=True, + src_address=IPAddress.create_ip_address_object( + ip_address(source_ip) + ), + dst_address=IPAddress.create_ip_address_object( + ip_address(destination_ip) + ), + mcast_sw_if_index=Constants.BITWISE_NON_ZERO, + encap_vrf_id=0, + decap_next_index=2, + teid=teid + ) + err_msg = f"Failed to create GTPU tunnel interface " \ + f"on host {node[u'host']}" + with PapiSocketExecutor(node) as papi_exec: + sw_if_index = papi_exec.add(cmd, **args).get_sw_if_index(err_msg) + + if_key = Topology.add_new_port(node, u"gtpu_tunnel") + Topology.update_interface_sw_if_index(node, if_key, sw_if_index) + ifc_name = InterfaceUtil.vpp_get_interface_name(node, sw_if_index) + Topology.update_interface_name(node, if_key, ifc_name) + + return sw_if_index + @staticmethod def vpp_create_loopback(node, mac=None): """Create loopback interface on VPP node. :param node: Node to create loopback interface on. - :param mac: Optional MAC address for Loopback interface. + :param mac: Optional MAC address for loopback interface. :type node: dict :type mac: str :returns: SW interface index. @@ -1004,14 +1117,17 @@ class InterfaceUtil(object): :raises RuntimeError: If it is not possible to create loopback on the node. """ - cmd = 'create_loopback' - args = dict(mac_address=L2Util.mac_to_bin(mac) if mac else 0) - err_msg = 'Failed to create loopback interface on host {host}'.format( - host=node['host']) + cmd = u"create_loopback_instance" + args = dict( + mac_address=L2Util.mac_to_bin(mac) if mac else 0, + is_specified=False, + user_instance=0, + ) + err_msg = f"Failed to create loopback interface on host {node[u'host']}" with PapiSocketExecutor(node) as papi_exec: sw_if_index = papi_exec.add(cmd, **args).get_sw_if_index(err_msg) - if_key = Topology.add_new_port(node, 'loopback') + if_key = Topology.add_new_port(node, u"loopback") Topology.update_interface_sw_if_index(node, if_key, sw_if_index) ifc_name = InterfaceUtil.vpp_get_interface_name(node, sw_if_index) Topology.update_interface_name(node, if_key, ifc_name) @@ -1022,55 +1138,70 @@ class InterfaceUtil(object): return sw_if_index @staticmethod - def vpp_create_bond_interface(node, mode, load_balance=None, mac=None): + def vpp_create_bond_interface( + node, mode, load_balance=None, mac=None, gso=False): """Create bond interface on VPP node. :param node: DUT node from topology. :param mode: Link bonding mode. :param load_balance: Load balance (optional, valid for xor and lacp - modes, otherwise ignored). + modes, otherwise ignored). Default: None. :param mac: MAC address to assign to the bond interface (optional). + Default: None. + :param gso: Enable GSO support (optional). Default: False. :type node: dict :type mode: str :type load_balance: str :type mac: str + :type gso: bool :returns: Interface key (name) in topology. :rtype: str :raises RuntimeError: If it is not possible to create bond interface on the node. """ - cmd = 'bond_create' - args = dict(id=int(Constants.BITWISE_NON_ZERO), - use_custom_mac=0 if mac is None else 1, - mac_address=0 if mac is None else L2Util.mac_to_bin(mac), - mode=getattr(LinkBondMode, '{md}'.format( - md=mode.replace('-', '_').upper())).value, - lb=0 if load_balance is None else getattr( - LinkBondLoadBalance, '{lb}'.format( - lb=load_balance.upper())).value) - err_msg = 'Failed to create bond interface on host {host}'.format( - host=node['host']) + cmd = u"bond_create2" + args = dict( + id=int(Constants.BITWISE_NON_ZERO), + use_custom_mac=bool(mac is not None), + mac_address=L2Util.mac_to_bin(mac) if mac else None, + mode=getattr( + LinkBondMode, + f"BOND_API_MODE_{mode.replace(u'-', u'_').upper()}" + ).value, + lb=0 if load_balance is None else getattr( + LinkBondLoadBalanceAlgo, + f"BOND_API_LB_ALGO_{load_balance.upper()}" + ).value, + numa_only=False, + enable_gso=gso + ) + err_msg = f"Failed to create bond interface on host {node[u'host']}" with PapiSocketExecutor(node) as papi_exec: sw_if_index = papi_exec.add(cmd, **args).get_sw_if_index(err_msg) - InterfaceUtil.add_eth_interface(node, sw_if_index=sw_if_index, - ifc_pfx='eth_bond') + InterfaceUtil.add_eth_interface( + node, sw_if_index=sw_if_index, ifc_pfx=u"eth_bond" + ) if_key = Topology.get_interface_by_sw_index(node, sw_if_index) return if_key @staticmethod - def add_eth_interface(node, ifc_name=None, sw_if_index=None, ifc_pfx=None): + def add_eth_interface( + node, ifc_name=None, sw_if_index=None, ifc_pfx=None, + host_if_key=None): """Add ethernet interface to current topology. :param node: DUT node from topology. :param ifc_name: Name of the interface. :param sw_if_index: SW interface index. :param ifc_pfx: Interface key prefix. + :param host_if_key: Host interface key from topology file. :type node: dict :type ifc_name: str :type sw_if_index: int :type ifc_pfx: str + :type host_if_key: str """ if_key = Topology.add_new_port(node, ifc_pfx) @@ -1083,42 +1214,171 @@ class InterfaceUtil(object): Topology.update_interface_name(node, if_key, ifc_name) ifc_mac = InterfaceUtil.vpp_get_interface_mac(node, sw_if_index) Topology.update_interface_mac_address(node, if_key, ifc_mac) + if host_if_key is not None: + Topology.set_interface_numa_node( + node, if_key, Topology.get_interface_numa_node( + node, host_if_key + ) + ) + Topology.update_interface_pci_address( + node, if_key, Topology.get_interface_pci_addr(node, host_if_key) + ) @staticmethod - def vpp_create_avf_interface(node, vf_pci_addr, num_rx_queues=None): + def vpp_create_avf_interface( + node, if_key, num_rx_queues=None, rxq_size=0, txq_size=0): """Create AVF interface on VPP node. :param node: DUT node from topology. - :param vf_pci_addr: Virtual Function PCI address. + :param if_key: Interface key from topology file of interface + to be bound to i40evf driver. :param num_rx_queues: Number of RX queues. + :param rxq_size: Size of RXQ (0 = Default API; 512 = Default VPP). + :param txq_size: Size of TXQ (0 = Default API; 512 = Default VPP). :type node: dict - :type vf_pci_addr: str + :type if_key: str :type num_rx_queues: int - :returns: Interface key (name) in topology. + :type rxq_size: int + :type txq_size: int + :returns: AVF interface key (name) in topology. :rtype: str :raises RuntimeError: If it is not possible to create AVF interface on the node. """ - cmd = 'avf_create' - args = dict(pci_addr=InterfaceUtil.pci_to_int(vf_pci_addr), - enable_elog=0, - rxq_num=int(num_rx_queues) if num_rx_queues else 0, - rxq_size=0, - txq_size=0) - err_msg = 'Failed to create AVF interface on host {host}'.format( - host=node['host']) + PapiSocketExecutor.run_cli_cmd( + node, u"set logging class avf level debug" + ) + + cmd = u"avf_create" + vf_pci_addr = Topology.get_interface_pci_addr(node, if_key) + args = dict( + pci_addr=InterfaceUtil.pci_to_int(vf_pci_addr), + enable_elog=0, + rxq_num=int(num_rx_queues) if num_rx_queues else 0, + rxq_size=rxq_size, + txq_size=txq_size + ) + err_msg = f"Failed to create AVF interface on host {node[u'host']}" with PapiSocketExecutor(node) as papi_exec: sw_if_index = papi_exec.add(cmd, **args).get_sw_if_index(err_msg) - InterfaceUtil.add_eth_interface(node, sw_if_index=sw_if_index, - ifc_pfx='eth_avf') - if_key = Topology.get_interface_by_sw_index(node, sw_if_index) + InterfaceUtil.add_eth_interface( + node, sw_if_index=sw_if_index, ifc_pfx=u"eth_avf", + host_if_key=if_key + ) - return if_key + return Topology.get_interface_by_sw_index(node, sw_if_index) @staticmethod - def vpp_enslave_physical_interface(node, interface, bond_if): - """Enslave physical interface to bond interface on VPP node. + def vpp_create_af_xdp_interface( + node, if_key, num_rx_queues=None, rxq_size=0, txq_size=0, + mode=u"auto"): + """Create AF_XDP interface on VPP node. + + :param node: DUT node from topology. + :param if_key: Physical interface key from topology file of interface + to be bound to compatible driver. + :param num_rx_queues: Number of RX queues. (Optional, Default: none) + :param rxq_size: Size of RXQ (0 = Default API; 512 = Default VPP). + :param txq_size: Size of TXQ (0 = Default API; 512 = Default VPP). + :param mode: AF_XDP interface mode. (Optional, Default: auto). + :type node: dict + :type if_key: str + :type num_rx_queues: int + :type rxq_size: int + :type txq_size: int + :type mode: str + :returns: Interface key (name) in topology file. + :rtype: str + :raises RuntimeError: If it is not possible to create AF_XDP interface + on the node. + """ + PapiSocketExecutor.run_cli_cmd( + node, u"set logging class af_xdp level debug" + ) + + cmd = u"af_xdp_create" + pci_addr = Topology.get_interface_pci_addr(node, if_key) + args = dict( + name=InterfaceUtil.pci_to_eth(node, pci_addr), + host_if=InterfaceUtil.pci_to_eth(node, pci_addr), + rxq_num=int(num_rx_queues) if num_rx_queues else 0, + rxq_size=rxq_size, + txq_size=txq_size, + mode=getattr(AfXdpMode, f"AF_XDP_API_MODE_{mode.upper()}").value + ) + err_msg = f"Failed to create AF_XDP interface on host {node[u'host']}" + with PapiSocketExecutor(node) as papi_exec: + sw_if_index = papi_exec.add(cmd, **args).get_sw_if_index(err_msg) + + InterfaceUtil.vpp_set_interface_mac( + node, sw_if_index, Topology.get_interface_mac(node, if_key) + ) + InterfaceUtil.add_eth_interface( + node, sw_if_index=sw_if_index, ifc_pfx=u"eth_af_xdp", + host_if_key=if_key + ) + + return Topology.get_interface_by_sw_index(node, sw_if_index) + + @staticmethod + def vpp_create_rdma_interface( + node, if_key, num_rx_queues=None, rxq_size=0, txq_size=0, + mode=u"auto"): + """Create RDMA interface on VPP node. + + :param node: DUT node from topology. + :param if_key: Physical interface key from topology file of interface + to be bound to rdma-core driver. + :param num_rx_queues: Number of RX queues. + :param rxq_size: Size of RXQ (0 = Default API; 512 = Default VPP). + :param txq_size: Size of TXQ (0 = Default API; 512 = Default VPP). + :param mode: RDMA interface mode - auto/ibv/dv. + :type node: dict + :type if_key: str + :type num_rx_queues: int + :type rxq_size: int + :type txq_size: int + :type mode: str + :returns: Interface key (name) in topology file. + :rtype: str + :raises RuntimeError: If it is not possible to create RDMA interface on + the node. + """ + PapiSocketExecutor.run_cli_cmd( + node, u"set logging class rdma level debug" + ) + + cmd = u"rdma_create_v2" + pci_addr = Topology.get_interface_pci_addr(node, if_key) + args = dict( + name=InterfaceUtil.pci_to_eth(node, pci_addr), + host_if=InterfaceUtil.pci_to_eth(node, pci_addr), + rxq_num=int(num_rx_queues) if num_rx_queues else 0, + rxq_size=rxq_size, + txq_size=txq_size, + mode=getattr(RdmaMode, f"RDMA_API_MODE_{mode.upper()}").value, + # Note: Set True for non-jumbo packets. + no_multi_seg=False, + max_pktlen=0, + ) + err_msg = f"Failed to create RDMA interface on host {node[u'host']}" + with PapiSocketExecutor(node) as papi_exec: + sw_if_index = papi_exec.add(cmd, **args).get_sw_if_index(err_msg) + + InterfaceUtil.vpp_set_interface_mac( + node, sw_if_index, Topology.get_interface_mac(node, if_key) + ) + InterfaceUtil.add_eth_interface( + node, sw_if_index=sw_if_index, ifc_pfx=u"eth_rdma", + host_if_key=if_key + ) + + return Topology.get_interface_by_sw_index(node, sw_if_index) + + @staticmethod + def vpp_add_bond_member(node, interface, bond_if): + """Add member interface to bond interface on VPP node. :param node: DUT node from topology. :param interface: Physical interface key from topology file. @@ -1126,19 +1386,18 @@ class InterfaceUtil(object): :type node: dict :type interface: str :type bond_if: str - :raises RuntimeError: If it is not possible to enslave physical - interface to bond interface on the node. + :raises RuntimeError: If it is not possible to add member to bond + interface on the node. """ - cmd = 'bond_enslave' + cmd = u"bond_add_member" args = dict( sw_if_index=Topology.get_interface_sw_index(node, interface), bond_sw_if_index=Topology.get_interface_sw_index(node, bond_if), - is_passive=0, - is_long_timeout=0) - err_msg = 'Failed to enslave physical interface {ifc} to bond ' \ - 'interface {bond} on host {host}'.format(ifc=interface, - bond=bond_if, - host=node['host']) + is_passive=False, + is_long_timeout=False + ) + err_msg = f"Failed to add member {interface} to bond interface " \ + f"{bond_if} on host {node[u'host']}" with PapiSocketExecutor(node) as papi_exec: papi_exec.add(cmd, **args).get_reply(err_msg) @@ -1151,37 +1410,41 @@ class InterfaceUtil(object): :type node: dict :type verbose: bool """ - cmd = 'sw_interface_bond_dump' - err_msg = 'Failed to get bond interface dump on host {host}'.format( - host=node['host']) + cmd = u"sw_bond_interface_dump" + err_msg = f"Failed to get bond interface dump on host {node[u'host']}" - data = ('Bond data on node {host}:\n'.format(host=node['host'])) + data = f"Bond data on node {node[u'host']}:\n" with PapiSocketExecutor(node) as papi_exec: details = papi_exec.add(cmd).get_details(err_msg) for bond in details: - data += ('{b}\n'.format(b=bond['interface_name'].rstrip('\x00'))) - data += (' mode: {m}\n'.format(m=bond['mode']).lower()) - data += (' load balance: {lb}\n'.format(lb=bond['lb']).lower()) - data += (' number of active slaves: {n}\n'.format( - n=bond['active_slaves'])) + data += f"{bond[u'interface_name']}\n" + data += u" mode: {m}\n".format( + m=bond[u"mode"].name.replace(u"BOND_API_MODE_", u"").lower() + ) + data += u" load balance: {lb}\n".format( + lb=bond[u"lb"].name.replace(u"BOND_API_LB_ALGO_", u"").lower() + ) + data += f" number of active members: {bond[u'active_members']}\n" if verbose: - slave_data = InterfaceUtil.vpp_bond_slave_dump( + member_data = InterfaceUtil.vpp_bond_member_dump( node, Topology.get_interface_by_sw_index( - node, bond['sw_if_index'])) - for slave in slave_data: - if not slave['is_passive']: - data += (' {s}\n'.format(s=slave['interface_name'])) - data += (' number of slaves: {n}\n'.format(n=bond['slaves'])) + node, bond[u"sw_if_index"] + ) + ) + for member in member_data: + if not member[u"is_passive"]: + data += f" {member[u'interface_name']}\n" + data += f" number of members: {bond[u'members']}\n" if verbose: - for slave in slave_data: - data += (' {s}\n'.format(s=slave['interface_name'])) - data += (' interface id: {i}\n'.format(i=bond['id'])) - data += (' sw_if_index: {i}\n'.format(i=bond['sw_if_index'])) + for member in member_data: + data += f" {member[u'interface_name']}\n" + data += f" interface id: {bond[u'id']}\n" + data += f" sw_if_index: {bond[u'sw_if_index']}\n" logger.info(data) @staticmethod - def vpp_bond_slave_dump(node, interface): + def vpp_bond_member_dump(node, interface): """Get bond interface slave(s) data on VPP node. :param node: DUT node from topology. @@ -1191,32 +1454,16 @@ class InterfaceUtil(object): :returns: Bond slave interface data. :rtype: dict """ - def process_slave_dump(slave_dump): - """Process slave dump. - - :param slave_dump: Slave interface dump. - :type slave_dump: dict - :returns: Processed slave interface dump. - :rtype: dict - """ - slave_dump['interface_name'] = slave_dump['interface_name'].\ - rstrip('\x00') - return slave_dump - - cmd = 'sw_interface_slave_dump' - args = dict(sw_if_index=Topology.get_interface_sw_index( - node, interface)) - err_msg = 'Failed to get slave dump on host {host}'.format( - host=node['host']) + cmd = u"sw_member_interface_dump" + args = dict( + sw_if_index=Topology.get_interface_sw_index(node, interface) + ) + err_msg = f"Failed to get slave dump on host {node[u'host']}" with PapiSocketExecutor(node) as papi_exec: details = papi_exec.add(cmd, **args).get_details(err_msg) - for dump in details: - # In-place edits. - process_slave_dump(dump) - - logger.debug('Slave data:\n{slave_data}'.format(slave_data=details)) + logger.debug(f"Member data:\n{details}") return details @staticmethod @@ -1229,12 +1476,12 @@ class InterfaceUtil(object): :type verbose: bool """ for node_data in nodes.values(): - if node_data['type'] == NodeType.DUT: + if node_data[u"type"] == NodeType.DUT: InterfaceUtil.vpp_show_bond_data_on_node(node_data, verbose) @staticmethod - def vpp_enable_input_acl_interface(node, interface, ip_version, - table_index): + def vpp_enable_input_acl_interface( + node, interface, ip_version, table_index): """Enable input acl on interface. :param node: VPP node to setup interface for input acl. @@ -1246,18 +1493,17 @@ class InterfaceUtil(object): :type ip_version: str :type table_index: int """ - cmd = 'input_acl_set_interface' + cmd = u"input_acl_set_interface" args = dict( sw_if_index=InterfaceUtil.get_interface_index(node, interface), - ip4_table_index=table_index if ip_version == 'ip4' + ip4_table_index=table_index if ip_version == u"ip4" else Constants.BITWISE_NON_ZERO, - ip6_table_index=table_index if ip_version == 'ip6' + ip6_table_index=table_index if ip_version == u"ip6" else Constants.BITWISE_NON_ZERO, - l2_table_index=table_index if ip_version == 'l2' + l2_table_index=table_index if ip_version == u"l2" else Constants.BITWISE_NON_ZERO, is_add=1) - err_msg = 'Failed to enable input acl on interface {ifc}'.format( - ifc=interface) + err_msg = f"Failed to enable input acl on interface {interface}" with PapiSocketExecutor(node) as papi_exec: papi_exec.add(cmd, **args).get_reply(err_msg) @@ -1274,15 +1520,16 @@ class InterfaceUtil(object): :returns: Classify table name. :rtype: str """ - if isinstance(interface, basestring): + if isinstance(interface, str): sw_if_index = InterfaceUtil.get_sw_if_index(node, interface) else: sw_if_index = interface - cmd = 'classify_table_by_interface' - args = dict(sw_if_index=sw_if_index) - err_msg = 'Failed to get classify table name by interface {ifc}'.format( - ifc=interface) + cmd = u"classify_table_by_interface" + args = dict( + sw_if_index=sw_if_index + ) + err_msg = f"Failed to get classify table name by interface {interface}" with PapiSocketExecutor(node) as papi_exec: reply = papi_exec.add(cmd, **args).get_reply(err_msg) @@ -1292,6 +1539,8 @@ class InterfaceUtil(object): def get_sw_if_index(node, interface_name): """Get sw_if_index for the given interface from actual interface dump. + FIXME: Delete and redirect callers to vpp_get_interface_sw_index. + :param node: VPP node to get interface data from. :param interface_name: Name of the specific interface. :type node: dict @@ -1300,8 +1549,9 @@ class InterfaceUtil(object): :rtype: str """ interface_data = InterfaceUtil.vpp_get_interface_data( - node, interface=interface_name) - return interface_data.get('sw_if_index') + node, interface=interface_name + ) + return interface_data.get(u"sw_if_index") @staticmethod def vxlan_gpe_dump(node, interface_name=None): @@ -1325,28 +1575,26 @@ class InterfaceUtil(object): :returns: Processed vxlan_gpe interface dump. :rtype: dict """ - if vxlan_dump['is_ipv6']: - vxlan_dump['local'] = \ - ip_address(unicode(vxlan_dump['local'])) - vxlan_dump['remote'] = \ - ip_address(unicode(vxlan_dump['remote'])) + if vxlan_dump[u"is_ipv6"]: + vxlan_dump[u"local"] = ip_address(vxlan_dump[u"local"]) + vxlan_dump[u"remote"] = ip_address(vxlan_dump[u"remote"]) else: - vxlan_dump['local'] = \ - ip_address(unicode(vxlan_dump['local'][0:4])) - vxlan_dump['remote'] = \ - ip_address(unicode(vxlan_dump['remote'][0:4])) + vxlan_dump[u"local"] = ip_address(vxlan_dump[u"local"][0:4]) + vxlan_dump[u"remote"] = ip_address(vxlan_dump[u"remote"][0:4]) return vxlan_dump if interface_name is not None: sw_if_index = InterfaceUtil.get_interface_index( - node, interface_name) + node, interface_name + ) else: sw_if_index = int(Constants.BITWISE_NON_ZERO) - cmd = 'vxlan_gpe_tunnel_dump' - args = dict(sw_if_index=sw_if_index) - err_msg = 'Failed to get VXLAN-GPE dump on host {host}'.format( - host=node['host']) + cmd = u"vxlan_gpe_tunnel_dump" + args = dict( + sw_if_index=sw_if_index + ) + err_msg = f"Failed to get VXLAN-GPE dump on host {node[u'host']}" with PapiSocketExecutor(node) as papi_exec: details = papi_exec.add(cmd, **args).get_details(err_msg) @@ -1354,12 +1602,11 @@ class InterfaceUtil(object): for dump in details: if interface_name is None: data.append(process_vxlan_gpe_dump(dump)) - elif dump['sw_if_index'] == sw_if_index: + elif dump[u"sw_if_index"] == sw_if_index: data = process_vxlan_gpe_dump(dump) break - logger.debug('VXLAN-GPE data:\n{vxlan_gpe_data}'.format( - vxlan_gpe_data=data)) + logger.debug(f"VXLAN-GPE data:\n{data}") return data @staticmethod @@ -1375,19 +1622,19 @@ class InterfaceUtil(object): :type table_id: int :type ipv6: bool """ - cmd = 'sw_interface_set_table' + cmd = u"sw_interface_set_table" args = dict( sw_if_index=InterfaceUtil.get_interface_index(node, interface), - is_ipv6=1 if ipv6 else 0, - vrf_id=int(table_id)) - err_msg = 'Failed to assign interface {ifc} to FIB table'.format( - ifc=interface) + is_ipv6=ipv6, + vrf_id=int(table_id) + ) + err_msg = f"Failed to assign interface {interface} to FIB table" with PapiSocketExecutor(node) as papi_exec: papi_exec.add(cmd, **args).get_reply(err_msg) @staticmethod - def set_linux_interface_mac(node, interface, mac, namespace=None, - vf_id=None): + def set_linux_interface_mac( + node, interface, mac, namespace=None, vf_id=None): """Set MAC address for interface in linux. :param node: Node where to execute command. @@ -1401,17 +1648,39 @@ class InterfaceUtil(object): :type namespace: str :type vf_id: int """ - mac_str = 'vf {vf_id} mac {mac}'.format(vf_id=vf_id, mac=mac) \ - if vf_id is not None else 'address {mac}'.format(mac=mac) - ns_str = 'ip netns exec {ns}'.format(ns=namespace) if namespace else '' + mac_str = f"vf {vf_id} mac {mac}" if vf_id is not None \ + else f"address {mac}" + ns_str = f"ip netns exec {namespace}" if namespace else u"" + + cmd = f"{ns_str} ip link set {interface} {mac_str}" + exec_cmd_no_error(node, cmd, sudo=True) + + @staticmethod + def set_linux_interface_promisc( + node, interface, namespace=None, vf_id=None, state=u"on"): + """Set promisc state for interface in linux. + + :param node: Node where to execute command. + :param interface: Interface in namespace. + :param namespace: Exec command in namespace. (Optional, Default: None) + :param vf_id: Virtual Function id. (Optional, Default: None) + :param state: State of feature. (Optional, Default: on) + :type node: dict + :type interface: str + :type namespace: str + :type vf_id: int + :type state: str + """ + promisc_str = f"vf {vf_id} promisc {state}" if vf_id is not None \ + else f"promisc {state}" + ns_str = f"ip netns exec {namespace}" if namespace else u"" - cmd = ('{ns} ip link set {interface} {mac}'. - format(ns=ns_str, interface=interface, mac=mac_str)) + cmd = f"{ns_str} ip link set dev {interface} {promisc_str}" exec_cmd_no_error(node, cmd, sudo=True) @staticmethod - def set_linux_interface_trust_on(node, interface, namespace=None, - vf_id=None): + def set_linux_interface_trust_on( + node, interface, namespace=None, vf_id=None): """Set trust on (promisc) for interface in linux. :param node: Node where to execute command. @@ -1423,17 +1692,15 @@ class InterfaceUtil(object): :type namespace: str :type vf_id: int """ - trust_str = 'vf {vf_id} trust on'.format(vf_id=vf_id) \ - if vf_id is not None else 'trust on' - ns_str = 'ip netns exec {ns}'.format(ns=namespace) if namespace else '' + trust_str = f"vf {vf_id} trust on" if vf_id is not None else u"trust on" + ns_str = f"ip netns exec {namespace}" if namespace else u"" - cmd = ('{ns} ip link set dev {interface} {trust}'. - format(ns=ns_str, interface=interface, trust=trust_str)) + cmd = f"{ns_str} ip link set dev {interface} {trust_str}" exec_cmd_no_error(node, cmd, sudo=True) @staticmethod - def set_linux_interface_spoof_off(node, interface, namespace=None, - vf_id=None): + def set_linux_interface_spoof_off( + node, interface, namespace=None, vf_id=None): """Set spoof off for interface in linux. :param node: Node where to execute command. @@ -1445,22 +1712,83 @@ class InterfaceUtil(object): :type namespace: str :type vf_id: int """ - spoof_str = 'vf {vf_id} spoof off'.format(vf_id=vf_id) \ - if vf_id is not None else 'spoof off' - ns_str = 'ip netns exec {ns}'.format(ns=namespace) if namespace else '' + spoof_str = f"vf {vf_id} spoof off" if vf_id is not None \ + else u"spoof off" + ns_str = f"ip netns exec {namespace}" if namespace else u"" + + cmd = f"{ns_str} ip link set dev {interface} {spoof_str}" + exec_cmd_no_error(node, cmd, sudo=True) + + @staticmethod + def set_linux_interface_state( + node, interface, namespace=None, state=u"up"): + """Set operational state for interface in linux. + + :param node: Node where to execute command. + :param interface: Interface in namespace. + :param namespace: Execute command in namespace. Optional + :param state: Up/Down. + :type node: dict + :type interface: str + :type namespace: str + :type state: str + """ + ns_str = f"ip netns exec {namespace}" if namespace else u"" - cmd = ('{ns} ip link set dev {interface} {spoof}'. - format(ns=ns_str, interface=interface, spoof=spoof_str)) + cmd = f"{ns_str} ip link set dev {interface} {state}" exec_cmd_no_error(node, cmd, sudo=True) @staticmethod - def init_avf_interface(node, ifc_key, numvfs=1, osi_layer='L2'): - """Init PCI device by creating VFs and bind them to vfio-pci for AVF - driver testing on DUT. + def init_interface(node, ifc_key, driver, numvfs=0, osi_layer=u"L2"): + """Init PCI device. Check driver compatibility and bind to proper + drivers. Optionally create NIC VFs. + + :param node: DUT node. + :param ifc_key: Interface key from topology file. + :param driver: Base driver to use. + :param numvfs: Number of VIFs to initialize, 0 - disable the VIFs. + :param osi_layer: OSI Layer type to initialize TG with. + Default value "L2" sets linux interface spoof off. + :type node: dict + :type ifc_key: str + :type driver: str + :type numvfs: int + :type osi_layer: str + :returns: Virtual Function topology interface keys. + :rtype: list + :raises RuntimeError: If a reason preventing initialization is found. + """ + kernel_driver = Topology.get_interface_driver(node, ifc_key) + vf_keys = [] + if driver == u"avf": + if kernel_driver not in ( + u"ice", u"iavf", u"i40e", u"i40evf"): + raise RuntimeError( + f"AVF needs ice or i40e compatible driver, not " + f"{kernel_driver} at node {node[u'host']} ifc {ifc_key}" + ) + vf_keys = InterfaceUtil.init_generic_interface( + node, ifc_key, numvfs=numvfs, osi_layer=osi_layer + ) + elif driver == u"af_xdp": + if kernel_driver not in ( + u"ice", u"iavf", u"i40e", u"i40evf", u"mlx5_core"): + raise RuntimeError( + f"AF_XDP needs ice or i40e or rdma compatible driver, not " + f"{kernel_driver} at node {node[u'host']} ifc {ifc_key}" + ) + vf_keys = InterfaceUtil.init_generic_interface( + node, ifc_key, numvfs=numvfs, osi_layer=osi_layer + ) + return vf_keys + + @staticmethod + def init_generic_interface(node, ifc_key, numvfs=0, osi_layer=u"L2"): + """Init PCI device. Bind to proper drivers. Optionally create NIC VFs. :param node: DUT node. :param ifc_key: Interface key from topology file. - :param numvfs: Number of VFs to initialize, 0 - disable the VFs. + :param numvfs: Number of VIFs to initialize, 0 - disable the VIFs. :param osi_layer: OSI Layer type to initialize TG with. Default value "L2" sets linux interface spoof off. :type node: dict @@ -1476,13 +1804,9 @@ class InterfaceUtil(object): pf_mac_addr = Topology.get_interface_mac(node, ifc_key).split(":") uio_driver = Topology.get_uio_driver(node) kernel_driver = Topology.get_interface_driver(node, ifc_key) - if kernel_driver not in ("i40e", "i40evf"): - raise RuntimeError( - "AVF needs i40e-compatible driver, not {driver} at node {host}" - " ifc {ifc}".format( - driver=kernel_driver, host=node["host"], ifc=ifc_key)) current_driver = DUTSetup.get_pci_dev_driver( - node, pf_pci_addr.replace(':', r'\:')) + node, pf_pci_addr.replace(u":", r"\:")) + pf_dev = f"`basename /sys/bus/pci/devices/{pf_pci_addr}/net/*`" VPPUtil.stop_vpp_service(node) if current_driver != kernel_driver: @@ -1497,34 +1821,50 @@ class InterfaceUtil(object): # Initialize PCI VFs. DUTSetup.set_sriov_numvfs(node, pf_pci_addr, numvfs) + if not numvfs: + if osi_layer == u"L2": + InterfaceUtil.set_linux_interface_promisc(node, pf_dev) + vf_ifc_keys = [] # Set MAC address and bind each virtual function to uio driver. for vf_id in range(numvfs): - vf_mac_addr = ":".join([pf_mac_addr[0], pf_mac_addr[2], - pf_mac_addr[3], pf_mac_addr[4], - pf_mac_addr[5], "{:02x}".format(vf_id)]) - - pf_dev = '`basename /sys/bus/pci/devices/{pci}/net/*`'.\ - format(pci=pf_pci_addr) - InterfaceUtil.set_linux_interface_trust_on(node, pf_dev, - vf_id=vf_id) - if osi_layer == 'L2': - InterfaceUtil.set_linux_interface_spoof_off(node, pf_dev, - vf_id=vf_id) - InterfaceUtil.set_linux_interface_mac(node, pf_dev, vf_mac_addr, - vf_id=vf_id) + vf_mac_addr = u":".join( + [pf_mac_addr[0], pf_mac_addr[2], pf_mac_addr[3], pf_mac_addr[4], + pf_mac_addr[5], f"{vf_id:02x}" + ] + ) + + InterfaceUtil.set_linux_interface_trust_on( + node, pf_dev, vf_id=vf_id + ) + if osi_layer == u"L2": + InterfaceUtil.set_linux_interface_spoof_off( + node, pf_dev, vf_id=vf_id + ) + InterfaceUtil.set_linux_interface_mac( + node, pf_dev, vf_mac_addr, vf_id=vf_id + ) + InterfaceUtil.set_linux_interface_state( + node, pf_dev, state=u"up" + ) DUTSetup.pci_vf_driver_unbind(node, pf_pci_addr, vf_id) DUTSetup.pci_vf_driver_bind(node, pf_pci_addr, vf_id, uio_driver) # Add newly created ports into topology file - vf_ifc_name = '{pf_if_key}_vf'.format(pf_if_key=ifc_key) + vf_ifc_name = f"{ifc_key}_vif" vf_pci_addr = DUTSetup.get_virtfn_pci_addr(node, pf_pci_addr, vf_id) vf_ifc_key = Topology.add_new_port(node, vf_ifc_name) - Topology.update_interface_name(node, vf_ifc_key, - vf_ifc_name+str(vf_id+1)) + Topology.update_interface_name( + node, vf_ifc_key, vf_ifc_name+str(vf_id+1) + ) Topology.update_interface_mac_address(node, vf_ifc_key, vf_mac_addr) Topology.update_interface_pci_address(node, vf_ifc_key, vf_pci_addr) + Topology.set_interface_numa_node( + node, vf_ifc_key, Topology.get_interface_numa_node( + node, ifc_key + ) + ) vf_ifc_keys.append(vf_ifc_key) return vf_ifc_keys @@ -1538,19 +1878,31 @@ class InterfaceUtil(object): :returns: Thread mapping information as a list of dictionaries. :rtype: list """ - cmd = 'sw_interface_rx_placement_dump' - err_msg = "Failed to run '{cmd}' PAPI command on host {host}!".format( - cmd=cmd, host=node['host']) + cmd = u"sw_interface_rx_placement_dump" + err_msg = f"Failed to run '{cmd}' PAPI command on host {node[u'host']}!" with PapiSocketExecutor(node) as papi_exec: - for ifc in node['interfaces'].values(): - if ifc['vpp_sw_index'] is not None: - papi_exec.add(cmd, sw_if_index=ifc['vpp_sw_index']) + for ifc in node[u"interfaces"].values(): + if ifc[u"vpp_sw_index"] is not None: + papi_exec.add(cmd, sw_if_index=ifc[u"vpp_sw_index"]) details = papi_exec.get_details(err_msg) - return sorted(details, key=lambda k: k['sw_if_index']) + return sorted(details, key=lambda k: k[u"sw_if_index"]) @staticmethod - def vpp_sw_interface_set_rx_placement(node, sw_if_index, queue_id, - worker_id): + def vpp_sw_interface_rx_placement_dump_on_all_duts(nodes): + """Dump VPP interface RX placement on all given nodes. + + :param nodes: Nodes to run command on. + :type nodes: dict + :returns: Thread mapping information as a list of dictionaries. + :rtype: list + """ + for node in nodes.values(): + if node[u"type"] == NodeType.DUT: + InterfaceUtil.vpp_sw_interface_rx_placement_dump(node) + + @staticmethod + def vpp_sw_interface_set_rx_placement( + node, sw_if_index, queue_id, worker_id): """Set interface RX placement to worker on node. :param node: Node to run command on. @@ -1564,47 +1916,78 @@ class InterfaceUtil(object): :raises RuntimeError: If failed to run command on host or if no API reply received. """ - cmd = 'sw_interface_set_rx_placement' - err_msg = "Failed to set interface RX placement to worker on host " \ - "{host}!".format(host=node['host']) - args = dict(sw_if_index=sw_if_index, queue_id=queue_id, - worker_id=worker_id) + cmd = u"sw_interface_set_rx_placement" + err_msg = f"Failed to set interface RX placement to worker " \ + f"on host {node[u'host']}!" + args = dict( + sw_if_index=sw_if_index, + queue_id=queue_id, + worker_id=worker_id, + is_main=False + ) with PapiSocketExecutor(node) as papi_exec: papi_exec.add(cmd, **args).get_reply(err_msg) @staticmethod - def vpp_round_robin_rx_placement(node, prefix): + def vpp_round_robin_rx_placement( + node, prefix, dp_worker_limit=None): """Set Round Robin interface RX placement on all worker threads on node. + If specified, dp_core_limit limits the number of physical cores used + for data plane I/O work. Other cores are presumed to do something else, + e.g. asynchronous crypto processing. + None means all workers are used for data plane work. + Note this keyword specifies workers, not cores. + :param node: Topology nodes. :param prefix: Interface name prefix. + :param dp_worker_limit: How many cores for data plane work. :type node: dict :type prefix: str + :type dp_worker_limit: Optional[int] """ worker_id = 0 worker_cnt = len(VPPUtil.vpp_show_threads(node)) - 1 + if dp_worker_limit is not None: + worker_cnt = min(worker_cnt, dp_worker_limit) if not worker_cnt: return for placement in InterfaceUtil.vpp_sw_interface_rx_placement_dump(node): - for interface in node['interfaces'].values(): - if placement['sw_if_index'] == interface['vpp_sw_index'] \ - and prefix in interface['name']: + for interface in node[u"interfaces"].values(): + if placement[u"sw_if_index"] == interface[u"vpp_sw_index"] \ + and prefix in interface[u"name"]: InterfaceUtil.vpp_sw_interface_set_rx_placement( - node, placement['sw_if_index'], placement['queue_id'], - worker_id % worker_cnt) + node, placement[u"sw_if_index"], placement[u"queue_id"], + worker_id % worker_cnt + ) worker_id += 1 @staticmethod - def vpp_round_robin_rx_placement_on_all_duts(nodes, prefix): + def vpp_round_robin_rx_placement_on_all_duts( + nodes, prefix, dp_core_limit=None): """Set Round Robin interface RX placement on all worker threads on all DUTs. + If specified, dp_core_limit limits the number of physical cores used + for data plane I/O work. Other cores are presumed to do something else, + e.g. asynchronous crypto processing. + None means all cores are used for data plane work. + Note this keyword specifies cores, not workers. + :param nodes: Topology nodes. :param prefix: Interface name prefix. + :param dp_worker_limit: How many cores for data plane work. :type nodes: dict :type prefix: str + :type dp_worker_limit: Optional[int] """ for node in nodes.values(): - if node['type'] == NodeType.DUT: - InterfaceUtil.vpp_round_robin_rx_placement(node, prefix) + if node[u"type"] == NodeType.DUT: + dp_worker_limit = CpuUtils.worker_count_from_cores_and_smt( + phy_cores=dp_core_limit, + smt_used=CpuUtils.is_smt_enabled(node[u"cpuinfo"]), + ) + InterfaceUtil.vpp_round_robin_rx_placement( + node, prefix, dp_worker_limit + )