X-Git-Url: https://gerrit.fd.io/r/gitweb?p=csit.git;a=blobdiff_plain;f=resources%2Flibraries%2Fpython%2FTrafficGenerator.py;h=455a21ebb8401a1f2355db8f843f335a81a9cce3;hp=ee65c1bf2afc08444136d32592a853cdb33b7c6a;hb=500dba02e62ba24f94972a0ef8f023418a7eee09;hpb=a9f251c649a5dea7428a43dc24380077a72dacba diff --git a/resources/libraries/python/TrafficGenerator.py b/resources/libraries/python/TrafficGenerator.py index ee65c1bf2a..455a21ebb8 100644 --- a/resources/libraries/python/TrafficGenerator.py +++ b/resources/libraries/python/TrafficGenerator.py @@ -1,4 +1,4 @@ -# Copyright (c) 2018 Cisco and/or its affiliates. +# Copyright (c) 2021 Cisco and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: @@ -13,45 +13,74 @@ """Performance testing traffic generator library.""" +import time + from robot.api import logger from robot.libraries.BuiltIn import BuiltIn +from .Constants import Constants +from .CpuUtils import CpuUtils from .DropRateSearch import DropRateSearch -from .constants import Constants -from .ssh import SSH +from .MLRsearch.AbstractMeasurer import AbstractMeasurer +from .MLRsearch.MultipleLossRatioSearch import MultipleLossRatioSearch +from .MLRsearch.ReceiveRateMeasurement import ReceiveRateMeasurement +from .PLRsearch.PLRsearch import PLRsearch +from .OptionString import OptionString +from .ssh import exec_cmd_no_error, exec_cmd from .topology import NodeType from .topology import NodeSubTypeTG from .topology import Topology -from .search.AbstractMeasurer import AbstractMeasurer -from .search.MultipleLossRatioSearch import MultipleLossRatioSearch -from .search.ReceiveRateMeasurement import ReceiveRateMeasurement -__all__ = ['TGDropRateSearchImpl', 'TrafficGenerator', 'OptimizedSearch'] +__all__ = [u"TGDropRateSearchImpl", u"TrafficGenerator", u"OptimizedSearch"] + + +def check_subtype(node): + """Return supported subtype of given node, or raise an exception. + + Currently only one subtype is supported, + but we want our code to be ready for other ones. + + :param node: Topology node to check. Can be None. + :type node: dict or NoneType + :returns: Subtype detected. + :rtype: NodeSubTypeTG + :raises RuntimeError: If node is not supported, message explains how. + """ + if node.get(u"type") is None: + msg = u"Node type is not defined" + elif node[u"type"] != NodeType.TG: + msg = f"Node type is {node[u'type']!r}, not a TG" + elif node.get(u"subtype") is None: + msg = u"TG subtype is not defined" + elif node[u"subtype"] != NodeSubTypeTG.TREX: + msg = f"TG subtype {node[u'subtype']!r} is not supported" + else: + return NodeSubTypeTG.TREX + raise RuntimeError(msg) class TGDropRateSearchImpl(DropRateSearch): """Drop Rate Search implementation.""" - def __init__(self): - super(TGDropRateSearchImpl, self).__init__() + # def __init__(self): + # super(TGDropRateSearchImpl, self).__init__() - def measure_loss(self, rate, frame_size, loss_acceptance, - loss_acceptance_type, traffic_type, skip_warmup=False): + def measure_loss( + self, rate, frame_size, loss_acceptance, loss_acceptance_type, + traffic_profile): """Runs the traffic and evaluate the measured results. :param rate: Offered traffic load. :param frame_size: Size of frame. :param loss_acceptance: Permitted drop ratio or frames count. :param loss_acceptance_type: Type of permitted loss. - :param traffic_type: Module name as a traffic type identifier. - See resources/traffic_profiles/trex for implemented modules. - :param skip_warmup: Start TRex without warmup traffic if true. + :param traffic_profile: Module name as a traffic profile identifier. + See GPL/traffic_profiles/trex for implemented modules. :type rate: float :type frame_size: str :type loss_acceptance: float :type loss_acceptance_type: LossAcceptanceType - :type traffic_type: str - :type skip_warmup: bool + :type traffic_profile: str :returns: Drop threshold exceeded? (True/False) :rtype: bool :raises NotImplementedError: If TG is not supported. @@ -60,35 +89,23 @@ class TGDropRateSearchImpl(DropRateSearch): # we need instance of TrafficGenerator instantiated by Robot Framework # to be able to use trex_stl-*() tg_instance = BuiltIn().get_library_instance( - 'resources.libraries.python.TrafficGenerator') - - if tg_instance.node['subtype'] is None: - raise RuntimeError('TG subtype not defined') - elif tg_instance.node['subtype'] == NodeSubTypeTG.TREX: + u"resources.libraries.python.TrafficGenerator" + ) + subtype = check_subtype(tg_instance.node) + if subtype == NodeSubTypeTG.TREX: unit_rate = str(rate) + self.get_rate_type_str() - if skip_warmup: - tg_instance.trex_stl_start_remote_exec(self.get_duration(), - unit_rate, frame_size, - traffic_type, - warmup_time=0.0) - else: - tg_instance.trex_stl_start_remote_exec(self.get_duration(), - unit_rate, frame_size, - traffic_type) + tg_instance.trex_stl_start_remote_exec( + self.get_duration(), unit_rate, frame_size, traffic_profile + ) loss = tg_instance.get_loss() sent = tg_instance.get_sent() if self.loss_acceptance_type_is_percentage(): loss = (float(loss) / float(sent)) * 100 - - logger.trace("comparing: {} < {} {}".format(loss, - loss_acceptance, - loss_acceptance_type)) - if float(loss) > float(loss_acceptance): - return False - else: - return True - else: - raise NotImplementedError("TG subtype not supported") + logger.trace( + f"comparing: {loss} < {loss_acceptance} {loss_acceptance_type}" + ) + return float(loss) <= float(loss_acceptance) + return False def get_latency(self): """Returns min/avg/max latency. @@ -97,34 +114,71 @@ class TGDropRateSearchImpl(DropRateSearch): :rtype: list """ tg_instance = BuiltIn().get_library_instance( - 'resources.libraries.python.TrafficGenerator') + u"resources.libraries.python.TrafficGenerator" + ) return tg_instance.get_latency_int() -class TrafficGenerator(AbstractMeasurer): - """Traffic Generator. +class TrexMode: + """Defines mode of T-Rex traffic generator.""" + # Advanced stateful mode + ASTF = u"ASTF" + # Stateless mode + STL = u"STL" - FIXME: Describe API.""" - # TODO: Decrease friction between various search and rate provider APIs. - # FIXME: Remove "trex" from lines which could work with other TGs. +# TODO: Pylint says too-many-instance-attributes. +class TrafficGenerator(AbstractMeasurer): + """Traffic Generator.""" - # use one instance of TrafficGenerator for all tests in test suite - ROBOT_LIBRARY_SCOPE = 'TEST SUITE' + # TODO: Remove "trex" from lines which could work with other TGs. + + # Use one instance of TrafficGenerator for all tests in test suite + ROBOT_LIBRARY_SCOPE = u"TEST SUITE" def __init__(self): + # TODO: Separate into few dataclasses/dicts. + # Pylint dislikes large unstructured state, and it is right. + self._node = None + self._mode = None + # TG interface order mapping + self._ifaces_reordered = False + # Result holding fields, to be removed. self._result = None self._loss = None self._sent = None self._latency = None self._received = None - self._node = None - # T-REX interface order mapping - self._ifaces_reordered = False - # Parameters not given by measure(). + self._approximated_rate = None + self._approximated_duration = None + self._l7_data = None + # Measurement input fields, needed for async stop result. + self._start_time = None + self._stop_time = None + self._rate = None + self._target_duration = None + self._duration = None + # Other input parameters, not knowable from measure() signature. self.frame_size = None - self.traffic_type = None - self.warmup_time = None + self.traffic_profile = None + self.traffic_directions = None + self.negative_loss = None + self.use_latency = None + self.ppta = None + self.resetter = None + self.transaction_scale = None + self.transaction_duration = None + self.sleep_till_duration = None + self.transaction_type = None + self.duration_limit = None + self.ramp_up_start = None + self.ramp_up_stop = None + self.ramp_up_rate = None + self.ramp_up_duration = None + self.state_timeout = None + # Transient data needed for async measurements. + self._xstats = (None, None) + # TODO: Rename "xstats" to something opaque, so T-Rex is not privileged? @property def node(self): @@ -167,13 +221,45 @@ class TrafficGenerator(AbstractMeasurer): """ return self._latency - def initialize_traffic_generator(self, tg_node, tg_if1, tg_if2, - tg_if1_adj_node, tg_if1_adj_if, - tg_if2_adj_node, tg_if2_adj_if, - test_type, - tg_if1_dst_mac=None, tg_if2_dst_mac=None): + def get_approximated_rate(self): + """Return approximated rate computed as ratio of transmitted packets + over duration of trial. + + :returns: Approximated rate. + :rtype: str + """ + return self._approximated_rate + + def get_l7_data(self): + """Return L7 data. + + :returns: Number of received packets. + :rtype: dict + """ + return self._l7_data + + def check_mode(self, expected_mode): + """Check TG mode. + + :param expected_mode: Expected traffic generator mode. + :type expected_mode: object + :raises RuntimeError: In case of unexpected TG mode. + """ + if self._mode == expected_mode: + return + raise RuntimeError( + f"{self._node[u'subtype']} not running in {expected_mode} mode!" + ) + + # TODO: pylint says disable=too-many-locals. + def initialize_traffic_generator( + self, tg_node, tg_if1, tg_if2, tg_if1_adj_node, tg_if1_adj_if, + tg_if2_adj_node, tg_if2_adj_if, osi_layer, tg_if1_dst_mac=None, + tg_if2_dst_mac=None): """TG initialization. + TODO: Document why do we need (and how do we use) _ifaces_reordered. + :param tg_node: Traffic generator node. :param tg_if1: TG - name of first interface. :param tg_if2: TG - name of second interface. @@ -181,7 +267,7 @@ class TrafficGenerator(AbstractMeasurer): :param tg_if1_adj_if: TG if1 adjecent interface. :param tg_if2_adj_node: TG if2 adjecent node. :param tg_if2_adj_if: TG if2 adjecent interface. - :param test_type: 'L2', 'L3' or 'L7' - OSI Layer testing type. + :param osi_layer: 'L2', 'L3' or 'L7' - OSI Layer testing type. :param tg_if1_dst_mac: Interface 1 destination MAC address. :param tg_if2_dst_mac: Interface 2 destination MAC address. :type tg_node: dict @@ -191,159 +277,168 @@ class TrafficGenerator(AbstractMeasurer): :type tg_if1_adj_if: str :type tg_if2_adj_node: dict :type tg_if2_adj_if: str - :type test_type: str + :type osi_layer: str :type tg_if1_dst_mac: str :type tg_if2_dst_mac: str :returns: nothing :raises RuntimeError: In case of issue during initialization. """ - if tg_node['type'] != NodeType.TG: - raise RuntimeError('Node type is not a TG') - self._node = tg_node - - if tg_node['subtype'] == NodeSubTypeTG.TREX: - ssh = SSH() - ssh.connect(tg_node) - - (ret, _, _) = ssh.exec_command( - "sudo -E sh -c '{0}/resources/tools/trex/" - "trex_installer.sh {1}'".format(Constants.REMOTE_FW_DIR, - Constants.TREX_INSTALL_VERSION), - timeout=1800) - if int(ret) != 0: - raise RuntimeError('TRex installation failed.') - - if1_pci = Topology().get_interface_pci_addr(tg_node, tg_if1) - if2_pci = Topology().get_interface_pci_addr(tg_node, tg_if2) - if1_addr = Topology().get_interface_mac(tg_node, tg_if1) - if2_addr = Topology().get_interface_mac(tg_node, tg_if2) - - if test_type == 'L2': - if1_adj_addr = if2_addr - if2_adj_addr = if1_addr - elif test_type == 'L3': - if1_adj_addr = Topology().get_interface_mac(tg_if1_adj_node, - tg_if1_adj_if) - if2_adj_addr = Topology().get_interface_mac(tg_if2_adj_node, - tg_if2_adj_if) - elif test_type == 'L7': - if1_addr = Topology().get_interface_ip4(tg_node, tg_if1) - if2_addr = Topology().get_interface_ip4(tg_node, tg_if2) - if1_adj_addr = Topology().get_interface_ip4(tg_if1_adj_node, - tg_if1_adj_if) - if2_adj_addr = Topology().get_interface_ip4(tg_if2_adj_node, - tg_if2_adj_if) + subtype = check_subtype(tg_node) + if subtype == NodeSubTypeTG.TREX: + self._node = tg_node + self._mode = TrexMode.ASTF if osi_layer == u"L7" else TrexMode.STL + if1 = dict() + if2 = dict() + if1[u"pci"] = Topology().get_interface_pci_addr(self._node, tg_if1) + if2[u"pci"] = Topology().get_interface_pci_addr(self._node, tg_if2) + if1[u"addr"] = Topology().get_interface_mac(self._node, tg_if1) + if2[u"addr"] = Topology().get_interface_mac(self._node, tg_if2) + + if osi_layer == u"L2": + if1[u"adj_addr"] = if2[u"addr"] + if2[u"adj_addr"] = if1[u"addr"] + elif osi_layer in (u"L3", u"L7"): + if1[u"adj_addr"] = Topology().get_interface_mac( + tg_if1_adj_node, tg_if1_adj_if + ) + if2[u"adj_addr"] = Topology().get_interface_mac( + tg_if2_adj_node, tg_if2_adj_if + ) else: - raise ValueError("Unknown Test Type") + raise ValueError(u"Unknown OSI layer!") # in case of switched environment we can override MAC addresses if tg_if1_dst_mac is not None and tg_if2_dst_mac is not None: - if1_adj_addr = tg_if1_dst_mac - if2_adj_addr = tg_if2_dst_mac + if1[u"adj_addr"] = tg_if1_dst_mac + if2[u"adj_addr"] = tg_if2_dst_mac - if min(if1_pci, if2_pci) != if1_pci: - if1_pci, if2_pci = if2_pci, if1_pci - if1_addr, if2_addr = if2_addr, if1_addr - if1_adj_addr, if2_adj_addr = if2_adj_addr, if1_adj_addr + if min(if1[u"pci"], if2[u"pci"]) != if1[u"pci"]: + if1, if2 = if2, if1 self._ifaces_reordered = True - if test_type == 'L2' or test_type == 'L3': - (ret, _, _) = ssh.exec_command( - "sudo sh -c 'cat << EOF > /etc/trex_cfg.yaml\n" - "- port_limit: 2\n" - " version: 2\n" - " interfaces: [\"{0}\",\"{1}\"]\n" - " port_info:\n" - " - dest_mac: [{2}]\n" - " src_mac: [{3}]\n" - " - dest_mac: [{4}]\n" - " src_mac: [{5}]\n" - "EOF'"\ - .format(if1_pci, if2_pci, - "0x"+if1_adj_addr.replace(":", ",0x"), - "0x"+if1_addr.replace(":", ",0x"), - "0x"+if2_adj_addr.replace(":", ",0x"), - "0x"+if2_addr.replace(":", ",0x"))) - elif test_type == 'L7': - (ret, _, _) = ssh.exec_command( - "sudo sh -c 'cat << EOF > /etc/trex_cfg.yaml\n" - "- port_limit: 2\n" - " version: 2\n" - " interfaces: [\"{0}\",\"{1}\"]\n" - " port_info:\n" - " - ip: [{2}]\n" - " default_gw: [{3}]\n" - " - ip: [{4}]\n" - " default_gw: [{5}]\n" - "EOF'"\ - .format(if1_pci, if2_pci, - if1_addr, if1_adj_addr, - if2_addr, if2_adj_addr)) + master_thread_id, latency_thread_id, socket, threads = \ + CpuUtils.get_affinity_trex( + self._node, tg_if1, tg_if2, + tg_dtc=Constants.TREX_CORE_COUNT) + + if osi_layer in (u"L2", u"L3", u"L7"): + exec_cmd_no_error( + self._node, + f"sh -c 'cat << EOF > /etc/trex_cfg.yaml\n" + f"- version: 2\n" + f" c: {len(threads)}\n" + f" limit_memory: {Constants.TREX_LIMIT_MEMORY}\n" + f" interfaces: [\"{if1[u'pci']}\",\"{if2[u'pci']}\"]\n" + f" port_info:\n" + f" - dest_mac: \'{if1[u'adj_addr']}\'\n" + f" src_mac: \'{if1[u'addr']}\'\n" + f" - dest_mac: \'{if2[u'adj_addr']}\'\n" + f" src_mac: \'{if2[u'addr']}\'\n" + f" platform :\n" + f" master_thread_id: {master_thread_id}\n" + f" latency_thread_id: {latency_thread_id}\n" + f" dual_if:\n" + f" - socket: {socket}\n" + f" threads: {threads}\n" + f"EOF'", + sudo=True, message=u"T-Rex config generation!" + ) else: - raise ValueError("Unknown Test Type") - if int(ret) != 0: - raise RuntimeError('TRex config generation error') + raise ValueError(u"Unknown OSI layer!") + + TrafficGenerator.startup_trex( + self._node, osi_layer, subtype=subtype + ) + @staticmethod + def startup_trex(tg_node, osi_layer, subtype=None): + """Startup sequence for the TRex traffic generator. + + :param tg_node: Traffic generator node. + :param osi_layer: 'L2', 'L3' or 'L7' - OSI Layer testing type. + :param subtype: Traffic generator sub-type. + :type tg_node: dict + :type osi_layer: str + :type subtype: NodeSubTypeTG + :raises RuntimeError: If T-Rex startup failed. + :raises ValueError: If OSI layer is not supported. + """ + if not subtype: + subtype = check_subtype(tg_node) + if subtype == NodeSubTypeTG.TREX: for _ in range(0, 3): - # kill TRex only if it is already running - ssh.exec_command( - "sh -c 'pgrep t-rex && sudo pkill t-rex && sleep 3'") - - # configure TRex - (ret, _, _) = ssh.exec_command( - "sh -c 'cd {0}/scripts/ && sudo ./trex-cfg'"\ - .format(Constants.TREX_INSTALL_DIR)) - if int(ret) != 0: - raise RuntimeError('trex-cfg failed') - - # start TRex - if test_type == 'L2' or test_type == 'L3': - (ret, _, _) = ssh.exec_command( - "sh -c 'cd {0}/scripts/ && " - "sudo nohup ./t-rex-64 -i -c 7 --iom 0 > /tmp/trex.log " - "2>&1 &' > /dev/null"\ - .format(Constants.TREX_INSTALL_DIR)) - elif test_type == 'L7': - (ret, _, _) = ssh.exec_command( - "sh -c 'cd {0}/scripts/ && " - "sudo nohup ./t-rex-64 --astf -i -c 7 --iom 0 > " - "/tmp/trex.log 2>&1 &' > /dev/null"\ - .format(Constants.TREX_INSTALL_DIR)) + # Kill TRex only if it is already running. + cmd = u"sh -c \"pgrep t-rex && pkill t-rex && sleep 3 || true\"" + exec_cmd_no_error( + tg_node, cmd, sudo=True, message=u"Kill TRex failed!" + ) + + # Configure TRex. + ports = '' + for port in tg_node[u"interfaces"].values(): + if u'Mellanox' not in port.get(u'model'): + ports += f" {port.get(u'pci_address')}" + + cmd = f"sh -c \"cd {Constants.TREX_INSTALL_DIR}/scripts/ && " \ + f"./dpdk_nic_bind.py -u {ports} || true\"" + exec_cmd_no_error( + tg_node, cmd, sudo=True, + message=u"Unbind PCI ports from driver failed!" + ) + + # Start TRex. + cd_cmd = f"cd '{Constants.TREX_INSTALL_DIR}/scripts/'" + trex_cmd = OptionString([u"nohup", u"./t-rex-64"]) + trex_cmd.add(u"-i") + trex_cmd.add(u"--prefix $(hostname)") + trex_cmd.add(u"--hdrh") + trex_cmd.add(u"--no-scapy-server") + trex_cmd.add_if(u"--astf", osi_layer == u"L7") + # OptionString does not create double space if extra is empty. + trex_cmd.add(f"{Constants.TREX_EXTRA_CMDLINE}") + inner_command = f"{cd_cmd} && {trex_cmd} > /tmp/trex.log 2>&1 &" + cmd = f"sh -c \"{inner_command}\" > /dev/null" + try: + exec_cmd_no_error(tg_node, cmd, sudo=True) + except RuntimeError: + cmd = u"sh -c \"cat /tmp/trex.log\"" + exec_cmd_no_error( + tg_node, cmd, sudo=True, + message=u"Get TRex logs failed!" + ) + raise RuntimeError(u"Start TRex failed!") + + # Test T-Rex API responsiveness. + cmd = f"python3 {Constants.REMOTE_FW_DIR}/GPL/tools/trex/" + if osi_layer in (u"L2", u"L3"): + cmd += u"trex_stl_assert.py" + elif osi_layer == u"L7": + cmd += u"trex_astf_assert.py" else: - raise ValueError("Unknown Test Type") - if int(ret) != 0: - ssh.exec_command("sh -c 'cat /tmp/trex.log'") - raise RuntimeError('t-rex-64 startup failed') - - # get TRex server info - (ret, _, _) = ssh.exec_command( - "sh -c 'sleep 3; " - "{0}/resources/tools/trex/trex_server_info.py'"\ - .format(Constants.REMOTE_FW_DIR), - timeout=120) - if int(ret) == 0: - # If we get info TRex is running - return - # after max retries TRex is still not responding to API - # critical error occurred - raise RuntimeError('t-rex-64 startup failed') + raise ValueError(u"Unknown OSI layer!") + try: + exec_cmd_no_error( + tg_node, cmd, sudo=True, + message=u"T-Rex API is not responding!", retries=20 + ) + except RuntimeError: + continue + return + # After max retries TRex is still not responding to API critical + # error occurred. + exec_cmd(tg_node, u"cat /tmp/trex.log", sudo=True) + raise RuntimeError(u"Start T-Rex failed after multiple retries!") @staticmethod def is_trex_running(node): - """Check if TRex is running using pidof. + """Check if T-Rex is running using pidof. :param node: Traffic generator node. :type node: dict - :returns: True if TRex is running otherwise False. + :returns: True if T-Rex is running otherwise False. :rtype: bool - :raises RuntimeError: If node type is not a TG. """ - if node['type'] != NodeType.TG: - raise RuntimeError('Node type is not a TG') - - ssh = SSH() - ssh.connect(node) - ret, _, _ = ssh.exec_command_sudo("pidof t-rex") + ret, _, _ = exec_cmd(node, u"pgrep t-rex", sudo=True) return bool(int(ret) == 0) @staticmethod @@ -354,164 +449,575 @@ class TrafficGenerator(AbstractMeasurer): :type node: dict :returns: nothing :raises RuntimeError: If node type is not a TG, - or if TRex teardown fails. - """ - if node['type'] != NodeType.TG: - raise RuntimeError('Node type is not a TG') - if node['subtype'] == NodeSubTypeTG.TREX: - ssh = SSH() - ssh.connect(node) - (ret, _, _) = ssh.exec_command( - "sh -c 'sudo pkill t-rex && sleep 3'") - if int(ret) != 0: - raise RuntimeError('pkill t-rex failed') - - @staticmethod - def trex_stl_stop_remote_exec(node): - """Execute script on remote node over ssh to stop running traffic. - - :param node: TRex generator node. + or if T-Rex teardown fails. + """ + subtype = check_subtype(node) + if subtype == NodeSubTypeTG.TREX: + exec_cmd_no_error( + node, + u"sh -c " + u"\"if pgrep t-rex; then sudo pkill t-rex && sleep 3; fi\"", + sudo=False, + message=u"T-Rex kill failed!" + ) + + def trex_astf_stop_remote_exec(self, node): + """Execute T-Rex ASTF script on remote node over ssh to stop running + traffic. + + Internal state is updated with measurement results. + + :param node: T-Rex generator node. + :type node: dict + :raises RuntimeError: If stop traffic script fails. + """ + command_line = OptionString().add(u"python3") + dirname = f"{Constants.REMOTE_FW_DIR}/GPL/tools/trex" + command_line.add(f"'{dirname}/trex_astf_stop.py'") + command_line.change_prefix(u"--") + for index, value in enumerate(self._xstats): + if value is not None: + value = value.replace(u"'", u"\"") + command_line.add_equals(f"xstat{index}", f"'{value}'") + stdout, _ = exec_cmd_no_error( + node, command_line, + message=u"T-Rex ASTF runtime error!" + ) + self._parse_traffic_results(stdout) + + def trex_stl_stop_remote_exec(self, node): + """Execute T-Rex STL script on remote node over ssh to stop running + traffic. + + Internal state is updated with measurement results. + + :param node: T-Rex generator node. :type node: dict - :returns: Nothing :raises RuntimeError: If stop traffic script fails. """ - ssh = SSH() - ssh.connect(node) - - (ret, _, _) = ssh.exec_command( - "sh -c '{}/resources/tools/trex/" - "trex_stateless_stop.py'".format(Constants.REMOTE_FW_DIR)) - - if int(ret) != 0: - raise RuntimeError('TRex stateless runtime error') - - def trex_stl_start_remote_exec(self, duration, rate, framesize, - traffic_type, async_call=False, - latency=True, warmup_time=5.0): - """Execute script on remote node over ssh to start traffic. - - :param duration: Time expresed in seconds for how long to send traffic. - :param rate: Traffic rate expressed with units (pps, %) - :param framesize: L2 frame size to send (without padding and IPG). - :param traffic_type: Module name as a traffic type identifier. - See resources/traffic_profiles/trex for implemented modules. - :param async_call: If enabled then don't wait for all incomming trafic. - :param latency: With latency measurement. - :param warmup_time: Warmup time period. + command_line = OptionString().add(u"python3") + dirname = f"{Constants.REMOTE_FW_DIR}/GPL/tools/trex" + command_line.add(f"'{dirname}/trex_stl_stop.py'") + command_line.change_prefix(u"--") + for index, value in enumerate(self._xstats): + if value is not None: + value = value.replace(u"'", u"\"") + command_line.add_equals(f"xstat{index}", f"'{value}'") + stdout, _ = exec_cmd_no_error( + node, command_line, + message=u"T-Rex STL runtime error!" + ) + self._parse_traffic_results(stdout) + + def stop_traffic_on_tg(self): + """Stop all traffic on TG. + + :returns: Structure containing the result of the measurement. + :rtype: ReceiveRateMeasurement + :raises ValueError: If TG traffic profile is not supported. + """ + subtype = check_subtype(self._node) + if subtype != NodeSubTypeTG.TREX: + raise ValueError(f"Unsupported TG subtype: {subtype!r}") + if u"trex-astf" in self.traffic_profile: + self.trex_astf_stop_remote_exec(self._node) + elif u"trex-stl" in self.traffic_profile: + self.trex_stl_stop_remote_exec(self._node) + else: + raise ValueError(u"Unsupported T-Rex traffic profile!") + self._stop_time = time.monotonic() + + return self._get_measurement_result() + + def _compute_duration(self, duration, multiplier): + """Compute duration for profile driver. + + The final result is influenced by transaction scale and duration limit. + It is assumed a higher level function has already set those to self. + The duration argument is the target value from search point of view, + before the overrides are applied here. + + Minus one (signalling async traffic start) is kept. + + Completeness flag is also included. Duration limited or async trials + are not considered complete for ramp-up purposes. + + :param duration: Time expressed in seconds for how long to send traffic. + :param multiplier: Traffic rate in transactions per second. + :type duration: float + :type multiplier: float + :returns: New duration and whether it was a complete ramp-up candidate. + :rtype: float, bool + """ + if duration < 0.0: + # Keep the async -1. + return duration, False + computed_duration = duration + if self.transaction_scale: + computed_duration = self.transaction_scale / multiplier + # Log the computed duration, + # so we can compare with what telemetry suggests + # the real duration was. + logger.debug(f"Expected duration {computed_duration}") + computed_duration += 0.1115 + if not self.duration_limit: + return computed_duration, True + limited_duration = min(computed_duration, self.duration_limit) + return limited_duration, (limited_duration == computed_duration) + + def trex_astf_start_remote_exec( + self, duration, multiplier, async_call=False): + """Execute T-Rex ASTF script on remote node over ssh to start running + traffic. + + In sync mode, measurement results are stored internally. + In async mode, initial data including xstats are stored internally. + + This method contains the logic to compute duration as maximum time + if transaction_scale is nonzero. + The transaction_scale argument defines (limits) how many transactions + will be started in total. As that amount of transaction can take + considerable time (sometimes due to explicit delays in the profile), + the real time a trial needs to finish is computed here. For now, + in that case the duration argument is ignored, assuming it comes + from ASTF-unaware search algorithm. The overall time a single + transaction needs is given in parameter transaction_duration, + it includes both explicit delays and implicit time it takes + to transfer data (or whatever the transaction does). + + Currently it is observed TRex does not start the ASTF traffic + immediately, an ad-hoc constant is added to the computed duration + to compensate for that. + + If transaction_scale is zero, duration is not recomputed. + It is assumed the subsequent result parsing gets the real duration + if the traffic stops sooner for any reason. + + Currently, it is assumed traffic profile defines a single transaction. + To avoid heavy logic here, the input rate is expected to be in + transactions per second, as that directly translates to TRex multiplier, + (assuming the profile does not override the default cps value of one). + + :param duration: Time expressed in seconds for how long to send traffic. + :param multiplier: Traffic rate in transactions per second. + :param async_call: If enabled then don't wait for all incoming traffic. + :type duration: float + :type multiplier: int + :type async_call: bool + :raises RuntimeError: In case of T-Rex driver issue. + """ + self.check_mode(TrexMode.ASTF) + p_0, p_1 = (1, 0) if self._ifaces_reordered else (0, 1) + if not isinstance(duration, (float, int)): + duration = float(duration) + + # TODO: Refactor the code so duration is computed only once, + # and both the initial and the computed durations are logged. + computed_duration, _ = self._compute_duration(duration, multiplier) + + command_line = OptionString().add(u"python3") + dirname = f"{Constants.REMOTE_FW_DIR}/GPL/tools/trex" + command_line.add(f"'{dirname}/trex_astf_profile.py'") + command_line.change_prefix(u"--") + dirname = f"{Constants.REMOTE_FW_DIR}/GPL/traffic_profiles/trex" + command_line.add_with_value( + u"profile", f"'{dirname}/{self.traffic_profile}.py'" + ) + command_line.add_with_value(u"duration", f"{computed_duration!r}") + command_line.add_with_value(u"frame_size", self.frame_size) + command_line.add_with_value(u"multiplier", multiplier) + command_line.add_with_value(u"port_0", p_0) + command_line.add_with_value(u"port_1", p_1) + command_line.add_with_value( + u"traffic_directions", self.traffic_directions + ) + command_line.add_if(u"async_start", async_call) + command_line.add_if(u"latency", self.use_latency) + command_line.add_if(u"force", Constants.TREX_SEND_FORCE) + + self._start_time = time.monotonic() + self._rate = multiplier + stdout, _ = exec_cmd_no_error( + self._node, command_line, timeout=computed_duration + 10.0, + message=u"T-Rex ASTF runtime error!" + ) + + if async_call: + # no result + self._target_duration = None + self._duration = None + self._received = None + self._sent = None + self._loss = None + self._latency = None + xstats = [None, None] + self._l7_data = dict() + self._l7_data[u"client"] = dict() + self._l7_data[u"client"][u"active_flows"] = None + self._l7_data[u"client"][u"established_flows"] = None + self._l7_data[u"client"][u"traffic_duration"] = None + self._l7_data[u"server"] = dict() + self._l7_data[u"server"][u"active_flows"] = None + self._l7_data[u"server"][u"established_flows"] = None + self._l7_data[u"server"][u"traffic_duration"] = None + if u"udp" in self.traffic_profile: + self._l7_data[u"client"][u"udp"] = dict() + self._l7_data[u"client"][u"udp"][u"connects"] = None + self._l7_data[u"client"][u"udp"][u"closed_flows"] = None + self._l7_data[u"client"][u"udp"][u"err_cwf"] = None + self._l7_data[u"server"][u"udp"] = dict() + self._l7_data[u"server"][u"udp"][u"accepted_flows"] = None + self._l7_data[u"server"][u"udp"][u"closed_flows"] = None + elif u"tcp" in self.traffic_profile: + self._l7_data[u"client"][u"tcp"] = dict() + self._l7_data[u"client"][u"tcp"][u"initiated_flows"] = None + self._l7_data[u"client"][u"tcp"][u"connects"] = None + self._l7_data[u"client"][u"tcp"][u"closed_flows"] = None + self._l7_data[u"client"][u"tcp"][u"connattempt"] = None + self._l7_data[u"server"][u"tcp"] = dict() + self._l7_data[u"server"][u"tcp"][u"accepted_flows"] = None + self._l7_data[u"server"][u"tcp"][u"connects"] = None + self._l7_data[u"server"][u"tcp"][u"closed_flows"] = None + else: + logger.warn(u"Unsupported T-Rex ASTF traffic profile!") + index = 0 + for line in stdout.splitlines(): + if f"Xstats snapshot {index}: " in line: + xstats[index] = line[19:] + index += 1 + if index == 2: + break + self._xstats = tuple(xstats) + else: + self._target_duration = duration + self._duration = computed_duration + self._parse_traffic_results(stdout) + + def trex_stl_start_remote_exec(self, duration, rate, async_call=False): + """Execute T-Rex STL script on remote node over ssh to start running + traffic. + + In sync mode, measurement results are stored internally. + In async mode, initial data including xstats are stored internally. + + Mode-unaware code (e.g. in search algorithms) works with transactions. + To keep the logic simple, multiplier is set to that value. + As bidirectional traffic profiles send packets in both directions, + they are treated as transactions with two packets (one per direction). + + :param duration: Time expressed in seconds for how long to send traffic. + :param rate: Traffic rate in transactions per second. + :param async_call: If enabled then don't wait for all incoming traffic. :type duration: float :type rate: str - :type framesize: str - :type traffic_type: str :type async_call: bool - :type latency: bool - :type warmup_time: float - :returns: Nothing - :raises RuntimeError: In case of TG driver issue. - """ - ssh = SSH() - ssh.connect(self._node) - - _async = "--async" if async_call else "" - _latency = "--latency" if latency else "" - _p0, _p1 = (2, 1) if self._ifaces_reordered else (1, 2) - - profile_path = ("{0}/resources/traffic_profiles/trex/" - "{1}.py".format(Constants.REMOTE_FW_DIR, - traffic_type)) - (ret, stdout, _) = ssh.exec_command( - "sh -c " - "'{0}/resources/tools/trex/trex_stateless_profile.py " - "--profile {1} " - "--duration {2} " - "--frame_size {3} " - "--rate {4} " - "--warmup_time {5} " - "--port_0 {6} " - "--port_1 {7} " - "{8} " # --async - "{9}'". # --latency - format(Constants.REMOTE_FW_DIR, profile_path, duration, framesize, - rate, warmup_time, _p0 - 1, _p1 - 1, _async, _latency), - timeout=float(duration) + 60) - - if int(ret) != 0: - raise RuntimeError('TRex stateless runtime error') - elif async_call: - #no result + :raises RuntimeError: In case of T-Rex driver issue. + """ + self.check_mode(TrexMode.STL) + p_0, p_1 = (1, 0) if self._ifaces_reordered else (0, 1) + if not isinstance(duration, (float, int)): + duration = float(duration) + + # TODO: Refactor the code so duration is computed only once, + # and both the initial and the computed durations are logged. + duration, _ = self._compute_duration(duration=duration, multiplier=rate) + + command_line = OptionString().add(u"python3") + dirname = f"{Constants.REMOTE_FW_DIR}/GPL/tools/trex" + command_line.add(f"'{dirname}/trex_stl_profile.py'") + command_line.change_prefix(u"--") + dirname = f"{Constants.REMOTE_FW_DIR}/GPL/traffic_profiles/trex" + command_line.add_with_value( + u"profile", f"'{dirname}/{self.traffic_profile}.py'" + ) + command_line.add_with_value(u"duration", f"{duration!r}") + command_line.add_with_value(u"frame_size", self.frame_size) + command_line.add_with_value(u"rate", f"{rate!r}") + command_line.add_with_value(u"port_0", p_0) + command_line.add_with_value(u"port_1", p_1) + command_line.add_with_value( + u"traffic_directions", self.traffic_directions + ) + command_line.add_if(u"async_start", async_call) + command_line.add_if(u"latency", self.use_latency) + command_line.add_if(u"force", Constants.TREX_SEND_FORCE) + + # TODO: This is ugly. Handle parsing better. + self._start_time = time.monotonic() + self._rate = float(rate[:-3]) if u"pps" in rate else float(rate) + stdout, _ = exec_cmd_no_error( + self._node, command_line, timeout=int(duration) + 60, + message=u"T-Rex STL runtime error" + ) + + if async_call: + # no result + self._target_duration = None + self._duration = None self._received = None self._sent = None self._loss = None self._latency = None + + xstats = [None, None] + index = 0 + for line in stdout.splitlines(): + if f"Xstats snapshot {index}: " in line: + xstats[index] = line[19:] + index += 1 + if index == 2: + break + self._xstats = tuple(xstats) else: - # last line from console output - line = stdout.splitlines()[-1] + self._target_duration = duration + self._duration = duration + self._parse_traffic_results(stdout) + + def send_traffic_on_tg( + self, + duration, + rate, + frame_size, + traffic_profile, + async_call=False, + ppta=1, + traffic_directions=2, + transaction_duration=0.0, + transaction_scale=0, + transaction_type=u"packet", + duration_limit=0.0, + use_latency=False, + ramp_up_rate=None, + ramp_up_duration=None, + state_timeout=300.0, + ramp_up_only=False, + ): + """Send traffic from all configured interfaces on TG. - self._result = line - logger.info('TrafficGen result: {0}'.format(self._result)) + In async mode, xstats is stored internally, + to enable getting correct result when stopping the traffic. + In both modes, stdout is returned, + but _parse_traffic_results only works in sync output. - self._received = self._result.split(', ')[1].split('=')[1] - self._sent = self._result.split(', ')[2].split('=')[1] - self._loss = self._result.split(', ')[3].split('=')[1] + Note that traffic generator uses DPDK driver which might + reorder port numbers based on wiring and PCI numbering. + This method handles that, so argument values are invariant, + but you can see swapped valued in debug logs. - self._latency = [] - self._latency.append(self._result.split(', ')[4].split('=')[1]) - self._latency.append(self._result.split(', ')[5].split('=')[1]) + When transaction_scale is specified, the duration value is ignored + and the needed time is computed. For cases where this results in + to too long measurement (e.g. teardown trial with small rate), + duration_limit is applied (of non-zero), so the trial is stopped sooner. - def stop_traffic_on_tg(self): - """Stop all traffic on TG. + Bidirectional STL profiles are treated as transactions with two packets. - :returns: Nothing - :raises RuntimeError: If TG is not set. - """ - if self._node is None: - raise RuntimeError("TG is not set") - if self._node['subtype'] == NodeSubTypeTG.TREX: - self.trex_stl_stop_remote_exec(self._node) + The return value is None for async. - def send_traffic_on_tg(self, duration, rate, framesize, - traffic_type, warmup_time=5, async_call=False, - latency=True): + :param duration: Duration of test traffic generation in seconds. + :param rate: Traffic rate in transactions per second. + :param frame_size: Frame size (L2) in Bytes. + :param traffic_profile: Module name as a traffic profile identifier. + See GPL/traffic_profiles/trex for implemented modules. + :param async_call: Async mode. + :param ppta: Packets per transaction, aggregated over directions. + Needed for udp_pps which does not have a good transaction counter, + so we need to compute expected number of packets. + Default: 1. + :param traffic_directions: Traffic is bi- (2) or uni- (1) directional. + Default: 2 + :param transaction_duration: Total expected time to close transaction. + :param transaction_scale: Number of transactions to perform. + 0 (default) means unlimited. + :param transaction_type: An identifier specifying which counters + and formulas to use when computing attempted and failed + transactions. Default: "packet". + :param duration_limit: Zero or maximum limit for computed (or given) + duration. + :param use_latency: Whether to measure latency during the trial. + Default: False. + :param ramp_up_rate: Rate to use in ramp-up trials [pps]. + :param ramp_up_duration: Duration of ramp-up trials [s]. + :param state_timeout: Time of life of DUT state [s]. + :param ramp_up_only: If true, do not perform main trial measurement. + :type duration: float + :type rate: float + :type frame_size: str + :type traffic_profile: str + :type async_call: bool + :type ppta: int + :type traffic_directions: int + :type transaction_duration: float + :type transaction_scale: int + :type transaction_type: str + :type duration_limit: float + :type use_latency: bool + :type ramp_up_rate: float + :type ramp_up_duration: float + :type state_timeout: float + :type ramp_up_only: bool + :returns: TG results. + :rtype: ReceiveRateMeasurement or None + :raises ValueError: If TG traffic profile is not supported. + """ + self.set_rate_provider_defaults( + frame_size=frame_size, + traffic_profile=traffic_profile, + ppta=ppta, + traffic_directions=traffic_directions, + transaction_duration=transaction_duration, + transaction_scale=transaction_scale, + transaction_type=transaction_type, + duration_limit=duration_limit, + use_latency=use_latency, + ramp_up_rate=ramp_up_rate, + ramp_up_duration=ramp_up_duration, + state_timeout=state_timeout, + ) + return self._send_traffic_on_tg_with_ramp_up( + duration=duration, + rate=rate, + async_call=async_call, + ramp_up_only=ramp_up_only, + ) + + def _send_traffic_on_tg_internal( + self, duration, rate, async_call=False): """Send traffic from all configured interfaces on TG. + This is an internal function, it assumes set_rate_provider_defaults + has been called to remember most values. + The reason why need to remember various values is that + the traffic can be asynchronous, and parsing needs those values. + The reason why this is is a separate function from the one + which calls set_rate_provider_defaults is that some search algorithms + need to specify their own values, and we do not want the measure call + to overwrite them with defaults. + + This function is used both for automated ramp-up trials + and for explicitly called trials. + :param duration: Duration of test traffic generation in seconds. - :param rate: Offered load per interface (e.g. 1%, 3gbps, 4mpps, ...). - :param framesize: Frame size (L2) in Bytes. - :param traffic_type: Module name as a traffic type identifier. - See resources/traffic_profiles/trex for implemented modules. - :param warmup_time: Warmup phase in seconds. + :param rate: Traffic rate in transactions per second. :param async_call: Async mode. - :param latency: With latency measurement. - :type duration: str - :type rate: str - :type framesize: str - :type traffic_type: str - :type warmup_time: float + :type duration: float + :type rate: float :type async_call: bool - :type latency: bool - :returns: TG output. - :rtype: str - :raises RuntimeError: If TG is not set, or if node is not TG, - or if subtype is not specified. - :raises NotImplementedError: If TG is not supported. + :returns: TG results. + :rtype: ReceiveRateMeasurement or None + :raises ValueError: If TG traffic profile is not supported. """ + subtype = check_subtype(self._node) + if subtype == NodeSubTypeTG.TREX: + if u"trex-astf" in self.traffic_profile: + self.trex_astf_start_remote_exec( + duration, float(rate), async_call + ) + elif u"trex-stl" in self.traffic_profile: + unit_rate_str = str(rate) + u"pps" + # TODO: Suport transaction_scale et al? + self.trex_stl_start_remote_exec( + duration, unit_rate_str, async_call + ) + else: + raise ValueError(u"Unsupported T-Rex traffic profile!") - node = self._node - if node is None: - raise RuntimeError("TG is not set") + return None if async_call else self._get_measurement_result() - if node['type'] != NodeType.TG: - raise RuntimeError('Node type is not a TG') + def _send_traffic_on_tg_with_ramp_up( + self, duration, rate, async_call=False, ramp_up_only=False): + """Send traffic from all interfaces on TG, maybe after ramp-up. - if node['subtype'] is None: - raise RuntimeError('TG subtype not defined') - elif node['subtype'] == NodeSubTypeTG.TREX: - self.trex_stl_start_remote_exec(duration, rate, framesize, - traffic_type, async_call, latency, - warmup_time=warmup_time) - else: - raise NotImplementedError("TG subtype not supported") + This is an internal function, it assumes set_rate_provider_defaults + has been called to remember most values. + The reason why need to remember various values is that + the traffic can be asynchronous, and parsing needs those values. + The reason why this is a separate function from the one + which calls set_rate_provider_defaults is that some search algorithms + need to specify their own values, and we do not want the measure call + to overwrite them with defaults. + + If ramp-up tracking is detected, a computation is performed, + and if state timeout is near, trial at ramp-up rate and duration + is inserted before the main trial measurement. - return self._result + The ramp_up_only parameter forces a ramp-up without immediate + trial measurement, which is useful in case self remembers + a previous ramp-up trial that belongs to a different test (phase). + + Return None if trial is async or ramp-up only. + + :param duration: Duration of test traffic generation in seconds. + :param rate: Traffic rate in transactions per second. + :param async_call: Async mode. + :param ramp_up_only: If true, do not perform main trial measurement. + :type duration: float + :type rate: float + :type async_call: bool + :type ramp_up_only: bool + :returns: TG results. + :rtype: ReceiveRateMeasurement or None + :raises ValueError: If TG traffic profile is not supported. + """ + complete = False + if self.ramp_up_rate: + # Figure out whether we need to insert a ramp-up trial. + # TODO: Give up on async_call=True? + if ramp_up_only or self.ramp_up_start is None: + # We never ramped up yet (at least not in this test case). + ramp_up_needed = True + else: + # We ramped up before, but maybe it was too long ago. + # Adding a constant overhead to be safe. + time_now = time.monotonic() + 1.0 + computed_duration, complete = self._compute_duration( + duration=duration, + multiplier=rate, + ) + # There are two conditions for inserting ramp-up. + # If early sessions are expiring already, + # or if late sessions are to expire before measurement is over. + ramp_up_start_delay = time_now - self.ramp_up_start + ramp_up_stop_delay = time_now - self.ramp_up_stop + ramp_up_stop_delay += computed_duration + bigger_delay = max(ramp_up_start_delay, ramp_up_stop_delay) + # Final boolean decision. + ramp_up_needed = (bigger_delay >= self.state_timeout) + if ramp_up_needed: + logger.debug( + u"State may time out during next real trial, " + u"inserting a ramp-up trial." + ) + self.ramp_up_start = time.monotonic() + self._send_traffic_on_tg_internal( + duration=self.ramp_up_duration, + rate=self.ramp_up_rate, + async_call=async_call, + ) + self.ramp_up_stop = time.monotonic() + logger.debug(u"Ramp-up done.") + else: + logger.debug( + u"State will probably not time out during next real trial, " + u"no ramp-up trial needed just yet." + ) + if ramp_up_only: + return None + trial_start = time.monotonic() + result = self._send_traffic_on_tg_internal( + duration=duration, + rate=rate, + async_call=async_call, + ) + trial_end = time.monotonic() + if self.ramp_up_rate: + # Optimization: No loss acts as a good ramp-up, if it was complete. + if complete and result is not None and result.loss_count == 0: + logger.debug(u"Good trial acts as a ramp-up") + self.ramp_up_start = trial_start + self.ramp_up_stop = trial_end + else: + logger.debug(u"Loss or incomplete, does not act as a ramp-up.") + return result def no_traffic_loss_occurred(self): """Fail if loss occurred in traffic run. @@ -520,23 +1026,25 @@ class TrafficGenerator(AbstractMeasurer): :raises Exception: If loss occured. """ if self._loss is None: - raise RuntimeError('The traffic generation has not been issued') - if self._loss != '0': - raise RuntimeError('Traffic loss occurred: {0}'.format(self._loss)) + raise RuntimeError(u"The traffic generation has not been issued") + if self._loss != u"0": + raise RuntimeError(f"Traffic loss occurred: {self._loss}") def fail_if_no_traffic_forwarded(self): """Fail if no traffic forwarded. + TODO: Check number of passed transactions instead. + :returns: nothing :raises Exception: If no traffic forwarded. """ if self._received is None: - raise RuntimeError('The traffic generation has not been issued') - if self._received == '0': - raise RuntimeError('No traffic forwarded') + raise RuntimeError(u"The traffic generation has not been issued") + if self._received == 0: + raise RuntimeError(u"No traffic forwarded") - def partial_traffic_loss_accepted(self, loss_acceptance, - loss_acceptance_type): + def partial_traffic_loss_accepted( + self, loss_acceptance, loss_acceptance_type): """Fail if loss is higher then accepted in traffic run. :param loss_acceptance: Permitted drop ratio or frames count. @@ -547,82 +1055,403 @@ class TrafficGenerator(AbstractMeasurer): :raises Exception: If loss is above acceptance criteria. """ if self._loss is None: - raise Exception('The traffic generation has not been issued') + raise Exception(u"The traffic generation has not been issued") - if loss_acceptance_type == 'percentage': + if loss_acceptance_type == u"percentage": loss = (float(self._loss) / float(self._sent)) * 100 - elif loss_acceptance_type == 'frames': + elif loss_acceptance_type == u"frames": loss = float(self._loss) else: - raise Exception('Loss acceptance type not supported') + raise Exception(u"Loss acceptance type not supported") if loss > float(loss_acceptance): - raise Exception("Traffic loss {} above loss acceptance: {}".format( - loss, loss_acceptance)) + raise Exception( + f"Traffic loss {loss} above loss acceptance: {loss_acceptance}" + ) - def set_rate_provider_defaults(self, frame_size, traffic_type, - warmup_time=0.0): - """Store values accessed by measure(). + def _parse_traffic_results(self, stdout): + """Parse stdout of scripts into fields of self. - :param frame_size: Frame size identifier or value [B]. - :param traffic_type: Module name as a traffic type identifier. - See resources/traffic_profiles/trex for implemented modules. - :param warmup_time: Traffic duration before measurement starts [s]. - :type frame_size: str or int - :type traffic_type: str - :type warmup_time: float + Block of code to reuse, by sync start, or stop after async. + + :param stdout: Text containing the standard output. + :type stdout: str """ - self.frame_size = frame_size - self.traffic_type = str(traffic_type) - self.warmup_time = float(warmup_time) + subtype = check_subtype(self._node) + if subtype == NodeSubTypeTG.TREX: + # Last line from console output + line = stdout.splitlines()[-1] + results = line.split(u";") + if results[-1] in (u" ", u""): + results.pop(-1) + self._result = dict() + for result in results: + key, value = result.split(u"=", maxsplit=1) + self._result[key.strip()] = value + logger.info(f"TrafficGen results:\n{self._result}") + self._received = int(self._result.get(u"total_received"), 0) + self._sent = int(self._result.get(u"total_sent", 0)) + self._loss = int(self._result.get(u"frame_loss", 0)) + self._approximated_duration = \ + self._result.get(u"approximated_duration", 0.0) + if u"manual" not in str(self._approximated_duration): + self._approximated_duration = float(self._approximated_duration) + self._latency = list() + self._latency.append(self._result.get(u"latency_stream_0(usec)")) + self._latency.append(self._result.get(u"latency_stream_1(usec)")) + if self._mode == TrexMode.ASTF: + self._l7_data = dict() + self._l7_data[u"client"] = dict() + self._l7_data[u"client"][u"sent"] = \ + int(self._result.get(u"client_sent", 0)) + self._l7_data[u"client"][u"received"] = \ + int(self._result.get(u"client_received", 0)) + self._l7_data[u"client"][u"active_flows"] = \ + int(self._result.get(u"client_active_flows", 0)) + self._l7_data[u"client"][u"established_flows"] = \ + int(self._result.get(u"client_established_flows", 0)) + self._l7_data[u"client"][u"traffic_duration"] = \ + float(self._result.get(u"client_traffic_duration", 0.0)) + self._l7_data[u"client"][u"err_rx_throttled"] = \ + int(self._result.get(u"client_err_rx_throttled", 0)) + self._l7_data[u"client"][u"err_c_nf_throttled"] = \ + int(self._result.get(u"client_err_nf_throttled", 0)) + self._l7_data[u"client"][u"err_flow_overflow"] = \ + int(self._result.get(u"client_err_flow_overflow", 0)) + self._l7_data[u"server"] = dict() + self._l7_data[u"server"][u"active_flows"] = \ + int(self._result.get(u"server_active_flows", 0)) + self._l7_data[u"server"][u"established_flows"] = \ + int(self._result.get(u"server_established_flows", 0)) + self._l7_data[u"server"][u"traffic_duration"] = \ + float(self._result.get(u"server_traffic_duration", 0.0)) + self._l7_data[u"server"][u"err_rx_throttled"] = \ + int(self._result.get(u"client_err_rx_throttled", 0)) + if u"udp" in self.traffic_profile: + self._l7_data[u"client"][u"udp"] = dict() + self._l7_data[u"client"][u"udp"][u"connects"] = \ + int(self._result.get(u"client_udp_connects", 0)) + self._l7_data[u"client"][u"udp"][u"closed_flows"] = \ + int(self._result.get(u"client_udp_closed", 0)) + self._l7_data[u"client"][u"udp"][u"tx_bytes"] = \ + int(self._result.get(u"client_udp_tx_bytes", 0)) + self._l7_data[u"client"][u"udp"][u"rx_bytes"] = \ + int(self._result.get(u"client_udp_rx_bytes", 0)) + self._l7_data[u"client"][u"udp"][u"tx_packets"] = \ + int(self._result.get(u"client_udp_tx_packets", 0)) + self._l7_data[u"client"][u"udp"][u"rx_packets"] = \ + int(self._result.get(u"client_udp_rx_packets", 0)) + self._l7_data[u"client"][u"udp"][u"keep_drops"] = \ + int(self._result.get(u"client_udp_keep_drops", 0)) + self._l7_data[u"client"][u"udp"][u"err_cwf"] = \ + int(self._result.get(u"client_err_cwf", 0)) + self._l7_data[u"server"][u"udp"] = dict() + self._l7_data[u"server"][u"udp"][u"accepted_flows"] = \ + int(self._result.get(u"server_udp_accepts", 0)) + self._l7_data[u"server"][u"udp"][u"closed_flows"] = \ + int(self._result.get(u"server_udp_closed", 0)) + self._l7_data[u"server"][u"udp"][u"tx_bytes"] = \ + int(self._result.get(u"server_udp_tx_bytes", 0)) + self._l7_data[u"server"][u"udp"][u"rx_bytes"] = \ + int(self._result.get(u"server_udp_rx_bytes", 0)) + self._l7_data[u"server"][u"udp"][u"tx_packets"] = \ + int(self._result.get(u"server_udp_tx_packets", 0)) + self._l7_data[u"server"][u"udp"][u"rx_packets"] = \ + int(self._result.get(u"server_udp_rx_packets", 0)) + elif u"tcp" in self.traffic_profile: + self._l7_data[u"client"][u"tcp"] = dict() + self._l7_data[u"client"][u"tcp"][u"initiated_flows"] = \ + int(self._result.get(u"client_tcp_connect_inits", 0)) + self._l7_data[u"client"][u"tcp"][u"connects"] = \ + int(self._result.get(u"client_tcp_connects", 0)) + self._l7_data[u"client"][u"tcp"][u"closed_flows"] = \ + int(self._result.get(u"client_tcp_closed", 0)) + self._l7_data[u"client"][u"tcp"][u"connattempt"] = \ + int(self._result.get(u"client_tcp_connattempt", 0)) + self._l7_data[u"client"][u"tcp"][u"tx_bytes"] = \ + int(self._result.get(u"client_tcp_tx_bytes", 0)) + self._l7_data[u"client"][u"tcp"][u"rx_bytes"] = \ + int(self._result.get(u"client_tcp_rx_bytes", 0)) + self._l7_data[u"server"][u"tcp"] = dict() + self._l7_data[u"server"][u"tcp"][u"accepted_flows"] = \ + int(self._result.get(u"server_tcp_accepts", 0)) + self._l7_data[u"server"][u"tcp"][u"connects"] = \ + int(self._result.get(u"server_tcp_connects", 0)) + self._l7_data[u"server"][u"tcp"][u"closed_flows"] = \ + int(self._result.get(u"server_tcp_closed", 0)) + self._l7_data[u"server"][u"tcp"][u"tx_bytes"] = \ + int(self._result.get(u"server_tcp_tx_bytes", 0)) + self._l7_data[u"server"][u"tcp"][u"rx_bytes"] = \ + int(self._result.get(u"server_tcp_rx_bytes", 0)) + + def _get_measurement_result(self): + """Return the result of last measurement as ReceiveRateMeasurement. + + Separate function, as measurements can end either by time + or by explicit call, this is the common block at the end. + + The target_tr field of ReceiveRateMeasurement is in + transactions per second. Transmit count and loss count units + depend on the transaction type. Usually they are in transactions + per second, or aggregate packets per second. + + TODO: Fail on running or already reported measurement. + + :returns: Structure containing the result of the measurement. + :rtype: ReceiveRateMeasurement + """ + try: + # Client duration seems to include a setup period + # where TRex does not send any packets yet. + # Server duration does not include it. + server_data = self._l7_data[u"server"] + approximated_duration = float(server_data[u"traffic_duration"]) + except (KeyError, AttributeError, ValueError, TypeError): + approximated_duration = None + try: + if not approximated_duration: + approximated_duration = float(self._approximated_duration) + except ValueError: # "manual" + approximated_duration = None + if not approximated_duration: + if self._duration and self._duration > 0: + # Known recomputed or target duration. + approximated_duration = self._duration + else: + # It was an explicit stop. + if not self._stop_time: + raise RuntimeError(u"Unable to determine duration.") + approximated_duration = self._stop_time - self._start_time + target_duration = self._target_duration + if not target_duration: + target_duration = approximated_duration + transmit_rate = self._rate + if self.transaction_type == u"packet": + partial_attempt_count = self._sent + expected_attempt_count = self._sent + fail_count = self._loss + elif self.transaction_type == u"udp_cps": + if not self.transaction_scale: + raise RuntimeError(u"Add support for no-limit udp_cps.") + partial_attempt_count = self._l7_data[u"client"][u"sent"] + # We do not care whether TG is slow, it should have attempted all. + expected_attempt_count = self.transaction_scale + pass_count = self._l7_data[u"client"][u"received"] + fail_count = expected_attempt_count - pass_count + elif self.transaction_type == u"tcp_cps": + if not self.transaction_scale: + raise RuntimeError(u"Add support for no-limit tcp_cps.") + ctca = self._l7_data[u"client"][u"tcp"][u"connattempt"] + partial_attempt_count = ctca + # We do not care whether TG is slow, it should have attempted all. + expected_attempt_count = self.transaction_scale + # From TCP point of view, server/connects counts full connections, + # but we are testing NAT session so client/connects counts that + # (half connections from TCP point of view). + pass_count = self._l7_data[u"client"][u"tcp"][u"connects"] + fail_count = expected_attempt_count - pass_count + elif self.transaction_type == u"udp_pps": + if not self.transaction_scale: + raise RuntimeError(u"Add support for no-limit udp_pps.") + partial_attempt_count = self._sent + expected_attempt_count = self.transaction_scale * self.ppta + fail_count = self._loss + (expected_attempt_count - self._sent) + elif self.transaction_type == u"tcp_pps": + if not self.transaction_scale: + raise RuntimeError(u"Add support for no-limit tcp_pps.") + partial_attempt_count = self._sent + expected_attempt_count = self.transaction_scale * self.ppta + # One loss-like scenario happens when TRex receives all packets + # on L2 level, but is not fast enough to process them all + # at L7 level, which leads to retransmissions. + # Those manifest as opackets larger than expected. + # A simple workaround is to add absolute difference. + # Probability of retransmissions exactly cancelling + # packets unsent due to duration stretching is quite low. + fail_count = self._loss + abs(expected_attempt_count - self._sent) + else: + raise RuntimeError(f"Unknown parsing {self.transaction_type!r}") + if fail_count < 0 and not self.negative_loss: + fail_count = 0 + measurement = ReceiveRateMeasurement( + duration=target_duration, + target_tr=transmit_rate, + transmit_count=expected_attempt_count, + loss_count=fail_count, + approximated_duration=approximated_duration, + partial_transmit_count=partial_attempt_count, + ) + measurement.latency = self.get_latency_int() + return measurement def measure(self, duration, transmit_rate): - """Run bi-directional measurement, parse and return results. + """Run trial measurement, parse and return results. + + The input rate is for transactions. Stateles bidirectional traffic + is understood as sequence of (asynchronous) transactions, + two packets each. + + The result units depend on test type, generally + the count either transactions or packets (aggregated over directions). + + Optionally, this method sleeps if measurement finished before + the time specified as duration. :param duration: Trial duration [s]. - :param transmit_rate: Target bidirectional transmit rate [pps]. + :param transmit_rate: Target rate in transactions per second. :type duration: float :type transmit_rate: float :returns: Structure containing the result of the measurement. :rtype: ReceiveRateMeasurement - :raises RuntimeError: If TG is not set, or if node is not TG, + :raises RuntimeError: If TG is not set or if node is not TG or if subtype is not specified. :raises NotImplementedError: If TG is not supported. """ duration = float(duration) - transmit_rate = float(transmit_rate) - # Trex needs target Tr per stream, but reports aggregate Tx and Dx. - unit_rate = str(transmit_rate / 2.0) + "pps" - self.send_traffic_on_tg( - duration, unit_rate, self.frame_size, self.traffic_type, - self.warmup_time, latency=True) - transmit_count = int(self.get_sent()) - loss_count = int(self.get_loss()) - measurement = ReceiveRateMeasurement( - duration, transmit_rate, transmit_count, loss_count) - measurement.latency = self.get_latency_int() - return measurement + time_start = time.monotonic() + time_stop = time_start + duration + if self.resetter: + self.resetter() + result = self._send_traffic_on_tg_with_ramp_up( + duration=duration, + rate=transmit_rate, + async_call=False, + ) + logger.debug(f"trial measurement result: {result!r}") + # In PLRsearch, computation needs the specified time to complete. + if self.sleep_till_duration: + sleeptime = time_stop - time.monotonic() + if sleeptime > 0.0: + # TODO: Sometimes we have time to do additional trials here, + # adapt PLRsearch to accept all the results. + time.sleep(sleeptime) + return result + def set_rate_provider_defaults( + self, + frame_size, + traffic_profile, + ppta=1, + resetter=None, + traffic_directions=2, + transaction_duration=0.0, + transaction_scale=0, + transaction_type=u"packet", + duration_limit=0.0, + negative_loss=True, + sleep_till_duration=False, + use_latency=False, + ramp_up_rate=None, + ramp_up_duration=None, + state_timeout=300.0, + ): + """Store values accessed by measure(). -class OptimizedSearch(object): - """Class to be imported as Robot Library, containing a single keyword.""" + :param frame_size: Frame size identifier or value [B]. + :param traffic_profile: Module name as a traffic profile identifier. + See GPL/traffic_profiles/trex for implemented modules. + :param ppta: Packets per transaction, aggregated over directions. + Needed for udp_pps which does not have a good transaction counter, + so we need to compute expected number of packets. + Default: 1. + :param resetter: Callable to reset DUT state for repeated trials. + :param traffic_directions: Traffic from packet counting point of view + is bi- (2) or uni- (1) directional. + Default: 2 + :param transaction_duration: Total expected time to close transaction. + :param transaction_scale: Number of transactions to perform. + 0 (default) means unlimited. + :param transaction_type: An identifier specifying which counters + and formulas to use when computing attempted and failed + transactions. Default: "packet". + TODO: Does this also specify parsing for the measured duration? + :param duration_limit: Zero or maximum limit for computed (or given) + duration. + :param negative_loss: If false, negative loss is reported as zero loss. + :param sleep_till_duration: If true and measurement returned faster, + sleep until it matches duration. Needed for PLRsearch. + :param use_latency: Whether to measure latency during the trial. + Default: False. + :param ramp_up_rate: Rate to use in ramp-up trials [pps]. + :param ramp_up_duration: Duration of ramp-up trials [s]. + :param state_timeout: Time of life of DUT state [s]. + :type frame_size: str or int + :type traffic_profile: str + :type ppta: int + :type resetter: Optional[Callable[[], None]] + :type traffic_directions: int + :type transaction_duration: float + :type transaction_scale: int + :type transaction_type: str + :type duration_limit: float + :type negative_loss: bool + :type sleep_till_duration: bool + :type use_latency: bool + :type ramp_up_rate: float + :type ramp_up_duration: float + :type state_timeout: float + """ + self.frame_size = frame_size + self.traffic_profile = str(traffic_profile) + self.resetter = resetter + self.ppta = ppta + self.traffic_directions = int(traffic_directions) + self.transaction_duration = float(transaction_duration) + self.transaction_scale = int(transaction_scale) + self.transaction_type = str(transaction_type) + self.duration_limit = float(duration_limit) + self.negative_loss = bool(negative_loss) + self.sleep_till_duration = bool(sleep_till_duration) + self.use_latency = bool(use_latency) + self.ramp_up_rate = float(ramp_up_rate) + self.ramp_up_duration = float(ramp_up_duration) + self.state_timeout = float(state_timeout) + + +class OptimizedSearch: + """Class to be imported as Robot Library, containing search keywords. + + Aside of setting up measurer and forwarding arguments, + the main business is to translate min/max rate from unidir to aggregate. + """ @staticmethod def perform_optimized_ndrpdr_search( - frame_size, traffic_type, minimum_transmit_rate, - maximum_transmit_rate, packet_loss_ratio=0.005, - final_relative_width=0.005, final_trial_duration=30.0, - initial_trial_duration=1.0, number_of_intermediate_phases=2, - timeout=600.0): + frame_size, + traffic_profile, + minimum_transmit_rate, + maximum_transmit_rate, + packet_loss_ratio=0.005, + final_relative_width=0.005, + final_trial_duration=30.0, + initial_trial_duration=1.0, + number_of_intermediate_phases=2, + timeout=720.0, + doublings=1, + ppta=1, + resetter=None, + traffic_directions=2, + transaction_duration=0.0, + transaction_scale=0, + transaction_type=u"packet", + use_latency=False, + ramp_up_rate=None, + ramp_up_duration=None, + state_timeout=300.0, + ): """Setup initialized TG, perform optimized search, return intervals. + If transaction_scale is nonzero, all non-init trial durations + are set to 2.0 (as they do not affect the real trial duration) + and zero intermediate phases are used. + The initial phase still uses 1.0 seconds, to force remeasurement. + That makes initial phase act as a warmup. + :param frame_size: Frame size identifier or value [B]. - :param traffic_type: Module name as a traffic type identifier. - See resources/traffic_profiles/trex for implemented modules. - :param minimum_transmit_rate: Minimal bidirectional - target transmit rate [pps]. - :param maximum_transmit_rate: Maximal bidirectional - target transmit rate [pps]. + :param traffic_profile: Module name as a traffic profile identifier. + See GPL/traffic_profiles/trex for implemented modules. + :param minimum_transmit_rate: Minimal load in transactions per second. + :param maximum_transmit_rate: Maximal load in transactions per second. :param packet_loss_ratio: Fraction of packets lost, for PDR [1]. :param final_relative_width: Final lower bound transmit rate cannot be more distant that this multiple of upper bound [1]. @@ -633,8 +1462,29 @@ class OptimizedSearch(object): to perform before the final phase [1]. :param timeout: The search will fail itself when not finished before this overall time [s]. + :param doublings: How many doublings to do in external search step. + Default 1 is suitable for fairly stable tests, + less stable tests might get better overal duration with 2 or more. + :param ppta: Packets per transaction, aggregated over directions. + Needed for udp_pps which does not have a good transaction counter, + so we need to compute expected number of packets. + Default: 1. + :param resetter: Callable to reset DUT state for repeated trials. + :param traffic_directions: Traffic is bi- (2) or uni- (1) directional. + Default: 2 + :param transaction_duration: Total expected time to close transaction. + :param transaction_scale: Number of transactions to perform. + 0 (default) means unlimited. + :param transaction_type: An identifier specifying which counters + and formulas to use when computing attempted and failed + transactions. Default: "packet". + :param use_latency: Whether to measure latency during the trial. + Default: False. + :param ramp_up_rate: Rate to use in ramp-up trials [pps]. + :param ramp_up_duration: Duration of ramp-up trials [s]. + :param state_timeout: Time of life of DUT state [s]. :type frame_size: str or int - :type traffic_type: str + :type traffic_profile: str :type minimum_transmit_rate: float :type maximum_transmit_rate: float :type packet_loss_ratio: float @@ -643,6 +1493,17 @@ class OptimizedSearch(object): :type initial_trial_duration: float :type number_of_intermediate_phases: int :type timeout: float + :type doublings: int + :type ppta: int + :type resetter: Optional[Callable[[], None]] + :type traffic_directions: int + :type transaction_duration: float + :type transaction_scale: int + :type transaction_type: str + :type use_latency: bool + :type ramp_up_rate: float + :type ramp_up_duration: float + :type state_timeout: float :returns: Structure containing narrowed down NDR and PDR intervals and their measurements. :rtype: NdrPdrResult @@ -651,13 +1512,164 @@ class OptimizedSearch(object): # we need instance of TrafficGenerator instantiated by Robot Framework # to be able to use trex_stl-*() tg_instance = BuiltIn().get_library_instance( - 'resources.libraries.python.TrafficGenerator') - tg_instance.set_rate_provider_defaults(frame_size, traffic_type) + u"resources.libraries.python.TrafficGenerator" + ) + # Overrides for fixed transaction amount. + # TODO: Move to robot code? We have two call sites, so this saves space, + # even though this is surprising for log readers. + if transaction_scale: + initial_trial_duration = 1.0 + final_trial_duration = 2.0 + number_of_intermediate_phases = 0 + timeout += transaction_scale * 3e-4 + tg_instance.set_rate_provider_defaults( + frame_size=frame_size, + traffic_profile=traffic_profile, + sleep_till_duration=False, + ppta=ppta, + resetter=resetter, + traffic_directions=traffic_directions, + transaction_duration=transaction_duration, + transaction_scale=transaction_scale, + transaction_type=transaction_type, + use_latency=use_latency, + ramp_up_rate=ramp_up_rate, + ramp_up_duration=ramp_up_duration, + state_timeout=state_timeout, + ) algorithm = MultipleLossRatioSearch( - measurer=tg_instance, final_trial_duration=final_trial_duration, + measurer=tg_instance, + final_trial_duration=final_trial_duration, final_relative_width=final_relative_width, number_of_intermediate_phases=number_of_intermediate_phases, - initial_trial_duration=initial_trial_duration, timeout=timeout) + initial_trial_duration=initial_trial_duration, + timeout=timeout, + doublings=doublings, + ) result = algorithm.narrow_down_ndr_and_pdr( - minimum_transmit_rate, maximum_transmit_rate, packet_loss_ratio) + min_rate=minimum_transmit_rate, + max_rate=maximum_transmit_rate, + packet_loss_ratio=packet_loss_ratio, + ) + return result + + @staticmethod + def perform_soak_search( + frame_size, + traffic_profile, + minimum_transmit_rate, + maximum_transmit_rate, + plr_target=1e-7, + tdpt=0.1, + initial_count=50, + timeout=7200.0, + ppta=1, + resetter=None, + trace_enabled=False, + traffic_directions=2, + transaction_duration=0.0, + transaction_scale=0, + transaction_type=u"packet", + use_latency=False, + ramp_up_rate=None, + ramp_up_duration=None, + state_timeout=300.0, + ): + """Setup initialized TG, perform soak search, return avg and stdev. + + :param frame_size: Frame size identifier or value [B]. + :param traffic_profile: Module name as a traffic profile identifier. + See GPL/traffic_profiles/trex for implemented modules. + :param minimum_transmit_rate: Minimal load in transactions per second. + :param maximum_transmit_rate: Maximal load in transactions per second. + :param plr_target: Fraction of packets lost to achieve [1]. + :param tdpt: Trial duration per trial. + The algorithm linearly increases trial duration with trial number, + this is the increment between succesive trials, in seconds. + :param initial_count: Offset to apply before the first trial. + For example initial_count=50 makes first trial to be 51*tdpt long. + This is needed because initial "search" phase of integrator + takes significant time even without any trial results. + :param timeout: The search will stop after this overall time [s]. + :param ppta: Packets per transaction, aggregated over directions. + Needed for udp_pps which does not have a good transaction counter, + so we need to compute expected number of packets. + Default: 1. + :param resetter: Callable to reset DUT state for repeated trials. + :param trace_enabled: True if trace enabled else False. + This is very verbose tracing on numeric computations, + do not use in production. + Default: False + :param traffic_directions: Traffic is bi- (2) or uni- (1) directional. + Default: 2 + :param transaction_duration: Total expected time to close transaction. + :param transaction_scale: Number of transactions to perform. + 0 (default) means unlimited. + :param transaction_type: An identifier specifying which counters + and formulas to use when computing attempted and failed + transactions. Default: "packet". + :param use_latency: Whether to measure latency during the trial. + Default: False. + :param ramp_up_rate: Rate to use in ramp-up trials [pps]. + :param ramp_up_duration: Duration of ramp-up trials [s]. + :param state_timeout: Time of life of DUT state [s]. + :type frame_size: str or int + :type traffic_profile: str + :type minimum_transmit_rate: float + :type maximum_transmit_rate: float + :type plr_target: float + :type initial_count: int + :type timeout: float + :type ppta: int + :type resetter: Optional[Callable[[], None]] + :type trace_enabled: bool + :type traffic_directions: int + :type transaction_duration: float + :type transaction_scale: int + :type transaction_type: str + :type use_latency: bool + :type ramp_up_rate: float + :type ramp_up_duration: float + :type state_timeout: float + :returns: Average and stdev of estimated aggregate rate giving PLR. + :rtype: 2-tuple of float + """ + tg_instance = BuiltIn().get_library_instance( + u"resources.libraries.python.TrafficGenerator" + ) + # Overrides for fixed transaction amount. + # TODO: Move to robot code? We have a single call site + # but MLRsearch has two and we want the two to be used similarly. + if transaction_scale: + # TODO: What is a good value for max scale? + # TODO: Scale the timeout with transaction scale. + timeout = 7200.0 + tg_instance.set_rate_provider_defaults( + frame_size=frame_size, + traffic_profile=traffic_profile, + negative_loss=False, + sleep_till_duration=True, + ppta=ppta, + resetter=resetter, + traffic_directions=traffic_directions, + transaction_duration=transaction_duration, + transaction_scale=transaction_scale, + transaction_type=transaction_type, + use_latency=use_latency, + ramp_up_rate=ramp_up_rate, + ramp_up_duration=ramp_up_duration, + state_timeout=state_timeout, + ) + algorithm = PLRsearch( + measurer=tg_instance, + trial_duration_per_trial=tdpt, + packet_loss_ratio_target=plr_target, + trial_number_offset=initial_count, + timeout=timeout, + trace_enabled=trace_enabled, + ) + result = algorithm.search( + min_rate=minimum_transmit_rate, + max_rate=maximum_transmit_rate, + ) return result