NAT44ed: Add pure throuhput bidir tests
[csit.git] / resources / libraries / python / TrafficGenerator.py
index fb9ae4d..30be3b9 100644 (file)
@@ -1,4 +1,4 @@
-# Copyright (c) 2016 Cisco and/or its affiliates.
+# Copyright (c) 2021 Cisco and/or its affiliates.
 # Licensed under the Apache License, Version 2.0 (the "License");
 # you may not use this file except in compliance with the License.
 # You may obtain a copy of the License at:
 
 """Performance testing traffic generator library."""
 
+import time
+
 from robot.api import logger
+from robot.libraries.BuiltIn import BuiltIn
+
+from .Constants import Constants
+from .CpuUtils import CpuUtils
+from .DropRateSearch import DropRateSearch
+from .MLRsearch.AbstractMeasurer import AbstractMeasurer
+from .MLRsearch.MultipleLossRatioSearch import MultipleLossRatioSearch
+from .MLRsearch.ReceiveRateMeasurement import ReceiveRateMeasurement
+from .PLRsearch.PLRsearch import PLRsearch
+from .OptionString import OptionString
+from .ssh import exec_cmd_no_error, exec_cmd
+from .topology import NodeType
+from .topology import NodeSubTypeTG
+from .topology import Topology
+
+__all__ = [u"TGDropRateSearchImpl", u"TrafficGenerator", u"OptimizedSearch"]
+
+
+def check_subtype(node):
+    """Return supported subtype of given node, or raise an exception.
+
+    Currently only one subtype is supported,
+    but we want our code to be ready for other ones.
 
-from resources.libraries.python.ssh import SSH
-from resources.libraries.python.topology import NodeType
-from resources.libraries.python.topology import NodeSubTypeTG
-from resources.libraries.python.topology import Topology
+    :param node: Topology node to check. Can be None.
+    :type node: dict or NoneType
+    :returns: Subtype detected.
+    :rtype: NodeSubTypeTG
+    :raises RuntimeError: If node is not supported, message explains how.
+    """
+    if node.get(u"type") is None:
+        msg = u"Node type is not defined"
+    elif node[u"type"] != NodeType.TG:
+        msg = f"Node type is {node[u'type']!r}, not a TG"
+    elif node.get(u"subtype") is None:
+        msg = u"TG subtype is not defined"
+    elif node[u"subtype"] != NodeSubTypeTG.TREX:
+        msg = f"TG subtype {node[u'subtype']!r} is not supported"
+    else:
+        return NodeSubTypeTG.TREX
+    raise RuntimeError(msg)
 
-__all__ = ['TrafficGenerator']
 
-class TrafficGenerator(object):
-    """Traffic Generator"""
+class TGDropRateSearchImpl(DropRateSearch):
+    """Drop Rate Search implementation."""
 
-    ROBOT_LIBRARY_SCOPE = 'TEST SUITE'
+    # def __init__(self):
+    #     super(TGDropRateSearchImpl, self).__init__()
+
+    def measure_loss(
+            self, rate, frame_size, loss_acceptance, loss_acceptance_type,
+            traffic_profile):
+        """Runs the traffic and evaluate the measured results.
+
+        :param rate: Offered traffic load.
+        :param frame_size: Size of frame.
+        :param loss_acceptance: Permitted drop ratio or frames count.
+        :param loss_acceptance_type: Type of permitted loss.
+        :param traffic_profile: Module name as a traffic profile identifier.
+            See GPL/traffic_profiles/trex for implemented modules.
+        :type rate: float
+        :type frame_size: str
+        :type loss_acceptance: float
+        :type loss_acceptance_type: LossAcceptanceType
+        :type traffic_profile: str
+        :returns: Drop threshold exceeded? (True/False)
+        :rtype: bool
+        :raises NotImplementedError: If TG is not supported.
+        :raises RuntimeError: If TG is not specified.
+        """
+        # we need instance of TrafficGenerator instantiated by Robot Framework
+        # to be able to use trex_stl-*()
+        tg_instance = BuiltIn().get_library_instance(
+            u"resources.libraries.python.TrafficGenerator"
+        )
+        subtype = check_subtype(tg_instance.node)
+        if subtype == NodeSubTypeTG.TREX:
+            unit_rate = str(rate) + self.get_rate_type_str()
+            tg_instance.trex_stl_start_remote_exec(
+                self.get_duration(), unit_rate, frame_size, traffic_profile
+            )
+            loss = tg_instance.get_loss()
+            sent = tg_instance.get_sent()
+            if self.loss_acceptance_type_is_percentage():
+                loss = (float(loss) / float(sent)) * 100
+            logger.trace(
+                f"comparing: {loss} < {loss_acceptance} {loss_acceptance_type}"
+            )
+            return float(loss) <= float(loss_acceptance)
+        return False
+
+    def get_latency(self):
+        """Returns min/avg/max latency.
+
+        :returns: Latency stats.
+        :rtype: list
+        """
+        tg_instance = BuiltIn().get_library_instance(
+            u"resources.libraries.python.TrafficGenerator"
+        )
+        return tg_instance.get_latency_int()
+
+
+class TrexMode:
+    """Defines mode of T-Rex traffic generator."""
+    # Advanced stateful mode
+    ASTF = u"ASTF"
+    # Stateless mode
+    STL = u"STL"
+
+
+# TODO: Pylint says too-many-instance-attributes.
+class TrafficGenerator(AbstractMeasurer):
+    """Traffic Generator."""
+
+    # TODO: Remove "trex" from lines which could work with other TGs.
+
+    # Use one instance of TrafficGenerator for all tests in test suite
+    ROBOT_LIBRARY_SCOPE = u"TEST SUITE"
 
     def __init__(self):
+        # TODO: Separate into few dataclasses/dicts.
+        #       Pylint dislikes large unstructured state, and it is right.
+        self._node = None
+        self._mode = None
+        # TG interface order mapping
+        self._ifaces_reordered = False
+        # Result holding fields, to be removed.
         self._result = None
         self._loss = None
         self._sent = None
+        self._latency = None
         self._received = None
-        #T-REX interface order mapping
-        self._ifaces_reordered = 0
-
-    def initialize_traffic_generator(self, tg_node, tg_if1, tg_if2,
-                                     dut1_node, dut1_if1, dut1_if2,
-                                     dut2_node, dut2_if1, dut2_if2,
-                                     test_type):
-        """TG initialization
-        :param tg_node: Traffic generator node
-        :param tg_if1: TG - name of first interface
-        :param tg_if2: TG - name of second interface
-        :param dut1_node: DUT1 node
-        :param dut1_if1: DUT1 - name of first interface
-        :param dut1_if2: DUT1 - name of second interface
-        :param dut2_node: DUT2 node
-        :param dut2_if1: DUT2 - name of first interface
-        :param dut2_if2: DUT2 - name of second interface
-        :test_type: 'L2' or 'L3' - src/dst MAC address
+        self._approximated_rate = None
+        self._approximated_duration = None
+        self._l7_data = None
+        # Measurement input fields, needed for async stop result.
+        self._start_time = None
+        self._stop_time = None
+        self._rate = None
+        self._target_duration = None
+        self._duration = None
+        # Other input parameters, not knowable from measure() signature.
+        self.frame_size = None
+        self.traffic_profile = None
+        self.traffic_directions = None
+        self.negative_loss = None
+        self.use_latency = None
+        self.ppta = None
+        self.resetter = None
+        self.transaction_scale = None
+        self.transaction_duration = None
+        self.sleep_till_duration = None
+        self.transaction_type = None
+        self.duration_limit = None
+        self.ramp_up_start = None
+        self.ramp_up_stop = None
+        self.ramp_up_rate = None
+        self.ramp_up_duration = None
+        self.state_timeout = None
+        # Transient data needed for async measurements.
+        self._xstats = (None, None)
+        # TODO: Rename "xstats" to something opaque, so T-Rex is not privileged?
+
+    @property
+    def node(self):
+        """Getter.
+
+        :returns: Traffic generator node.
+        :rtype: dict
+        """
+        return self._node
+
+    def get_loss(self):
+        """Return number of lost packets.
+
+        :returns: Number of lost packets.
+        :rtype: str
+        """
+        return self._loss
+
+    def get_sent(self):
+        """Return number of sent packets.
+
+        :returns: Number of sent packets.
+        :rtype: str
+        """
+        return self._sent
+
+    def get_received(self):
+        """Return number of received packets.
+
+        :returns: Number of received packets.
+        :rtype: str
+        """
+        return self._received
+
+    def get_latency_int(self):
+        """Return rounded min/avg/max latency.
+
+        :returns: Latency stats.
+        :rtype: list
+        """
+        return self._latency
+
+    def get_approximated_rate(self):
+        """Return approximated rate computed as ratio of transmitted packets
+        over duration of trial.
+
+        :returns: Approximated rate.
+        :rtype: str
+        """
+        return self._approximated_rate
+
+    def get_l7_data(self):
+        """Return L7 data.
+
+        :returns: Number of received packets.
+        :rtype: dict
+        """
+        return self._l7_data
+
+    def check_mode(self, expected_mode):
+        """Check TG mode.
+
+        :param expected_mode: Expected traffic generator mode.
+        :type expected_mode: object
+        :raises RuntimeError: In case of unexpected TG mode.
+        """
+        if self._mode == expected_mode:
+            return
+        raise RuntimeError(
+            f"{self._node[u'subtype']} not running in {expected_mode} mode!"
+        )
+
+    # TODO: pylint says disable=too-many-locals.
+    def initialize_traffic_generator(
+            self, tg_node, tg_if1, tg_if2, tg_if1_adj_node, tg_if1_adj_if,
+            tg_if2_adj_node, tg_if2_adj_if, osi_layer, tg_if1_dst_mac=None,
+            tg_if2_dst_mac=None):
+        """TG initialization.
+
+        TODO: Document why do we need (and how do we use) _ifaces_reordered.
+
+        :param tg_node: Traffic generator node.
+        :param tg_if1: TG - name of first interface.
+        :param tg_if2: TG - name of second interface.
+        :param tg_if1_adj_node: TG if1 adjecent node.
+        :param tg_if1_adj_if: TG if1 adjecent interface.
+        :param tg_if2_adj_node: TG if2 adjecent node.
+        :param tg_if2_adj_if: TG if2 adjecent interface.
+        :param osi_layer: 'L2', 'L3' or 'L7' - OSI Layer testing type.
+        :param tg_if1_dst_mac: Interface 1 destination MAC address.
+        :param tg_if2_dst_mac: Interface 2 destination MAC address.
         :type tg_node: dict
         :type tg_if1: str
         :type tg_if2: str
-        :type dut1_node: dict
-        :type dut1_if1: str
-        :type dut1_if2: str
-        :type dut2_node: dict
-        :type dut2_if1: str
-        :type dut2_if2: str
-        :type test_type: str
-        :return: nothing
-        """
-
-        trex_path = "/opt/trex-core-1.91"
-
-        topo = Topology()
-
-        if tg_node['type'] != NodeType.TG:
-            raise Exception('Node type is not a TG')
-        if tg_node['subtype'] == NodeSubTypeTG.TREX:
-            ssh = SSH()
-            ssh.connect(tg_node)
-
-            if1_pci = topo.get_interface_pci_addr(tg_node, tg_if1)
-            if2_pci = topo.get_interface_pci_addr(tg_node, tg_if2)
-            if1_mac = topo.get_interface_mac(tg_node, tg_if1)
-            if2_mac = topo.get_interface_mac(tg_node, tg_if2)
-
-            if test_type == 'L2':
-                if1_adj_mac = if2_mac
-                if2_adj_mac = if1_mac
-            elif test_type == 'L3':
-                if1_adj_mac = topo.get_interface_mac(dut1_node, dut1_if1)
-                if2_adj_mac = topo.get_interface_mac(dut2_node, dut2_if2)
+        :type tg_if1_adj_node: dict
+        :type tg_if1_adj_if: str
+        :type tg_if2_adj_node: dict
+        :type tg_if2_adj_if: str
+        :type osi_layer: str
+        :type tg_if1_dst_mac: str
+        :type tg_if2_dst_mac: str
+        :returns: nothing
+        :raises RuntimeError: In case of issue during initialization.
+        """
+        subtype = check_subtype(tg_node)
+        if subtype == NodeSubTypeTG.TREX:
+            self._node = tg_node
+            self._mode = TrexMode.ASTF if osi_layer == u"L7" else TrexMode.STL
+            if1 = dict()
+            if2 = dict()
+            if1[u"pci"] = Topology().get_interface_pci_addr(self._node, tg_if1)
+            if2[u"pci"] = Topology().get_interface_pci_addr(self._node, tg_if2)
+            if1[u"addr"] = Topology().get_interface_mac(self._node, tg_if1)
+            if2[u"addr"] = Topology().get_interface_mac(self._node, tg_if2)
+
+            if osi_layer == u"L2":
+                if1[u"adj_addr"] = if2[u"addr"]
+                if2[u"adj_addr"] = if1[u"addr"]
+            elif osi_layer in (u"L3", u"L7"):
+                if1[u"adj_addr"] = Topology().get_interface_mac(
+                    tg_if1_adj_node, tg_if1_adj_if
+                )
+                if2[u"adj_addr"] = Topology().get_interface_mac(
+                    tg_if2_adj_node, tg_if2_adj_if
+                )
+            else:
+                raise ValueError(u"Unknown OSI layer!")
+
+            # in case of switched environment we can override MAC addresses
+            if tg_if1_dst_mac is not None and tg_if2_dst_mac is not None:
+                if1[u"adj_addr"] = tg_if1_dst_mac
+                if2[u"adj_addr"] = tg_if2_dst_mac
+
+            if min(if1[u"pci"], if2[u"pci"]) != if1[u"pci"]:
+                if1, if2 = if2, if1
+                self._ifaces_reordered = True
+
+            master_thread_id, latency_thread_id, socket, threads = \
+                CpuUtils.get_affinity_trex(
+                    self._node, tg_if1, tg_if2,
+                    tg_dtc=Constants.TREX_CORE_COUNT)
+
+            if osi_layer in (u"L2", u"L3", u"L7"):
+                exec_cmd_no_error(
+                    self._node,
+                    f"sh -c 'cat << EOF > /etc/trex_cfg.yaml\n"
+                    f"- version: 2\n"
+                    f"  c: {len(threads)}\n"
+                    f"  limit_memory: {Constants.TREX_LIMIT_MEMORY}\n"
+                    f"  interfaces: [\"{if1[u'pci']}\",\"{if2[u'pci']}\"]\n"
+                    f"  port_info:\n"
+                    f"      - dest_mac: \'{if1[u'adj_addr']}\'\n"
+                    f"        src_mac: \'{if1[u'addr']}\'\n"
+                    f"      - dest_mac: \'{if2[u'adj_addr']}\'\n"
+                    f"        src_mac: \'{if2[u'addr']}\'\n"
+                    f"  platform :\n"
+                    f"      master_thread_id: {master_thread_id}\n"
+                    f"      latency_thread_id: {latency_thread_id}\n"
+                    f"      dual_if:\n"
+                    f"          - socket: {socket}\n"
+                    f"            threads: {threads}\n"
+                    f"EOF'",
+                    sudo=True, message=u"T-Rex config generation!"
+                )
             else:
-                raise Exception("test_type unknown")
-
-            if min(if1_pci, if2_pci) != if1_pci:
-                if1_mac, if2_mac = if2_mac, if1_mac
-                if1_pci, if2_pci = if2_pci, if1_pci
-                if1_adj_mac, if2_adj_mac = if2_adj_mac, if1_adj_mac
-                self._ifaces_reordered = 1
-
-            if1_mac_hex = "0x"+if1_mac.replace(":", ",0x")
-            if2_mac_hex = "0x"+if2_mac.replace(":", ",0x")
-            if1_adj_mac_hex = "0x"+if1_adj_mac.replace(":", ",0x")
-            if2_adj_mac_hex = "0x"+if2_adj_mac.replace(":", ",0x")
-
-            (ret, stdout, stderr) = ssh.exec_command(
-                "sudo sh -c 'cat << EOF > /etc/trex_cfg.yaml\n"
-                "- port_limit      : 2\n"
-                "  version         : 2\n"
-                "  interfaces      : [\"{}\",\"{}\"]\n"
-                "  port_info       :\n"
-                "          - dest_mac        :   [{}]\n"
-                "            src_mac         :   [{}]\n"
-                "          - dest_mac        :   [{}]\n"
-                "            src_mac         :   [{}]\n"
-                "EOF'"\
-                .format(if1_pci, if2_pci,
-                        if1_adj_mac_hex, if1_mac_hex,
-                        if2_adj_mac_hex, if2_mac_hex))
-            if int(ret) != 0:
-                logger.error("failed to create t-rex config: {}"\
-                .format(stdout + stderr))
-                raise RuntimeError('trex config generation error')
-
-            (ret, stdout, stderr) = ssh.exec_command(
-                "sh -c 'cd {0}/scripts/ && "
-                "sudo ./trex-cfg'"\
-                .format(trex_path))
-            if int(ret) != 0:
-                logger.error('trex-cfg failed: {0}'.format(stdout + stderr))
-                raise RuntimeError('trex-cfg failed')
-
-            (ret, _, _) = ssh.exec_command(
-                "sh -c 'cd {0}/scripts/ && "
-                "sudo nohup ./t-rex-64 -i -c 4 --iom 0 > /dev/null 2>&1 &'"
-                "> /dev/null"\
-                .format(trex_path))
-            if int(ret) != 0:
-                raise RuntimeError('t-rex-64 startup failed')
+                raise ValueError(u"Unknown OSI layer!")
+
+            TrafficGenerator.startup_trex(
+                self._node, osi_layer, subtype=subtype
+            )
+
+    @staticmethod
+    def startup_trex(tg_node, osi_layer, subtype=None):
+        """Startup sequence for the TRex traffic generator.
+
+        :param tg_node: Traffic generator node.
+        :param osi_layer: 'L2', 'L3' or 'L7' - OSI Layer testing type.
+        :param subtype: Traffic generator sub-type.
+        :type tg_node: dict
+        :type osi_layer: str
+        :type subtype: NodeSubTypeTG
+        :raises RuntimeError: If T-Rex startup failed.
+        :raises ValueError: If OSI layer is not supported.
+        """
+        if not subtype:
+            subtype = check_subtype(tg_node)
+        if subtype == NodeSubTypeTG.TREX:
+            for _ in range(0, 3):
+                # Kill TRex only if it is already running.
+                cmd = u"sh -c \"pgrep t-rex && pkill t-rex && sleep 3 || true\""
+                exec_cmd_no_error(
+                    tg_node, cmd, sudo=True, message=u"Kill TRex failed!"
+                )
+
+                # Configure TRex.
+                ports = ''
+                for port in tg_node[u"interfaces"].values():
+                    if u'Mellanox' not in port.get(u'model'):
+                        ports += f" {port.get(u'pci_address')}"
+
+                cmd = f"sh -c \"cd {Constants.TREX_INSTALL_DIR}/scripts/ && " \
+                    f"./dpdk_nic_bind.py -u {ports} || true\""
+                exec_cmd_no_error(
+                    tg_node, cmd, sudo=True,
+                    message=u"Unbind PCI ports from driver failed!"
+                )
+
+                # Start TRex.
+                cd_cmd = f"cd '{Constants.TREX_INSTALL_DIR}/scripts/'"
+                trex_cmd = OptionString([u"nohup", u"./t-rex-64"])
+                trex_cmd.add(u"-i")
+                trex_cmd.add(u"--prefix $(hostname)")
+                trex_cmd.add(u"--hdrh")
+                trex_cmd.add(u"--no-scapy-server")
+                trex_cmd.add_if(u"--astf", osi_layer == u"L7")
+                # OptionString does not create double space if extra is empty.
+                trex_cmd.add(f"{Constants.TREX_EXTRA_CMDLINE}")
+                inner_command = f"{cd_cmd} && {trex_cmd} > /tmp/trex.log 2>&1 &"
+                cmd = f"sh -c \"{inner_command}\" > /dev/null"
+                try:
+                    exec_cmd_no_error(tg_node, cmd, sudo=True)
+                except RuntimeError:
+                    cmd = u"sh -c \"cat /tmp/trex.log\""
+                    exec_cmd_no_error(
+                        tg_node, cmd, sudo=True,
+                        message=u"Get TRex logs failed!"
+                    )
+                    raise RuntimeError(u"Start TRex failed!")
+
+                # Test T-Rex API responsiveness.
+                cmd = f"python3 {Constants.REMOTE_FW_DIR}/GPL/tools/trex/"
+                if osi_layer in (u"L2", u"L3"):
+                    cmd += u"trex_stl_assert.py"
+                elif osi_layer == u"L7":
+                    cmd += u"trex_astf_assert.py"
+                else:
+                    raise ValueError(u"Unknown OSI layer!")
+                try:
+                    exec_cmd_no_error(
+                        tg_node, cmd, sudo=True,
+                        message=u"T-Rex API is not responding!", retries=20
+                    )
+                except RuntimeError:
+                    continue
+                return
+            # After max retries TRex is still not responding to API critical
+            # error occurred.
+            exec_cmd(tg_node, u"cat /tmp/trex.log", sudo=True)
+            raise RuntimeError(u"Start T-Rex failed after multiple retries!")
+
+    @staticmethod
+    def is_trex_running(node):
+        """Check if T-Rex is running using pidof.
+
+        :param node: Traffic generator node.
+        :type node: dict
+        :returns: True if T-Rex is running otherwise False.
+        :rtype: bool
+        """
+        ret, _, _ = exec_cmd(node, u"pgrep t-rex", sudo=True)
+        return bool(int(ret) == 0)
 
     @staticmethod
     def teardown_traffic_generator(node):
-        """TG teardown
-        :param node: Traffic generator node
+        """TG teardown.
+
+        :param node: Traffic generator node.
         :type node: dict
-        :return: nothing
-        """
-
-        if node['type'] != NodeType.TG:
-            raise Exception('Node type is not a TG')
-        if node['subtype'] == NodeSubTypeTG.TREX:
-            ssh = SSH()
-            ssh.connect(node)
-            (ret, stdout, stderr) = ssh.exec_command(
-                "sh -c 'sudo pkill t-rex'")
-            if int(ret) != 0:
-                logger.error('pkill t-rex failed: {0}'.format(stdout + stderr))
-                raise RuntimeError('pkill t-rex failed')
-
-    def send_traffic_on(self, nodes_info, duration, rate,
-                        framesize, traffic_type):
-        """Send traffic from all configured interfaces on TG
-        :param nodes_info: Dictionary containing information on all nodes
-        in topology.
-        :param duration: Duration of test traffic generation in seconds
-        :param rate: Offered load per interface (e.g. 1%, 3gbps, 4mpps, ...)
-        :param framesize: Frame size (L2) in Bytes
-        :param traffic_type: Traffic profile
-        :type nodes_info: dict
-        :type duration: str
-        :type rate: str
-        :type framesize: str
-        :type traffic_type: str
-        :return: TG output
-        :rtype: str
+        :returns: nothing
+        :raises RuntimeError: If node type is not a TG,
+            or if T-Rex teardown fails.
+        """
+        subtype = check_subtype(node)
+        if subtype == NodeSubTypeTG.TREX:
+            exec_cmd_no_error(
+                node,
+                u"sh -c "
+                u"\"if pgrep t-rex; then sudo pkill t-rex && sleep 3; fi\"",
+                sudo=False,
+                message=u"T-Rex kill failed!"
+            )
+
+    def trex_astf_stop_remote_exec(self, node):
+        """Execute T-Rex ASTF script on remote node over ssh to stop running
+        traffic.
+
+        Internal state is updated with measurement results.
+
+        :param node: T-Rex generator node.
+        :type node: dict
+        :raises RuntimeError: If stop traffic script fails.
+        """
+        command_line = OptionString().add(u"python3")
+        dirname = f"{Constants.REMOTE_FW_DIR}/GPL/tools/trex"
+        command_line.add(f"'{dirname}/trex_astf_stop.py'")
+        command_line.change_prefix(u"--")
+        for index, value in enumerate(self._xstats):
+            if value is not None:
+                value = value.replace(u"'", u"\"")
+                command_line.add_equals(f"xstat{index}", f"'{value}'")
+        stdout, _ = exec_cmd_no_error(
+            node, command_line,
+            message=u"T-Rex ASTF runtime error!"
+        )
+        self._parse_traffic_results(stdout)
+
+    def trex_stl_stop_remote_exec(self, node):
+        """Execute T-Rex STL script on remote node over ssh to stop running
+        traffic.
+
+        Internal state is updated with measurement results.
+
+        :param node: T-Rex generator node.
+        :type node: dict
+        :raises RuntimeError: If stop traffic script fails.
         """
+        command_line = OptionString().add(u"python3")
+        dirname = f"{Constants.REMOTE_FW_DIR}/GPL/tools/trex"
+        command_line.add(f"'{dirname}/trex_stl_stop.py'")
+        command_line.change_prefix(u"--")
+        for index, value in enumerate(self._xstats):
+            if value is not None:
+                value = value.replace(u"'", u"\"")
+                command_line.add_equals(f"xstat{index}", f"'{value}'")
+        stdout, _ = exec_cmd_no_error(
+            node, command_line,
+            message=u"T-Rex STL runtime error!"
+        )
+        self._parse_traffic_results(stdout)
+
+    def stop_traffic_on_tg(self):
+        """Stop all traffic on TG.
+
+        :returns: Structure containing the result of the measurement.
+        :rtype: ReceiveRateMeasurement
+        :raises ValueError: If TG traffic profile is not supported.
+        """
+        subtype = check_subtype(self._node)
+        if subtype != NodeSubTypeTG.TREX:
+            raise ValueError(f"Unsupported TG subtype: {subtype!r}")
+        if u"trex-astf" in self.traffic_profile:
+            self.trex_astf_stop_remote_exec(self._node)
+        elif u"trex-stl" in self.traffic_profile:
+            self.trex_stl_stop_remote_exec(self._node)
+        else:
+            raise ValueError(u"Unsupported T-Rex traffic profile!")
+        self._stop_time = time.monotonic()
+
+        return self._get_measurement_result()
+
+    def _compute_duration(self, duration, multiplier):
+        """Compute duration for profile driver.
+
+        The final result is influenced by transaction scale and duration limit.
+        It is assumed a higher level function has already set those to self.
+        The duration argument is the target value from search point of view,
+        before the overrides are applied here.
 
-        node = nodes_info["TG"]
-
-        if node['type'] != NodeType.TG:
-            raise Exception('Node type is not a TG')
-
-        if node['subtype'] is None:
-            raise Exception('TG subtype not defined')
-
-        ssh = SSH()
-        ssh.connect(node)
-
-        if node['subtype'] == NodeSubTypeTG.TREX:
-
-            _p0 = 1
-            _p1 = 2
-
-            if self._ifaces_reordered != 0:
-                _p0, _p1 = _p1, _p0
-
-            if traffic_type in ["3-node-xconnect", "3-node-bridge"]:
-                (ret, stdout, stderr) = ssh.exec_command(
-                    "sh -c '/tmp/openvpp-testing/resources/tools/t-rex/"
-                    "t-rex-stateless.py "
-                    "-d {0} -r {1} -s {2} "
-                    "--p{3}_src_start_ip 10.10.10.1 "
-                    "--p{3}_src_end_ip 10.10.10.254 "
-                    "--p{3}_dst_start_ip 20.20.20.1 "
-                    "--p{4}_src_start_ip 20.20.20.1 "
-                    "--p{4}_src_end_ip 20.20.20.254 "
-                    "--p{4}_dst_start_ip 10.10.10.1'".\
-                    format(duration, rate, framesize, _p0, _p1),\
-                    timeout=int(duration)+60)
-            elif traffic_type in ["3-node-IPv4"]:
-                (ret, stdout, stderr) = ssh.exec_command(
-                    "sh -c '/tmp/openvpp-testing/resources/tools/t-rex/"
-                    "t-rex-stateless.py "
-                    "-d {0} -r {1} -s {2} "
-                    "--p{3}_src_start_ip 10.10.10.2 "
-                    "--p{3}_src_end_ip 10.10.10.254 "
-                    "--p{3}_dst_start_ip 20.20.20.2 "
-                    "--p{4}_src_start_ip 20.20.20.2 "
-                    "--p{4}_src_end_ip 20.20.20.254 "
-                    "--p{4}_dst_start_ip 10.10.10.2'".\
-                    format(duration, rate, framesize, _p0, _p1),\
-                    timeout=int(duration)+60)
+        Minus one (signalling async traffic start) is kept.
+
+        Completeness flag is also included. Duration limited or async trials
+        are not considered complete for ramp-up purposes.
+
+        :param duration: Time expressed in seconds for how long to send traffic.
+        :param multiplier: Traffic rate in transactions per second.
+        :type duration: float
+        :type multiplier: float
+        :returns: New duration and whether it was a complete ramp-up candidate.
+        :rtype: float, bool
+        """
+        if duration < 0.0:
+            # Keep the async -1.
+            return duration, False
+        computed_duration = duration
+        if self.transaction_scale:
+            computed_duration = self.transaction_scale / multiplier
+            # Log the computed duration,
+            # so we can compare with what telemetry suggests
+            # the real duration was.
+            logger.debug(f"Expected duration {computed_duration}")
+            computed_duration += 0.1115
+        if not self.duration_limit:
+            return computed_duration, True
+        limited_duration = min(computed_duration, self.duration_limit)
+        return limited_duration, (limited_duration == computed_duration)
+
+    def trex_astf_start_remote_exec(
+            self, duration, multiplier, async_call=False):
+        """Execute T-Rex ASTF script on remote node over ssh to start running
+        traffic.
+
+        In sync mode, measurement results are stored internally.
+        In async mode, initial data including xstats are stored internally.
+
+        This method contains the logic to compute duration as maximum time
+        if transaction_scale is nonzero.
+        The transaction_scale argument defines (limits) how many transactions
+        will be started in total. As that amount of transaction can take
+        considerable time (sometimes due to explicit delays in the profile),
+        the real time a trial needs to finish is computed here. For now,
+        in that case the duration argument is ignored, assuming it comes
+        from ASTF-unaware search algorithm. The overall time a single
+        transaction needs is given in parameter transaction_duration,
+        it includes both explicit delays and implicit time it takes
+        to transfer data (or whatever the transaction does).
+
+        Currently it is observed TRex does not start the ASTF traffic
+        immediately, an ad-hoc constant is added to the computed duration
+        to compensate for that.
+
+        If transaction_scale is zero, duration is not recomputed.
+        It is assumed the subsequent result parsing gets the real duration
+        if the traffic stops sooner for any reason.
+
+        Currently, it is assumed traffic profile defines a single transaction.
+        To avoid heavy logic here, the input rate is expected to be in
+        transactions per second, as that directly translates to TRex multiplier,
+        (assuming the profile does not override the default cps value of one).
+
+        :param duration: Time expressed in seconds for how long to send traffic.
+        :param multiplier: Traffic rate in transactions per second.
+        :param async_call: If enabled then don't wait for all incoming traffic.
+        :type duration: float
+        :type multiplier: int
+        :type async_call: bool
+        :raises RuntimeError: In case of T-Rex driver issue.
+        """
+        self.check_mode(TrexMode.ASTF)
+        p_0, p_1 = (1, 0) if self._ifaces_reordered else (0, 1)
+        if not isinstance(duration, (float, int)):
+            duration = float(duration)
+
+        # TODO: Refactor the code so duration is computed only once,
+        # and both the initial and the computed durations are logged.
+        computed_duration, _ = self._compute_duration(duration, multiplier)
+
+        command_line = OptionString().add(u"python3")
+        dirname = f"{Constants.REMOTE_FW_DIR}/GPL/tools/trex"
+        command_line.add(f"'{dirname}/trex_astf_profile.py'")
+        command_line.change_prefix(u"--")
+        dirname = f"{Constants.REMOTE_FW_DIR}/GPL/traffic_profiles/trex"
+        command_line.add_with_value(
+            u"profile", f"'{dirname}/{self.traffic_profile}.py'"
+        )
+        command_line.add_with_value(u"duration", f"{computed_duration!r}")
+        command_line.add_with_value(u"frame_size", self.frame_size)
+        command_line.add_with_value(u"multiplier", multiplier)
+        command_line.add_with_value(u"port_0", p_0)
+        command_line.add_with_value(u"port_1", p_1)
+        command_line.add_with_value(
+            u"traffic_directions", self.traffic_directions
+        )
+        command_line.add_if(u"async_start", async_call)
+        command_line.add_if(u"latency", self.use_latency)
+        command_line.add_if(u"force", Constants.TREX_SEND_FORCE)
+
+        self._start_time = time.monotonic()
+        self._rate = multiplier
+        stdout, _ = exec_cmd_no_error(
+            self._node, command_line, timeout=computed_duration + 10.0,
+            message=u"T-Rex ASTF runtime error!"
+        )
+
+        if async_call:
+            # no result
+            self._target_duration = None
+            self._duration = None
+            self._received = None
+            self._sent = None
+            self._loss = None
+            self._latency = None
+            xstats = [None, None]
+            self._l7_data = dict()
+            self._l7_data[u"client"] = dict()
+            self._l7_data[u"client"][u"active_flows"] = None
+            self._l7_data[u"client"][u"established_flows"] = None
+            self._l7_data[u"client"][u"traffic_duration"] = None
+            self._l7_data[u"server"] = dict()
+            self._l7_data[u"server"][u"active_flows"] = None
+            self._l7_data[u"server"][u"established_flows"] = None
+            self._l7_data[u"server"][u"traffic_duration"] = None
+            if u"udp" in self.traffic_profile:
+                self._l7_data[u"client"][u"udp"] = dict()
+                self._l7_data[u"client"][u"udp"][u"connects"] = None
+                self._l7_data[u"client"][u"udp"][u"closed_flows"] = None
+                self._l7_data[u"client"][u"udp"][u"err_cwf"] = None
+                self._l7_data[u"server"][u"udp"] = dict()
+                self._l7_data[u"server"][u"udp"][u"accepted_flows"] = None
+                self._l7_data[u"server"][u"udp"][u"closed_flows"] = None
+            elif u"tcp" in self.traffic_profile:
+                self._l7_data[u"client"][u"tcp"] = dict()
+                self._l7_data[u"client"][u"tcp"][u"initiated_flows"] = None
+                self._l7_data[u"client"][u"tcp"][u"connects"] = None
+                self._l7_data[u"client"][u"tcp"][u"closed_flows"] = None
+                self._l7_data[u"client"][u"tcp"][u"connattempt"] = None
+                self._l7_data[u"server"][u"tcp"] = dict()
+                self._l7_data[u"server"][u"tcp"][u"accepted_flows"] = None
+                self._l7_data[u"server"][u"tcp"][u"connects"] = None
+                self._l7_data[u"server"][u"tcp"][u"closed_flows"] = None
             else:
-                raise NotImplementedError('Unsupported traffic type')
+                logger.warn(u"Unsupported T-Rex ASTF traffic profile!")
+            index = 0
+            for line in stdout.splitlines():
+                if f"Xstats snapshot {index}: " in line:
+                    xstats[index] = line[19:]
+                    index += 1
+                if index == 2:
+                    break
+            self._xstats = tuple(xstats)
+        else:
+            self._target_duration = duration
+            self._duration = computed_duration
+            self._parse_traffic_results(stdout)
+
+    def trex_stl_start_remote_exec(self, duration, rate, async_call=False):
+        """Execute T-Rex STL script on remote node over ssh to start running
+        traffic.
+
+        In sync mode, measurement results are stored internally.
+        In async mode, initial data including xstats are stored internally.
+
+        Mode-unaware code (e.g. in search algorithms) works with transactions.
+        To keep the logic simple, multiplier is set to that value.
+        As bidirectional traffic profiles send packets in both directions,
+        they are treated as transactions with two packets (one per direction).
+
+        :param duration: Time expressed in seconds for how long to send traffic.
+        :param rate: Traffic rate in transactions per second.
+        :param async_call: If enabled then don't wait for all incoming traffic.
+        :type duration: float
+        :type rate: str
+        :type async_call: bool
+        :raises RuntimeError: In case of T-Rex driver issue.
+        """
+        self.check_mode(TrexMode.STL)
+        p_0, p_1 = (1, 0) if self._ifaces_reordered else (0, 1)
+        if not isinstance(duration, (float, int)):
+            duration = float(duration)
+
+        # TODO: Refactor the code so duration is computed only once,
+        # and both the initial and the computed durations are logged.
+        duration, _ = self._compute_duration(duration=duration, multiplier=rate)
 
+        command_line = OptionString().add(u"python3")
+        dirname = f"{Constants.REMOTE_FW_DIR}/GPL/tools/trex"
+        command_line.add(f"'{dirname}/trex_stl_profile.py'")
+        command_line.change_prefix(u"--")
+        dirname = f"{Constants.REMOTE_FW_DIR}/GPL/traffic_profiles/trex"
+        command_line.add_with_value(
+            u"profile", f"'{dirname}/{self.traffic_profile}.py'"
+        )
+        command_line.add_with_value(u"duration", f"{duration!r}")
+        command_line.add_with_value(u"frame_size", self.frame_size)
+        command_line.add_with_value(u"rate", f"{rate!r}")
+        command_line.add_with_value(u"port_0", p_0)
+        command_line.add_with_value(u"port_1", p_1)
+        command_line.add_with_value(
+            u"traffic_directions", self.traffic_directions
+        )
+        command_line.add_if(u"async_start", async_call)
+        command_line.add_if(u"latency", self.use_latency)
+        command_line.add_if(u"force", Constants.TREX_SEND_FORCE)
+
+        # TODO: This is ugly. Handle parsing better.
+        self._start_time = time.monotonic()
+        self._rate = float(rate[:-3]) if u"pps" in rate else float(rate)
+        stdout, _ = exec_cmd_no_error(
+            self._node, command_line, timeout=int(duration) + 60,
+            message=u"T-Rex STL runtime error"
+        )
+
+        if async_call:
+            # no result
+            self._target_duration = None
+            self._duration = None
+            self._received = None
+            self._sent = None
+            self._loss = None
+            self._latency = None
+
+            xstats = [None, None]
+            index = 0
+            for line in stdout.splitlines():
+                if f"Xstats snapshot {index}: " in line:
+                    xstats[index] = line[19:]
+                    index += 1
+                if index == 2:
+                    break
+            self._xstats = tuple(xstats)
         else:
-            raise NotImplementedError("TG subtype not supported")
+            self._target_duration = duration
+            self._duration = duration
+            self._parse_traffic_results(stdout)
+
+    def send_traffic_on_tg(
+            self,
+            duration,
+            rate,
+            frame_size,
+            traffic_profile,
+            async_call=False,
+            ppta=1,
+            traffic_directions=2,
+            transaction_duration=0.0,
+            transaction_scale=0,
+            transaction_type=u"packet",
+            duration_limit=0.0,
+            use_latency=False,
+            ramp_up_rate=None,
+            ramp_up_duration=None,
+            state_timeout=300.0,
+            ramp_up_only=False,
+        ):
+        """Send traffic from all configured interfaces on TG.
+
+        In async mode, xstats is stored internally,
+        to enable getting correct result when stopping the traffic.
+        In both modes, stdout is returned,
+        but _parse_traffic_results only works in sync output.
+
+        Note that traffic generator uses DPDK driver which might
+        reorder port numbers based on wiring and PCI numbering.
+        This method handles that, so argument values are invariant,
+        but you can see swapped valued in debug logs.
+
+        When transaction_scale is specified, the duration value is ignored
+        and the needed time is computed. For cases where this results in
+        to too long measurement (e.g. teardown trial with small rate),
+        duration_limit is applied (of non-zero), so the trial is stopped sooner.
+
+        Bidirectional STL profiles are treated as transactions with two packets.
+
+        The return value is None for async.
+
+        :param duration: Duration of test traffic generation in seconds.
+        :param rate: Traffic rate in transactions per second.
+        :param frame_size: Frame size (L2) in Bytes.
+        :param traffic_profile: Module name as a traffic profile identifier.
+            See GPL/traffic_profiles/trex for implemented modules.
+        :param async_call: Async mode.
+        :param ppta: Packets per transaction, aggregated over directions.
+            Needed for udp_pps which does not have a good transaction counter,
+            so we need to compute expected number of packets.
+            Default: 1.
+        :param traffic_directions: Traffic is bi- (2) or uni- (1) directional.
+            Default: 2
+        :param transaction_duration: Total expected time to close transaction.
+        :param transaction_scale: Number of transactions to perform.
+            0 (default) means unlimited.
+        :param transaction_type: An identifier specifying which counters
+            and formulas to use when computing attempted and failed
+            transactions. Default: "packet".
+        :param duration_limit: Zero or maximum limit for computed (or given)
+            duration.
+        :param use_latency: Whether to measure latency during the trial.
+            Default: False.
+        :param ramp_up_rate: Rate to use in ramp-up trials [pps].
+        :param ramp_up_duration: Duration of ramp-up trials [s].
+        :param state_timeout: Time of life of DUT state [s].
+        :param ramp_up_only: If true, do not perform main trial measurement.
+        :type duration: float
+        :type rate: float
+        :type frame_size: str
+        :type traffic_profile: str
+        :type async_call: bool
+        :type ppta: int
+        :type traffic_directions: int
+        :type transaction_duration: float
+        :type transaction_scale: int
+        :type transaction_type: str
+        :type duration_limit: float
+        :type use_latency: bool
+        :type ramp_up_rate: float
+        :type ramp_up_duration: float
+        :type state_timeout: float
+        :type ramp_up_only: bool
+        :returns: TG results.
+        :rtype: ReceiveRateMeasurement or None
+        :raises ValueError: If TG traffic profile is not supported.
+        """
+        self.set_rate_provider_defaults(
+            frame_size=frame_size,
+            traffic_profile=traffic_profile,
+            ppta=ppta,
+            traffic_directions=traffic_directions,
+            transaction_duration=transaction_duration,
+            transaction_scale=transaction_scale,
+            transaction_type=transaction_type,
+            duration_limit=duration_limit,
+            use_latency=use_latency,
+            ramp_up_rate=ramp_up_rate,
+            ramp_up_duration=ramp_up_duration,
+            state_timeout=state_timeout,
+        )
+        return self._send_traffic_on_tg_with_ramp_up(
+            duration=duration,
+            rate=rate,
+            async_call=async_call,
+            ramp_up_only=ramp_up_only,
+        )
+
+    def _send_traffic_on_tg_internal(
+            self, duration, rate, async_call=False):
+        """Send traffic from all configured interfaces on TG.
+
+        This is an internal function, it assumes set_rate_provider_defaults
+        has been called to remember most values.
+        The reason why need to remember various values is that
+        the traffic can be asynchronous, and parsing needs those values.
+        The reason why this is is a separate function from the one
+        which calls set_rate_provider_defaults is that some search algorithms
+        need to specify their own values, and we do not want the measure call
+        to overwrite them with defaults.
+
+        This function is used both for automated ramp-up trials
+        and for explicitly called trials.
+
+        :param duration: Duration of test traffic generation in seconds.
+        :param rate: Traffic rate in transactions per second.
+        :param async_call: Async mode.
+        :type duration: float
+        :type rate: float
+        :type async_call: bool
+        :returns: TG results.
+        :rtype: ReceiveRateMeasurement or None
+        :raises ValueError: If TG traffic profile is not supported.
+        """
+        subtype = check_subtype(self._node)
+        if subtype == NodeSubTypeTG.TREX:
+            if u"trex-astf" in self.traffic_profile:
+                self.trex_astf_start_remote_exec(
+                    duration, float(rate), async_call
+                )
+            elif u"trex-stl" in self.traffic_profile:
+                unit_rate_str = str(rate) + u"pps"
+                # TODO: Suport transaction_scale et al?
+                self.trex_stl_start_remote_exec(
+                    duration, unit_rate_str, async_call
+                )
+            else:
+                raise ValueError(u"Unsupported T-Rex traffic profile!")
+
+        return None if async_call else self._get_measurement_result()
+
+    def _send_traffic_on_tg_with_ramp_up(
+            self, duration, rate, async_call=False, ramp_up_only=False):
+        """Send traffic from all interfaces on TG, maybe after ramp-up.
+
+        This is an internal function, it assumes set_rate_provider_defaults
+        has been called to remember most values.
+        The reason why need to remember various values is that
+        the traffic can be asynchronous, and parsing needs those values.
+        The reason why this is a separate function from the one
+        which calls set_rate_provider_defaults is that some search algorithms
+        need to specify their own values, and we do not want the measure call
+        to overwrite them with defaults.
+
+        If ramp-up tracking is detected, a computation is performed,
+        and if state timeout is near, trial at ramp-up rate and duration
+        is inserted before the main trial measurement.
+
+        The ramp_up_only parameter forces a ramp-up without immediate
+        trial measurement, which is useful in case self remembers
+        a previous ramp-up trial that belongs to a different test (phase).
 
-        logger.trace(ret)
-        logger.trace(stdout)
-        logger.trace(stderr)
+        Return None if trial is async or ramp-up only.
 
-        for line in stdout.splitlines():
-            pass
+        :param duration: Duration of test traffic generation in seconds.
+        :param rate: Traffic rate in transactions per second.
+        :param async_call: Async mode.
+        :param ramp_up_only: If true, do not perform main trial measurement.
+        :type duration: float
+        :type rate: float
+        :type async_call: bool
+        :type ramp_up_only: bool
+        :returns: TG results.
+        :rtype: ReceiveRateMeasurement or None
+        :raises ValueError: If TG traffic profile is not supported.
+        """
+        complete = False
+        if self.ramp_up_rate:
+            # Figure out whether we need to insert a ramp-up trial.
+            # TODO: Give up on async_call=True?
+            if self.ramp_up_start is None:
+                # We never ramped up yet.
+                ramp_up_needed = True
+            else:
+                # We ramped up before, but maybe it was too long ago.
+                # Adding a constant overhead to be safe.
+                time_now = time.monotonic() + 1.0
+                computed_duration, complete = self._compute_duration(
+                    duration=duration,
+                    multiplier=rate,
+                )
+                # There are two conditions for inserting ramp-up.
+                # If early sessions are expiring already,
+                # or if late sessions are to expire before measurement is over.
+                ramp_up_start_delay = time_now - self.ramp_up_start
+                ramp_up_stop_delay = time_now - self.ramp_up_stop
+                ramp_up_stop_delay += computed_duration
+                bigger_delay = max(ramp_up_start_delay, ramp_up_stop_delay)
+                # Final boolean decision.
+                ramp_up_needed = (bigger_delay >= self.state_timeout)
+            if ramp_up_needed:
+                logger.debug(
+                    u"State may time out during next real trial, "
+                    u"inserting a ramp-up trial."
+                )
+                self.ramp_up_start = time.monotonic()
+                self._send_traffic_on_tg_internal(
+                    duration=self.ramp_up_duration,
+                    rate=self.ramp_up_rate,
+                    async_call=async_call,
+                )
+                self.ramp_up_stop = time.monotonic()
+                logger.debug(u"Ramp-up done.")
+            else:
+                logger.debug(
+                    u"State will probably not time out during next real trial, "
+                    u"no ramp-up trial needed just yet."
+                )
+        if ramp_up_only:
+            return None
+        trial_start = time.monotonic()
+        result = self._send_traffic_on_tg_internal(
+            duration=duration,
+            rate=rate,
+            async_call=async_call,
+        )
+        trial_end = time.monotonic()
+        if self.ramp_up_rate:
+            # Optimization: No loss acts as a good ramp-up, if it was complete.
+            if complete and result is not None and result.loss_count == 0:
+                logger.debug(u"Good trial acts as a ramp-up")
+                self.ramp_up_start = trial_start
+                self.ramp_up_stop = trial_end
+            else:
+                logger.debug(u"Loss or incomplete, does not act as a ramp-up.")
+        return result
+
+    def no_traffic_loss_occurred(self):
+        """Fail if loss occurred in traffic run.
 
-        self._result = line
-        logger.info('TrafficGen result: {0}'.format(self._result))
+        :returns: nothing
+        :raises Exception: If loss occured.
+        """
+        if self._loss is None:
+            raise RuntimeError(u"The traffic generation has not been issued")
+        if self._loss != u"0":
+            raise RuntimeError(f"Traffic loss occurred: {self._loss}")
 
-        self._loss = self._result.split(', ')[3].split('=')[1]
+    def fail_if_no_traffic_forwarded(self):
+        """Fail if no traffic forwarded.
 
-        return self._result
+        TODO: Check number of passed transactions instead.
 
-    def no_traffic_loss_occured(self):
-        """Fail is loss occured in traffic run
-        :return: nothing
+        :returns: nothing
+        :raises Exception: If no traffic forwarded.
         """
+        if self._received is None:
+            raise RuntimeError(u"The traffic generation has not been issued")
+        if self._received == u"0":
+            raise RuntimeError(u"No traffic forwarded")
+
+    def partial_traffic_loss_accepted(
+            self, loss_acceptance, loss_acceptance_type):
+        """Fail if loss is higher then accepted in traffic run.
 
+        :param loss_acceptance: Permitted drop ratio or frames count.
+        :param loss_acceptance_type: Type of permitted loss.
+        :type loss_acceptance: float
+        :type loss_acceptance_type: LossAcceptanceType
+        :returns: nothing
+        :raises Exception: If loss is above acceptance criteria.
+        """
         if self._loss is None:
-            raise Exception('The traffic generation has not been issued')
-        if self._loss != '0':
-            raise Exception('Traffic loss occured: {0}'.format(self._loss))
+            raise Exception(u"The traffic generation has not been issued")
+
+        if loss_acceptance_type == u"percentage":
+            loss = (float(self._loss) / float(self._sent)) * 100
+        elif loss_acceptance_type == u"frames":
+            loss = float(self._loss)
+        else:
+            raise Exception(u"Loss acceptance type not supported")
+
+        if loss > float(loss_acceptance):
+            raise Exception(
+                f"Traffic loss {loss} above loss acceptance: {loss_acceptance}"
+            )
+
+    def _parse_traffic_results(self, stdout):
+        """Parse stdout of scripts into fields of self.
+
+        Block of code to reuse, by sync start, or stop after async.
+
+        :param stdout: Text containing the standard output.
+        :type stdout: str
+        """
+        subtype = check_subtype(self._node)
+        if subtype == NodeSubTypeTG.TREX:
+            # Last line from console output
+            line = stdout.splitlines()[-1]
+            results = line.split(u";")
+            if results[-1] in (u" ", u""):
+                results.pop(-1)
+            self._result = dict()
+            for result in results:
+                key, value = result.split(u"=", maxsplit=1)
+                self._result[key.strip()] = value
+            logger.info(f"TrafficGen results:\n{self._result}")
+            self._received = int(self._result.get(u"total_received"), 0)
+            self._sent = int(self._result.get(u"total_sent", 0))
+            self._loss = int(self._result.get(u"frame_loss", 0))
+            self._approximated_duration = \
+                self._result.get(u"approximated_duration", 0.0)
+            if u"manual" not in str(self._approximated_duration):
+                self._approximated_duration = float(self._approximated_duration)
+            self._latency = list()
+            self._latency.append(self._result.get(u"latency_stream_0(usec)"))
+            self._latency.append(self._result.get(u"latency_stream_1(usec)"))
+            if self._mode == TrexMode.ASTF:
+                self._l7_data = dict()
+                self._l7_data[u"client"] = dict()
+                self._l7_data[u"client"][u"sent"] = \
+                    int(self._result.get(u"client_sent", 0))
+                self._l7_data[u"client"][u"received"] = \
+                    int(self._result.get(u"client_received", 0))
+                self._l7_data[u"client"][u"active_flows"] = \
+                    int(self._result.get(u"client_active_flows", 0))
+                self._l7_data[u"client"][u"established_flows"] = \
+                    int(self._result.get(u"client_established_flows", 0))
+                self._l7_data[u"client"][u"traffic_duration"] = \
+                    float(self._result.get(u"client_traffic_duration", 0.0))
+                self._l7_data[u"client"][u"err_rx_throttled"] = \
+                    int(self._result.get(u"client_err_rx_throttled", 0))
+                self._l7_data[u"client"][u"err_c_nf_throttled"] = \
+                    int(self._result.get(u"client_err_nf_throttled", 0))
+                self._l7_data[u"client"][u"err_flow_overflow"] = \
+                    int(self._result.get(u"client_err_flow_overflow", 0))
+                self._l7_data[u"server"] = dict()
+                self._l7_data[u"server"][u"active_flows"] = \
+                    int(self._result.get(u"server_active_flows", 0))
+                self._l7_data[u"server"][u"established_flows"] = \
+                    int(self._result.get(u"server_established_flows", 0))
+                self._l7_data[u"server"][u"traffic_duration"] = \
+                    float(self._result.get(u"server_traffic_duration", 0.0))
+                self._l7_data[u"server"][u"err_rx_throttled"] = \
+                    int(self._result.get(u"client_err_rx_throttled", 0))
+                if u"udp" in self.traffic_profile:
+                    self._l7_data[u"client"][u"udp"] = dict()
+                    self._l7_data[u"client"][u"udp"][u"connects"] = \
+                        int(self._result.get(u"client_udp_connects", 0))
+                    self._l7_data[u"client"][u"udp"][u"closed_flows"] = \
+                        int(self._result.get(u"client_udp_closed", 0))
+                    self._l7_data[u"client"][u"udp"][u"tx_bytes"] = \
+                        int(self._result.get(u"client_udp_tx_bytes", 0))
+                    self._l7_data[u"client"][u"udp"][u"rx_bytes"] = \
+                        int(self._result.get(u"client_udp_rx_bytes", 0))
+                    self._l7_data[u"client"][u"udp"][u"tx_packets"] = \
+                        int(self._result.get(u"client_udp_tx_packets", 0))
+                    self._l7_data[u"client"][u"udp"][u"rx_packets"] = \
+                        int(self._result.get(u"client_udp_rx_packets", 0))
+                    self._l7_data[u"client"][u"udp"][u"keep_drops"] = \
+                        int(self._result.get(u"client_udp_keep_drops", 0))
+                    self._l7_data[u"client"][u"udp"][u"err_cwf"] = \
+                        int(self._result.get(u"client_err_cwf", 0))
+                    self._l7_data[u"server"][u"udp"] = dict()
+                    self._l7_data[u"server"][u"udp"][u"accepted_flows"] = \
+                        int(self._result.get(u"server_udp_accepts", 0))
+                    self._l7_data[u"server"][u"udp"][u"closed_flows"] = \
+                        int(self._result.get(u"server_udp_closed", 0))
+                    self._l7_data[u"server"][u"udp"][u"tx_bytes"] = \
+                        int(self._result.get(u"server_udp_tx_bytes", 0))
+                    self._l7_data[u"server"][u"udp"][u"rx_bytes"] = \
+                        int(self._result.get(u"server_udp_rx_bytes", 0))
+                    self._l7_data[u"server"][u"udp"][u"tx_packets"] = \
+                        int(self._result.get(u"server_udp_tx_packets", 0))
+                    self._l7_data[u"server"][u"udp"][u"rx_packets"] = \
+                        int(self._result.get(u"server_udp_rx_packets", 0))
+                elif u"tcp" in self.traffic_profile:
+                    self._l7_data[u"client"][u"tcp"] = dict()
+                    self._l7_data[u"client"][u"tcp"][u"initiated_flows"] = \
+                        int(self._result.get(u"client_tcp_connect_inits", 0))
+                    self._l7_data[u"client"][u"tcp"][u"connects"] = \
+                        int(self._result.get(u"client_tcp_connects", 0))
+                    self._l7_data[u"client"][u"tcp"][u"closed_flows"] = \
+                        int(self._result.get(u"client_tcp_closed", 0))
+                    self._l7_data[u"client"][u"tcp"][u"connattempt"] = \
+                        int(self._result.get(u"client_tcp_connattempt", 0))
+                    self._l7_data[u"client"][u"tcp"][u"tx_bytes"] = \
+                        int(self._result.get(u"client_tcp_tx_bytes", 0))
+                    self._l7_data[u"client"][u"tcp"][u"rx_bytes"] = \
+                        int(self._result.get(u"client_tcp_rx_bytes", 0))
+                    self._l7_data[u"server"][u"tcp"] = dict()
+                    self._l7_data[u"server"][u"tcp"][u"accepted_flows"] = \
+                        int(self._result.get(u"server_tcp_accepts", 0))
+                    self._l7_data[u"server"][u"tcp"][u"connects"] = \
+                        int(self._result.get(u"server_tcp_connects", 0))
+                    self._l7_data[u"server"][u"tcp"][u"closed_flows"] = \
+                        int(self._result.get(u"server_tcp_closed", 0))
+                    self._l7_data[u"server"][u"tcp"][u"tx_bytes"] = \
+                        int(self._result.get(u"server_tcp_tx_bytes", 0))
+                    self._l7_data[u"server"][u"tcp"][u"rx_bytes"] = \
+                        int(self._result.get(u"server_tcp_rx_bytes", 0))
+
+    def _get_measurement_result(self):
+        """Return the result of last measurement as ReceiveRateMeasurement.
+
+        Separate function, as measurements can end either by time
+        or by explicit call, this is the common block at the end.
+
+        The target_tr field of ReceiveRateMeasurement is in
+        transactions per second. Transmit count and loss count units
+        depend on the transaction type. Usually they are in transactions
+        per second, or aggregate packets per second.
+
+        TODO: Fail on running or already reported measurement.
+
+        :returns: Structure containing the result of the measurement.
+        :rtype: ReceiveRateMeasurement
+        """
+        try:
+            # Client duration seems to include a setup period
+            # where TRex does not send any packets yet.
+            # Server duration does not include it.
+            server_data = self._l7_data[u"server"]
+            approximated_duration = float(server_data[u"traffic_duration"])
+        except (KeyError, AttributeError, ValueError, TypeError):
+            approximated_duration = None
+        try:
+            if not approximated_duration:
+                approximated_duration = float(self._approximated_duration)
+        except ValueError:  # "manual"
+            approximated_duration = None
+        if not approximated_duration:
+            if self._duration and self._duration > 0:
+                # Known recomputed or target duration.
+                approximated_duration = self._duration
+            else:
+                # It was an explicit stop.
+                if not self._stop_time:
+                    raise RuntimeError(u"Unable to determine duration.")
+                approximated_duration = self._stop_time - self._start_time
+        target_duration = self._target_duration
+        if not target_duration:
+            target_duration = approximated_duration
+        transmit_rate = self._rate
+        if self.transaction_type == u"packet":
+            partial_attempt_count = self._sent
+            expected_attempt_count = self._sent
+            fail_count = self._loss
+        elif self.transaction_type == u"udp_cps":
+            if not self.transaction_scale:
+                raise RuntimeError(u"Add support for no-limit udp_cps.")
+            partial_attempt_count = self._l7_data[u"client"][u"sent"]
+            # We do not care whether TG is slow, it should have attempted all.
+            expected_attempt_count = self.transaction_scale
+            pass_count = self._l7_data[u"client"][u"received"]
+            fail_count = expected_attempt_count - pass_count
+        elif self.transaction_type == u"tcp_cps":
+            if not self.transaction_scale:
+                raise RuntimeError(u"Add support for no-limit tcp_cps.")
+            ctca = self._l7_data[u"client"][u"tcp"][u"connattempt"]
+            partial_attempt_count = ctca
+            # We do not care whether TG is slow, it should have attempted all.
+            expected_attempt_count = self.transaction_scale
+            # From TCP point of view, server/connects counts full connections,
+            # but we are testing NAT session so client/connects counts that
+            # (half connections from TCP point of view).
+            pass_count = self._l7_data[u"client"][u"tcp"][u"connects"]
+            fail_count = expected_attempt_count - pass_count
+        elif self.transaction_type == u"udp_pps":
+            if not self.transaction_scale:
+                raise RuntimeError(u"Add support for no-limit udp_pps.")
+            partial_attempt_count = self._sent
+            expected_attempt_count = self.transaction_scale * self.ppta
+            fail_count = self._loss + (expected_attempt_count - self._sent)
+        elif self.transaction_type == u"tcp_pps":
+            if not self.transaction_scale:
+                raise RuntimeError(u"Add support for no-limit tcp_pps.")
+            partial_attempt_count = self._sent
+            expected_attempt_count = self.transaction_scale * self.ppta
+            # One loss-like scenario happens when TRex receives all packets
+            # on L2 level, but is not fast enough to process them all
+            # at L7 level, which leads to retransmissions.
+            # Those manifest as opackets larger than expected.
+            # A simple workaround is to add absolute difference.
+            # Probability of retransmissions exactly cancelling
+            # packets unsent due to duration stretching is quite low.
+            fail_count = self._loss + abs(expected_attempt_count - self._sent)
+        else:
+            raise RuntimeError(f"Unknown parsing {self.transaction_type!r}")
+        if fail_count < 0 and not self.negative_loss:
+            fail_count = 0
+        measurement = ReceiveRateMeasurement(
+            duration=target_duration,
+            target_tr=transmit_rate,
+            transmit_count=expected_attempt_count,
+            loss_count=fail_count,
+            approximated_duration=approximated_duration,
+            partial_transmit_count=partial_attempt_count,
+        )
+        measurement.latency = self.get_latency_int()
+        return measurement
+
+    def measure(self, duration, transmit_rate):
+        """Run trial measurement, parse and return results.
+
+        The input rate is for transactions. Stateles bidirectional traffic
+        is understood as sequence of (asynchronous) transactions,
+        two packets each.
+
+        The result units depend on test type, generally
+        the count either transactions or packets (aggregated over directions).
+
+        Optionally, this method sleeps if measurement finished before
+        the time specified as duration.
+
+        :param duration: Trial duration [s].
+        :param transmit_rate: Target rate in transactions per second.
+        :type duration: float
+        :type transmit_rate: float
+        :returns: Structure containing the result of the measurement.
+        :rtype: ReceiveRateMeasurement
+        :raises RuntimeError: If TG is not set or if node is not TG
+            or if subtype is not specified.
+        :raises NotImplementedError: If TG is not supported.
+        """
+        duration = float(duration)
+        time_start = time.monotonic()
+        time_stop = time_start + duration
+        if self.resetter:
+            self.resetter()
+        result = self._send_traffic_on_tg_with_ramp_up(
+            duration=duration,
+            rate=transmit_rate,
+            async_call=False,
+        )
+        logger.debug(f"trial measurement result: {result!r}")
+        # In PLRsearch, computation needs the specified time to complete.
+        if self.sleep_till_duration:
+            sleeptime = time_stop - time.monotonic()
+            if sleeptime > 0.0:
+                # TODO: Sometimes we have time to do additional trials here,
+                # adapt PLRsearch to accept all the results.
+                time.sleep(sleeptime)
+        return result
+
+    def set_rate_provider_defaults(
+            self,
+            frame_size,
+            traffic_profile,
+            ppta=1,
+            resetter=None,
+            traffic_directions=2,
+            transaction_duration=0.0,
+            transaction_scale=0,
+            transaction_type=u"packet",
+            duration_limit=0.0,
+            negative_loss=True,
+            sleep_till_duration=False,
+            use_latency=False,
+            ramp_up_rate=None,
+            ramp_up_duration=None,
+            state_timeout=300.0,
+        ):
+        """Store values accessed by measure().
+
+        :param frame_size: Frame size identifier or value [B].
+        :param traffic_profile: Module name as a traffic profile identifier.
+            See GPL/traffic_profiles/trex for implemented modules.
+        :param ppta: Packets per transaction, aggregated over directions.
+            Needed for udp_pps which does not have a good transaction counter,
+            so we need to compute expected number of packets.
+            Default: 1.
+        :param resetter: Callable to reset DUT state for repeated trials.
+        :param traffic_directions: Traffic from packet counting point of view
+            is bi- (2) or uni- (1) directional.
+            Default: 2
+        :param transaction_duration: Total expected time to close transaction.
+        :param transaction_scale: Number of transactions to perform.
+            0 (default) means unlimited.
+        :param transaction_type: An identifier specifying which counters
+            and formulas to use when computing attempted and failed
+            transactions. Default: "packet".
+            TODO: Does this also specify parsing for the measured duration?
+        :param duration_limit: Zero or maximum limit for computed (or given)
+            duration.
+        :param negative_loss: If false, negative loss is reported as zero loss.
+        :param sleep_till_duration: If true and measurement returned faster,
+            sleep until it matches duration. Needed for PLRsearch.
+        :param use_latency: Whether to measure latency during the trial.
+            Default: False.
+        :param ramp_up_rate: Rate to use in ramp-up trials [pps].
+        :param ramp_up_duration: Duration of ramp-up trials [s].
+        :param state_timeout: Time of life of DUT state [s].
+        :type frame_size: str or int
+        :type traffic_profile: str
+        :type ppta: int
+        :type resetter: Optional[Callable[[], None]]
+        :type traffic_directions: int
+        :type transaction_duration: float
+        :type transaction_scale: int
+        :type transaction_type: str
+        :type duration_limit: float
+        :type negative_loss: bool
+        :type sleep_till_duration: bool
+        :type use_latency: bool
+        :type ramp_up_rate: float
+        :type ramp_up_duration: float
+        :type state_timeout: float
+        """
+        self.frame_size = frame_size
+        self.traffic_profile = str(traffic_profile)
+        self.resetter = resetter
+        self.ppta = ppta
+        self.traffic_directions = int(traffic_directions)
+        self.transaction_duration = float(transaction_duration)
+        self.transaction_scale = int(transaction_scale)
+        self.transaction_type = str(transaction_type)
+        self.duration_limit = float(duration_limit)
+        self.negative_loss = bool(negative_loss)
+        self.sleep_till_duration = bool(sleep_till_duration)
+        self.use_latency = bool(use_latency)
+        self.ramp_up_rate = float(ramp_up_rate)
+        self.ramp_up_duration = float(ramp_up_duration)
+        self.state_timeout = float(state_timeout)
+
+
+class OptimizedSearch:
+    """Class to be imported as Robot Library, containing search keywords.
+
+    Aside of setting up measurer and forwarding arguments,
+    the main business is to translate min/max rate from unidir to aggregate.
+    """
+
+    @staticmethod
+    def perform_optimized_ndrpdr_search(
+            frame_size,
+            traffic_profile,
+            minimum_transmit_rate,
+            maximum_transmit_rate,
+            packet_loss_ratio=0.005,
+            final_relative_width=0.005,
+            final_trial_duration=30.0,
+            initial_trial_duration=1.0,
+            number_of_intermediate_phases=2,
+            timeout=720.0,
+            doublings=1,
+            ppta=1,
+            resetter=None,
+            traffic_directions=2,
+            transaction_duration=0.0,
+            transaction_scale=0,
+            transaction_type=u"packet",
+            use_latency=False,
+            ramp_up_rate=None,
+            ramp_up_duration=None,
+            state_timeout=300.0,
+    ):
+        """Setup initialized TG, perform optimized search, return intervals.
+
+        If transaction_scale is nonzero, all non-init trial durations
+        are set to 2.0 (as they do not affect the real trial duration)
+        and zero intermediate phases are used.
+        The initial phase still uses 1.0 seconds, to force remeasurement.
+        That makes initial phase act as a warmup.
+
+        :param frame_size: Frame size identifier or value [B].
+        :param traffic_profile: Module name as a traffic profile identifier.
+            See GPL/traffic_profiles/trex for implemented modules.
+        :param minimum_transmit_rate: Minimal load in transactions per second.
+        :param maximum_transmit_rate: Maximal load in transactions per second.
+        :param packet_loss_ratio: Fraction of packets lost, for PDR [1].
+        :param final_relative_width: Final lower bound transmit rate
+            cannot be more distant that this multiple of upper bound [1].
+        :param final_trial_duration: Trial duration for the final phase [s].
+        :param initial_trial_duration: Trial duration for the initial phase
+            and also for the first intermediate phase [s].
+        :param number_of_intermediate_phases: Number of intermediate phases
+            to perform before the final phase [1].
+        :param timeout: The search will fail itself when not finished
+            before this overall time [s].
+        :param doublings: How many doublings to do in external search step.
+            Default 1 is suitable for fairly stable tests,
+            less stable tests might get better overal duration with 2 or more.
+        :param ppta: Packets per transaction, aggregated over directions.
+            Needed for udp_pps which does not have a good transaction counter,
+            so we need to compute expected number of packets.
+            Default: 1.
+        :param resetter: Callable to reset DUT state for repeated trials.
+        :param traffic_directions: Traffic is bi- (2) or uni- (1) directional.
+            Default: 2
+        :param transaction_duration: Total expected time to close transaction.
+        :param transaction_scale: Number of transactions to perform.
+            0 (default) means unlimited.
+        :param transaction_type: An identifier specifying which counters
+            and formulas to use when computing attempted and failed
+            transactions. Default: "packet".
+        :param use_latency: Whether to measure latency during the trial.
+            Default: False.
+        :param ramp_up_rate: Rate to use in ramp-up trials [pps].
+        :param ramp_up_duration: Duration of ramp-up trials [s].
+        :param state_timeout: Time of life of DUT state [s].
+        :type frame_size: str or int
+        :type traffic_profile: str
+        :type minimum_transmit_rate: float
+        :type maximum_transmit_rate: float
+        :type packet_loss_ratio: float
+        :type final_relative_width: float
+        :type final_trial_duration: float
+        :type initial_trial_duration: float
+        :type number_of_intermediate_phases: int
+        :type timeout: float
+        :type doublings: int
+        :type ppta: int
+        :type resetter: Optional[Callable[[], None]]
+        :type traffic_directions: int
+        :type transaction_duration: float
+        :type transaction_scale: int
+        :type transaction_type: str
+        :type use_latency: bool
+        :type ramp_up_rate: float
+        :type ramp_up_duration: float
+        :type state_timeout: float
+        :returns: Structure containing narrowed down NDR and PDR intervals
+            and their measurements.
+        :rtype: NdrPdrResult
+        :raises RuntimeError: If total duration is larger than timeout.
+        """
+        # we need instance of TrafficGenerator instantiated by Robot Framework
+        # to be able to use trex_stl-*()
+        tg_instance = BuiltIn().get_library_instance(
+            u"resources.libraries.python.TrafficGenerator"
+        )
+        # Overrides for fixed transaction amount.
+        # TODO: Move to robot code? We have two call sites, so this saves space,
+        #       even though this is surprising for log readers.
+        if transaction_scale:
+            initial_trial_duration = 1.0
+            final_trial_duration = 2.0
+            number_of_intermediate_phases = 0
+            timeout += transaction_scale * 3e-4
+        tg_instance.set_rate_provider_defaults(
+            frame_size=frame_size,
+            traffic_profile=traffic_profile,
+            sleep_till_duration=False,
+            ppta=ppta,
+            resetter=resetter,
+            traffic_directions=traffic_directions,
+            transaction_duration=transaction_duration,
+            transaction_scale=transaction_scale,
+            transaction_type=transaction_type,
+            use_latency=use_latency,
+            ramp_up_rate=ramp_up_rate,
+            ramp_up_duration=ramp_up_duration,
+            state_timeout=state_timeout,
+        )
+        algorithm = MultipleLossRatioSearch(
+            measurer=tg_instance,
+            final_trial_duration=final_trial_duration,
+            final_relative_width=final_relative_width,
+            number_of_intermediate_phases=number_of_intermediate_phases,
+            initial_trial_duration=initial_trial_duration,
+            timeout=timeout,
+            doublings=doublings,
+        )
+        result = algorithm.narrow_down_ndr_and_pdr(
+            min_rate=minimum_transmit_rate,
+            max_rate=maximum_transmit_rate,
+            packet_loss_ratio=packet_loss_ratio,
+        )
+        return result
+
+    @staticmethod
+    def perform_soak_search(
+            frame_size,
+            traffic_profile,
+            minimum_transmit_rate,
+            maximum_transmit_rate,
+            plr_target=1e-7,
+            tdpt=0.1,
+            initial_count=50,
+            timeout=7200.0,
+            ppta=1,
+            resetter=None,
+            trace_enabled=False,
+            traffic_directions=2,
+            transaction_duration=0.0,
+            transaction_scale=0,
+            transaction_type=u"packet",
+            use_latency=False,
+            ramp_up_rate=None,
+            ramp_up_duration=None,
+            state_timeout=300.0,
+    ):
+        """Setup initialized TG, perform soak search, return avg and stdev.
+
+        :param frame_size: Frame size identifier or value [B].
+        :param traffic_profile: Module name as a traffic profile identifier.
+            See GPL/traffic_profiles/trex for implemented modules.
+        :param minimum_transmit_rate: Minimal load in transactions per second.
+        :param maximum_transmit_rate: Maximal load in transactions per second.
+        :param plr_target: Fraction of packets lost to achieve [1].
+        :param tdpt: Trial duration per trial.
+            The algorithm linearly increases trial duration with trial number,
+            this is the increment between succesive trials, in seconds.
+        :param initial_count: Offset to apply before the first trial.
+            For example initial_count=50 makes first trial to be 51*tdpt long.
+            This is needed because initial "search" phase of integrator
+            takes significant time even without any trial results.
+        :param timeout: The search will stop after this overall time [s].
+        :param ppta: Packets per transaction, aggregated over directions.
+            Needed for udp_pps which does not have a good transaction counter,
+            so we need to compute expected number of packets.
+            Default: 1.
+        :param resetter: Callable to reset DUT state for repeated trials.
+        :param trace_enabled: True if trace enabled else False.
+            This is very verbose tracing on numeric computations,
+            do not use in production.
+            Default: False
+        :param traffic_directions: Traffic is bi- (2) or uni- (1) directional.
+            Default: 2
+        :param transaction_duration: Total expected time to close transaction.
+        :param transaction_scale: Number of transactions to perform.
+            0 (default) means unlimited.
+        :param transaction_type: An identifier specifying which counters
+            and formulas to use when computing attempted and failed
+            transactions. Default: "packet".
+        :param use_latency: Whether to measure latency during the trial.
+            Default: False.
+        :param ramp_up_rate: Rate to use in ramp-up trials [pps].
+        :param ramp_up_duration: Duration of ramp-up trials [s].
+        :param state_timeout: Time of life of DUT state [s].
+        :type frame_size: str or int
+        :type traffic_profile: str
+        :type minimum_transmit_rate: float
+        :type maximum_transmit_rate: float
+        :type plr_target: float
+        :type initial_count: int
+        :type timeout: float
+        :type ppta: int
+        :type resetter: Optional[Callable[[], None]]
+        :type trace_enabled: bool
+        :type traffic_directions: int
+        :type transaction_duration: float
+        :type transaction_scale: int
+        :type transaction_type: str
+        :type use_latency: bool
+        :type ramp_up_rate: float
+        :type ramp_up_duration: float
+        :type state_timeout: float
+        :returns: Average and stdev of estimated aggregate rate giving PLR.
+        :rtype: 2-tuple of float
+        """
+        tg_instance = BuiltIn().get_library_instance(
+            u"resources.libraries.python.TrafficGenerator"
+        )
+        # Overrides for fixed transaction amount.
+        # TODO: Move to robot code? We have a single call site
+        #       but MLRsearch has two and we want the two to be used similarly.
+        if transaction_scale:
+            # TODO: What is a good value for max scale?
+            # TODO: Scale the timeout with transaction scale.
+            timeout = 7200.0
+        tg_instance.set_rate_provider_defaults(
+            frame_size=frame_size,
+            traffic_profile=traffic_profile,
+            negative_loss=False,
+            sleep_till_duration=True,
+            ppta=ppta,
+            resetter=resetter,
+            traffic_directions=traffic_directions,
+            transaction_duration=transaction_duration,
+            transaction_scale=transaction_scale,
+            transaction_type=transaction_type,
+            use_latency=use_latency,
+            ramp_up_rate=ramp_up_rate,
+            ramp_up_duration=ramp_up_duration,
+            state_timeout=state_timeout,
+        )
+        algorithm = PLRsearch(
+            measurer=tg_instance,
+            trial_duration_per_trial=tdpt,
+            packet_loss_ratio_target=plr_target,
+            trial_number_offset=initial_count,
+            timeout=timeout,
+            trace_enabled=trace_enabled,
+        )
+        result = algorithm.search(
+            min_rate=minimum_transmit_rate,
+            max_rate=maximum_transmit_rate,
+        )
+        return result