+| | | Import Library | resources.libraries.python.VppConfigGenerator
+| | | ... | WITH NAME | ${dut}
+| | | Run keyword | ${dut}.Set Node | ${nodes['${dut}']}
+| | | Run keyword | ${dut}.Add Unix Log
+| | | Run keyword | ${dut}.Add Unix CLI Listen
+| | | Run keyword | ${dut}.Add Unix Nodaemon
+| | | Run keyword | ${dut}.Add DPDK Socketmem | "1024,1024"
+| | | Run keyword | ${dut}.Add Heapsize | "3G"
+| | | Run keyword | ${dut}.Add IP6 Hash Buckets | "2000000"
+| | | Run keyword | ${dut}.Add IP6 Heap Size | "3G"
+
+| Add '${m}' worker threads and '${n}' rxqueues in 3-node single-link circular topology
+| | [Documentation] | Setup M worker threads and N rxqueues in vpp startup\
+| | ... | configuration on all DUTs in 3-node single-link topology.
+| | ...
+| | ${m_int}= | Convert To Integer | ${m}
+| | ${dut1_numa}= | Get interfaces numa node | ${dut1}
+| | ... | ${dut1_if1} | ${dut1_if2}
+| | ${dut2_numa}= | Get interfaces numa node | ${dut2}
+| | ... | ${dut2_if1} | ${dut2_if2}
+| | ${dut1_cpu_main}= | Cpu list per node str | ${dut1} | ${dut1_numa}
+| | ... | skip_cnt=${1} | cpu_cnt=${1}
+| | ${dut1_cpu_w}= | Cpu list per node str | ${dut1} | ${dut1_numa}
+| | ... | skip_cnt=${2} | cpu_cnt=${m_int}
+| | ${dut2_cpu_main}= | Cpu list per node str | ${dut2} | ${dut2_numa}
+| | ... | skip_cnt=${1} | cpu_cnt=${1}
+| | ${dut2_cpu_w}= | Cpu list per node str | ${dut2} | ${dut2_numa}
+| | ... | skip_cnt=${2} | cpu_cnt=${m_int}
+| | Run keyword | DUT1.Add CPU Main Core | ${dut1_cpu_main}
+| | Run keyword | DUT2.Add CPU Main Core | ${dut2_cpu_main}
+| | Run keyword | DUT1.Add CPU Corelist Workers | ${dut1_cpu_w}
+| | Run keyword | DUT2.Add CPU Corelist Workers | ${dut2_cpu_w}
+| | Run keyword | DUT1.Add DPDK Dev Default RXQ | ${n}
+| | Run keyword | DUT2.Add DPDK Dev Default RXQ | ${n}
+
+| Add '${m}' worker threads and '${n}' rxqueues in 2-node single-link circular topology
+| | [Documentation] | Setup M worker threads and N rxqueues in vpp startup\
+| | ... | configuration on all DUTs in 2-node single-link topology.
+| | ...
+| | ${m_int}= | Convert To Integer | ${m}
+| | ${dut1_numa}= | Get interfaces numa node | ${dut1}
+| | ... | ${dut1_if1} | ${dut1_if2}
+| | ${dut1_cpu_main}= | Cpu list per node str | ${dut1} | ${dut1_numa}
+| | ... | skip_cnt=${1} | cpu_cnt=${1}
+| | ${dut1_cpu_w}= | Cpu list per node str | ${dut1} | ${dut1_numa}
+| | ... | skip_cnt=${2} | cpu_cnt=${m_int}
+| | Run keyword | DUT1.Add CPU Main Core | ${dut1_cpu_main}
+| | Run keyword | DUT1.Add CPU Corelist Workers | ${dut1_cpu_w}
+| | Run keyword | DUT1.Add DPDK Dev Default RXQ | ${n}
+
+| Add '${m}' worker threads using SMT and '${n}' rxqueues in 3-node single-link circular topology
+| | [Documentation] | Setup M worker threads using SMT and N rxqueues in vpp\
+| | ... | startup configuration on all DUTs in 3-node single-link topology.
+| | ...
+| | ${m_int}= | Convert To Integer | ${m}
+| | ${dut1_numa}= | Get interfaces numa node | ${dut1}
+| | ... | ${dut1_if1} | ${dut1_if2}
+| | ${dut2_numa}= | Get interfaces numa node | ${dut2}
+| | ... | ${dut2_if1} | ${dut2_if2}
+| | ${dut1_cpu_main}= | Cpu list per node str | ${dut1} | ${dut1_numa}
+| | ... | skip_cnt=${1} | cpu_cnt=${1} | smt_used=${True}
+| | ${dut1_cpu_w}= | Cpu list per node str | ${dut1} | ${dut1_numa}
+| | ... | skip_cnt=${2} | cpu_cnt=${m_int} | smt_used=${True}
+| | ${dut2_cpu_main}= | Cpu list per node str | ${dut2} | ${dut2_numa}
+| | ... | skip_cnt=${1} | cpu_cnt=${1} | smt_used=${True}
+| | ${dut2_cpu_w}= | Cpu list per node str | ${dut2} | ${dut2_numa}
+| | ... | skip_cnt=${2} | cpu_cnt=${m_int} | smt_used=${True}
+| | Run keyword | DUT1.Add CPU Main Core | ${dut1_cpu_main}
+| | Run keyword | DUT2.Add CPU Main Core | ${dut2_cpu_main}
+| | Run keyword | DUT1.Add CPU Corelist Workers | ${dut1_cpu_w}
+| | Run keyword | DUT2.Add CPU Corelist Workers | ${dut2_cpu_w}
+| | Run keyword | DUT1.Add DPDK Dev Default RXQ | ${n}
+| | Run keyword | DUT2.Add DPDK Dev Default RXQ | ${n}
+
+| Add '${m}' worker threads using SMT and '${n}' rxqueues in 2-node single-link circular topology
+| | [Documentation] | Setup M worker threads and N rxqueues in vpp startup\
+| | ... | configuration on all DUTs in 2-node single-link topology.
+| | ...
+| | ${m_int}= | Convert To Integer | ${m}
+| | ${dut1_numa}= | Get interfaces numa node | ${dut1}
+| | ... | ${dut1_if1} | ${dut1_if2}
+| | ${dut1_cpu_main}= | Cpu list per node str | ${dut1} | ${dut1_numa}
+| | ... | skip_cnt=${1} | cpu_cnt=${1} | smt_used=${True}
+| | ${dut1_cpu_w}= | Cpu list per node str | ${dut1} | ${dut1_numa}
+| | ... | skip_cnt=${2} | cpu_cnt=${m_int} | smt_used=${True}
+| | Run keyword | DUT1.Add CPU Main Core | ${dut1_cpu_main}
+| | Run keyword | DUT1.Add CPU Corelist Workers | ${dut1_cpu_w}
+| | Run keyword | DUT1.Add DPDK Dev Default RXQ | ${n}
+
+| Add no multi seg to all DUTs
+| | [Documentation] | Add No Multi Seg to VPP startup configuration to all DUTs.
+| | ...
+| | ${duts}= | Get Matches | ${nodes} | DUT*
+| | :FOR | ${dut} | IN | @{duts}
+| | | Run keyword | ${dut}.Add DPDK No Multi Seg
+
+| Add SNAT to all DUTs
+| | [Documentation] | Add SNAT configuration to all DUTs.
+| | ...
+| | ${duts}= | Get Matches | ${nodes} | DUT*
+| | :FOR | ${dut} | IN | @{duts}
+| | | Run keyword | ${dut}.Add SNAT
+
+| Add cryptodev to all DUTs
+| | [Documentation] | Add Cryptodev to VPP startup configuration to all DUTs.
+| | ...
+| | ... | *Arguments:*
+| | ... | - ${count} - Number of QAT devices. Type: integer
+| | ...
+| | ... | *Example:*
+| | ...
+| | ... | \| Add cryptodev to all DUTs \| ${4} \|
+| | ...
+| | [Arguments] | ${count}
+| | ${duts}= | Get Matches | ${nodes} | DUT*
+| | :FOR | ${dut} | IN | @{duts}
+| | | Run keyword | ${dut}.Add DPDK Cryptodev | ${count}
+
+| Apply startup configuration on all VPP DUTs
+| | [Documentation] | Write startup configuration and restart VPP on all DUTs.
+| | ...
+| | ${duts}= | Get Matches | ${nodes} | DUT*
+| | :FOR | ${dut} | IN | @{duts}
+| | | Run keyword | ${dut}.Apply Config
+| | Update All Interface Data On All Nodes | ${nodes} | skip_tg=${TRUE}
+
+| Save VPP PIDs
+| | [Documentation] | Get PIDs of VPP processes from all DUTs in topology and\
+| | ... | set it as a test variable. The PIDs are stored as dictionary items\
+| | ... | where the key is the host and the value is the PID.
+| | ...
+| | ${setup_vpp_pids}= | Get VPP PIDs | ${nodes}
+| | ${keys}= | Get Dictionary Keys | ${setup_vpp_pids}
+| | :FOR | ${key} | IN | @{keys}
+| | | ${pid}= | Get From Dictionary | ${setup_vpp_pids} | ${key}
+| | | Run Keyword If | $pid is None | FAIL | No VPP PID found on node ${key}
+| | | Run Keyword If | ',' in '${pid}'
+| | | ... | FAIL | More then one VPP PID found on node ${key}: ${pid}
+| | Set Test Variable | ${setup_vpp_pids}
+
+| Verify VPP PID in Teardown
+| | [Documentation] | Check if the VPP PIDs on all DUTs are the same at the end\
+| | ... | of test as they were at the begining. If they are not, only a message\
+| | ... | is printed on console and to log. The test will not fail.
+| | ...
+| | ${teardown_vpp_pids}= | Get VPP PIDs | ${nodes}
+| | ${err_msg}= | Catenate | ${SUITE NAME} - ${TEST NAME}
+| | ... | \nThe VPP PIDs are not equal!\nTest Setup VPP PIDs:
+| | ... | ${setup_vpp_pids}\nTest Teardown VPP PIDs: ${teardown_vpp_pids}
+| | ${rc} | ${msg}= | Run keyword and ignore error
+| | ... | Dictionaries Should Be Equal
+| | ... | ${setup_vpp_pids} | ${teardown_vpp_pids}
+| | Run Keyword And Return If | '${rc}'=='FAIL' | Log | ${err_msg}
+| | ... | console=yes | level=WARN
+
+| Set up functional test
+| | [Documentation] | Common test setup for functional tests.
+| | ...
+| | Configure all DUTs before test
+| | Save VPP PIDs
+| | Configure all TGs for traffic script
+| | Update All Interface Data On All Nodes | ${nodes}
+| | Reset VAT History On All DUTs | ${nodes}