X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=docs%2Freport%2Fvpp_performance_tests%2Foverview.rst;h=d835df4684ebbd194a8b683fa44a0c290b726f98;hb=0f6410d3c95eb28164fdd349778155c6525e1a79;hp=c1302866fc2cbf1218aba18814508b52420cdf23;hpb=737fbd3341638e90478979779b5ee6c0b0bf5c39;p=csit.git diff --git a/docs/report/vpp_performance_tests/overview.rst b/docs/report/vpp_performance_tests/overview.rst index c1302866fc..d835df4684 100644 --- a/docs/report/vpp_performance_tests/overview.rst +++ b/docs/report/vpp_performance_tests/overview.rst @@ -54,7 +54,7 @@ performance labs to address larger scale multi-interface and multi-NIC performance testing scenarios. For test cases that require DUT (VPP) to communicate with -VirtualMachines (VMs) / Containers (Linux or Docker Containers) over +VirtualMachines (VMs) / Linux or Docker Containers (Ctrs) over vhost-user/memif interfaces, N of VM/Ctr instances are created on SUT1 and SUT2. For N=1 DUT forwards packets between vhost/memif and physical interfaces. For N>1 DUT a logical service chain forwarding topology is @@ -104,9 +104,14 @@ much higher dependency on intensive memory operations and sensitivity to Linux kernel scheduler settings and behaviour, this estimation may not always yield good enough accuracy. -For detailed :abbr:`LF (Linux Foundation)` FD.io test bed specification and -physical topology please refer to `LF FD.io CSIT testbed wiki page -`_. +For detailed FD.io CSIT testbed specification and topology, as well as +configuration and setup of SUTs and DUTs testbeds please refer to +:ref:`test_environment`. + +Similar SUT compute node and DUT VPP settings can be arrived to in a +standalone VPP setup by using a `vpp-config configuration tool +`_ developed within the +VPP project using CSIT recommended settings and scripts. Performance Tests Coverage -------------------------- @@ -157,7 +162,7 @@ CSIT |release| includes following performance test suites, listed per NIC type: number of users and ports per user. - **Container memif connections** - VPP memif virtual interface tests to interconnect VPP instances with L2XC and L2BD. - - **Container Orchestrated Topologies** - Container topologies connected over + - **Container K8s Orchestrated Topologies** - Container topologies connected over the memif virtual interface. - 2port40GE XL710 Intel @@ -170,10 +175,14 @@ CSIT |release| includes following performance test suites, listed per NIC type: - **VMs with vhost-user** - virtual topologies with 1 VM and service chains of 2 VMs using vhost-user interfaces, with VPP forwarding modes incl. L2 Cross-Connect, L2 Bridge-Domain, VXLAN with L2BD, IPv4 routed-forwarding. - - **IPSec** - IPSec encryption with AES-GCM, CBC-SHA1 ciphers, in combination - with IPv4 routed-forwarding. + - **IPSecSW** - IPSec encryption with AES-GCM, CBC-SHA1 ciphers, in + combination with IPv4 routed-forwarding. + - **IPSecHW** - IPSec encryption with AES-GCM, CBC-SHA1 ciphers, in + combination with IPv4 routed-forwarding. Intel QAT HW acceleration. - **IPSec+LISP** - IPSec encryption with CBC-SHA1 ciphers, in combination with LISP-GPE overlay tunneling for IPv4-over-IPv4. + - **VPP TCP/IP stack** - VPP builtin TCP based HTTP server. WRK traffic + generator is used. - 2port10GE X710 Intel @@ -394,7 +403,7 @@ TRex is installed and run on the TG compute node. The typical procedure is: - TRex is started in the background mode :: - $ sh -c 'cd /opt/trex-core-2.25/scripts/ && sudo nohup ./t-rex-64 -i -c 7 --iom 0 > /dev/null 2>&1 &' > /dev/null + $ sh -c 'cd /scripts/ && sudo nohup ./t-rex-64 -i -c 7 --iom 0 > /tmp/trex.log 2>&1 &' > /dev/null - There are traffic streams dynamically prepared for each test, based on traffic profiles. The traffic is sent and the statistics obtained using