Looks like disabling all plugins in container leaving
only memif_plugin.so results into issue that could lead
into lost of SSH connectivity to TB or in better case
to not handling any traffic.
This fix is supose to return previous behavior for Container
instance of VPP.
Change-Id: Ie8b3ce2e35a40c8f3498e490abb6967cd4dd912a
Signed-off-by: Peter Mikus <pmikus@cisco.com>
set -x
# Space separated list of available testbeds, described by topology files
-TOPOLOGIES="topologies/available/lf_testbed1.yaml \
- topologies/available/lf_testbed2.yaml \
- topologies/available/lf_testbed3.yaml"
+TOPOLOGIES="topologies/available/lf_testbed3.yaml"
SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-s "tests.vpp.perf" \
+ -i THIS \
tests/
RETURN_STATUS=$(echo $?)
esac
if cpuset_cpus:
corelist_workers = ','.join(str(cpu) for cpu in cpuset_cpus)
vpp_config.add_cpu_corelist_workers(corelist_workers)
- vpp_config.add_plugin('disable', 'default')
- vpp_config.add_plugin('enable', 'memif_plugin.so')
+ vpp_config.add_plugin('disable', 'dpdk_plugin.so')
self.execute('mkdir -p /etc/vpp/')
self.execute('echo "{c}" | tee {f}'
| | ... | [Ver] Measure MaxReceivedRate for 64B frames using single trial\
| | ... | throughput test.
| | ...
-| | [Tags] | 64B | 1T1C | STHREAD
+| | [Tags] | 64B | 1T1C | STHREAD | THIS
| | [Template] | Check RR for l2xcbase-eth-2memif-1dcr
| | framesize=${64} | wt=1 | rxq=1