-# Copyright (c) 2020 Cisco and/or its affiliates.
+# Copyright (c) 2021 Cisco and/or its affiliates.
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at:
class Constants:
- """Constants used in CSIT.
-
- TODO: Yaml files are easier for humans to edit.
- Figure out how to set the attributes by parsing a file
- that works regardless of current working directory.
- """
+ """Constants used in CSIT."""
# OpenVPP testing directory location at topology nodes
REMOTE_FW_DIR = u"/tmp/openvpp-testing"
# shell scripts location
RESOURCES_LIB_SH = u"resources/libraries/bash"
+ # python scripts location
+ RESOURCES_LIB_PY = u"resources/libraries/python"
+
+ # shell scripts location
+ RESOURCES_TOOLS = u"resources/tools"
+
# Python API provider location
RESOURCES_PAPI_PROVIDER = u"resources/tools/papi/vpp_papi_provider.py"
+ # Templates location
+ RESOURCES_TPL = u"resources/templates"
+
# vat templates location
RESOURCES_TPL_VAT = u"resources/templates/vat"
# Kubernetes templates location
RESOURCES_TPL_K8S = u"resources/templates/kubernetes"
- # KernelVM templates location
- RESOURCES_TPL_VM = u"resources/templates/vm"
-
# Container templates location
RESOURCES_TPL_CONTAINER = u"resources/templates/container"
# VPP Communications Library templates location
RESOURCES_TPL_VCL = u"resources/templates/vcl"
- # HTTP Server www root directory
- RESOURCES_TP_WRK_WWW = u"resources/traffic_profiles/wrk/www"
+ # VPP Communications Library templates location
+ RESOURCES_TPL_TELEMETRY = u"resources/templates/telemetry"
# VPP Communications Library LD_PRELOAD library
VCL_LDPRELOAD_LIBRARY = u"/usr/lib/x86_64-linux-gnu/libvcl_ldpreload.so"
QEMU_VM_KERNEL_INITRD = u"/opt/boot/initrd.img"
# QEMU VM nested image path
- QEMU_VM_IMAGE = u"/var/lib/vm/vhost-nested.img"
+ QEMU_VM_IMAGE = u"/var/lib/vm/image.iso"
# QEMU VM DPDK path
- QEMU_VM_DPDK = u"/opt/dpdk-20.02"
+ QEMU_VM_DPDK = u"/opt/dpdk-21.08"
# Docker container SUT image
- DOCKER_SUT_IMAGE_UBUNTU = u"snergster/csit-sut:latest"
+ DOCKER_SUT_IMAGE_UBUNTU = u"csit_sut-ubuntu2004:local"
# Docker container arm SUT image
- DOCKER_SUT_IMAGE_UBUNTU_ARM = u"snergster/csit-arm-sut:latest"
+ DOCKER_SUT_IMAGE_UBUNTU_ARM = u"csit_sut-ubuntu2004:local"
# TRex install directory
- TREX_INSTALL_DIR = u"/opt/trex-core-2.73"
+ TREX_INSTALL_DIR = u"/opt/trex-core-2.88"
+
+ # TRex pcap files directory
+ TREX_PCAP_DIR = f"{TREX_INSTALL_DIR}/scripts/avl"
# TRex limit memory.
- TREX_LIMIT_MEMORY = get_int_from_env(u"TREX_LIMIT_MEMORY", 4096)
+ TREX_LIMIT_MEMORY = get_int_from_env(u"TREX_LIMIT_MEMORY", 8192)
# TRex number of cores
- TREX_CORE_COUNT = get_int_from_env(u"TREX_CORE_COUNT", 7)
+ TREX_CORE_COUNT = get_int_from_env(u"TREX_CORE_COUNT", 8)
- # Trex force start regardles ports state
+ # TRex set number of RX/TX descriptors
+ # Set to 0 to use default values
+ TREX_TX_DESCRIPTORS_COUNT = get_int_from_env(
+ u"TREX_TX_DESCRIPTORS_COUNT", 0
+ )
+ TREX_RX_DESCRIPTORS_COUNT = get_int_from_env(
+ u"TREX_RX_DESCRIPTORS_COUNT", 0
+ )
+
+ # Trex force start regardless ports state
TREX_SEND_FORCE = get_pessimistic_bool_from_env(u"TREX_SEND_FORCE")
# TRex extra commandline arguments
- TREX_EXTRA_CMDLINE = get_str_from_env(u"TREX_EXTRA_CMDLINE", u"")
+ TREX_EXTRA_CMDLINE = get_str_from_env(
+ u"TREX_EXTRA_CMDLINE", u"--mbuf-factor 32")
+
+ # graph node variant value
+ GRAPH_NODE_VARIANT = get_str_from_env(
+ u"GRAPH_NODE_VARIANT", u"")
# Sysctl kernel.core_pattern
KERNEL_CORE_PATTERN = u"/tmp/%p-%u-%g-%s-%t-%h-%e.core"
# Core dump directory
CORE_DUMP_DIR = u"/tmp"
+ # Perf stat events (comma separated).
+ PERF_STAT_EVENTS = get_str_from_env(
+ u"PERF_STAT_EVENTS",
+ u"cpu-clock,context-switches,cpu-migrations,page-faults,"
+ u"cycles,instructions,branches,branch-misses,L1-icache-load-misses")
+
# Equivalent to ~0 used in vpp code
BITWISE_NON_ZERO = 0xffffffff
# Number of trials to execute in MRR test.
PERF_TRIAL_MULTIPLICITY = get_int_from_env(u"PERF_TRIAL_MULTIPLICITY", 10)
- # Duration of one trial in MRR test.
+ # Duration [s] of one trial in MRR test.
PERF_TRIAL_DURATION = get_float_from_env(u"PERF_TRIAL_DURATION", 1.0)
+ # Whether to use latency streams in main search trials.
+ PERF_USE_LATENCY = get_pessimistic_bool_from_env(u"PERF_USE_LATENCY")
+
# Duration of one latency-specific trial in NDRPDR test.
PERF_TRIAL_LATENCY_DURATION = get_float_from_env(
u"PERF_TRIAL_LATENCY_DURATION", 5.0)
+ # For some testbeds TG takes longer than usual to start sending traffic.
+ # This constant [s] allows longer wait, without affecting
+ # the approximate duration. For example, use 0.098 for AWS.
+ PERF_TRIAL_STL_DELAY = get_float_from_env(u"PERF_TRIAL_STL_DELAY", 0.0)
+
+ # ASTF usually needs a different value for the delay.
+ PERF_TRIAL_ASTF_DELAY = get_float_from_env(
+ u"PERF_TRIAL_ASTF_DELAY", 0.112
+ )
+
+ # Extended debug (incl. vpp packet trace, linux perf stat, ...).
+ # Full list is available as suite variable (__init__.robot) or is
+ # override by test.
+ EXTENDED_DEBUG = get_pessimistic_bool_from_env(u"EXTENDED_DEBUG")
+
# UUID string of DUT1 /tmp volume created outside of the
# DUT1 docker in case of vpp-device test. ${EMPTY} value means that
# /tmp directory is inside the DUT1 docker.
# Mapping from NIC name to its bps limit.
NIC_NAME_TO_BPS_LIMIT = {
- u"Cisco-VIC-1227": 10000000000,
- u"Cisco-VIC-1385": 24500000000,
u"Intel-X520-DA2": 10000000000,
u"Intel-X553": 10000000000,
u"Intel-X710": 10000000000,
u"Intel-XL710": 24500000000,
u"Intel-XXV710": 24500000000,
+ u"Intel-E810CQ": 100000000000,
u"Mellanox-CX556A": 100000000000,
u"Amazon-Nitro-50G": 10000000000,
u"virtual": 100000000,
# Mapping from NIC name to its pps limit.
NIC_NAME_TO_PPS_LIMIT = {
- u"Cisco-VIC-1227": 14880952,
- u"Cisco-VIC-1385": 18750000,
u"Intel-X520-DA2": 14880952,
u"Intel-X553": 14880952,
u"Intel-X710": 14880952,
u"Intel-XL710": 18750000,
u"Intel-XXV710": 18750000,
- u"Mellanox-CX556A": 60000000, # 148809523,
- u"Amazon-Nitro-50G": 1500000,
+ u"Intel-E810CQ": 58500000,
+ # 2n-clx testbeds show duration stretching on high rates,
+ # depending on encapsulation TRex has to generate.
+ # 40 Mpps is still too much for dot1q (~8% stretching).
+ # 36 Mpps is around the maximal VPP throughput (l2patch 4c8t).
+ # Vxlan traffic will still show stretching at 36 Mpps (>12%),
+ # but we do not care about those tests that much.
+ u"Mellanox-CX556A": 36000000, # 148809523,
+ u"Amazon-Nitro-50G": 1200000,
u"virtual": 14880952,
}
# Suite file names use codes for NICs.
NIC_NAME_TO_CODE = {
- u"Cisco-VIC-1227": u"10ge2p1vic1227",
- u"Cisco-VIC-1385": u"40ge2p1vic1385",
u"Intel-X520-DA2": u"10ge2p1x520",
u"Intel-X553": u"10ge2p1x553",
u"Intel-X710": u"10ge2p1x710",
u"Intel-XL710": u"40ge2p1xl710",
u"Intel-XXV710": u"25ge2p1xxv710",
+ u"Intel-E810CQ": u"100ge2p1e810cq",
u"Amazon-Nitro-50G": u"50ge1p1ENA",
u"Mellanox-CX556A": u"100ge2p1cx556a",
}
# Not each driver is supported by each NIC.
NIC_NAME_TO_DRIVER = {
- u"Cisco-VIC-1227": [u"vfio-pci"],
- u"Cisco-VIC-1385": [u"vfio-pci"],
- u"Intel-X520-DA2": [u"vfio-pci"],
- u"Intel-X553": [u"vfio-pci"],
- u"Intel-X710": [u"vfio-pci", u"avf"],
- u"Intel-XL710": [u"vfio-pci", u"avf"],
- u"Intel-XXV710": [u"vfio-pci", u"avf"],
+ u"Intel-X520-DA2": [u"vfio-pci", u"af_xdp"],
+ u"Intel-X553": [u"vfio-pci", u"af_xdp"],
+ u"Intel-X710": [u"vfio-pci", u"avf", u"af_xdp"],
+ u"Intel-XL710": [u"vfio-pci", u"avf", u"af_xdp"],
+ u"Intel-XXV710": [u"vfio-pci", u"avf", u"af_xdp"],
+ u"Intel-E810CQ": [u"vfio-pci", u"avf", u"af_xdp"],
u"Amazon-Nitro-50G": [u"vfio-pci"],
- u"Mellanox-CX556A": [u"rdma-core"],
+ u"Mellanox-CX556A": [u"rdma-core", u"af_xdp"],
}
# Each driver needs different prugin to work.
u"vfio-pci": u"dpdk_plugin.so",
u"avf": u"avf_plugin.so",
u"rdma-core": u"rdma_plugin.so",
+ u"af_xdp": u"af_xdp_plugin.so",
}
# Tags to differentiate tests for different NIC driver.
u"vfio-pci": u"DRV_VFIO_PCI",
u"avf": u"DRV_AVF",
u"rdma-core": u"DRV_RDMA_CORE",
+ u"af_xdp": u"DRV_AF_XDP",
}
# Suite names have to be different, add prefix.
u"vfio-pci": u"",
u"avf": u"avf-",
u"rdma-core": u"rdma-",
+ u"af_xdp": u"af-xdp-",
+ }
+
+ # Number of virtual functions of physical nic.
+ NIC_DRIVER_TO_VFS = {
+ u"vfio-pci": u"nic_vfs}= | 0",
+ u"avf": u"nic_vfs}= | 1",
+ u"rdma-core": u"nic_vfs}= | 0",
+ u"af_xdp": u"nic_vfs}= | 0",
+ }
+
+ # Not each driver is supported by each NIC.
+ DPDK_NIC_NAME_TO_DRIVER = {
+ u"Intel-X520-DA2": [u"vfio-pci"],
+ u"Intel-X553": [u"vfio-pci"],
+ u"Intel-X710": [u"vfio-pci"],
+ u"Intel-XL710": [u"vfio-pci"],
+ u"Intel-XXV710": [u"vfio-pci"],
+ u"Intel-E810CQ": [u"vfio-pci"],
+ u"Amazon-Nitro-50G": [u"vfio-pci"],
+ u"Mellanox-CX556A": [u"mlx5_core"],
+ }
+
+ # Tags to differentiate tests for different NIC driver.
+ DPDK_NIC_DRIVER_TO_TAG = {
+ u"vfio-pci": u"DRV_VFIO_PCI",
+ u"mlx5_core": u"DRV_MLX5_CORE",
+ }
+
+ # Suite names have to be different, add prefix.
+ DPDK_NIC_DRIVER_TO_SUITE_PREFIX = {
+ u"vfio-pci": u"",
+ u"mlx5_core": u"mlx5-",
}
# Some identifiers constructed from suite names
FORBIDDEN_SUITE_PREFIX_LIST = [
prefix for prefix in NIC_DRIVER_TO_SUITE_PREFIX.values() if prefix
]
-
- # Additional step for perf needs to know driver type.
- # Contains part of suite setup line, matching both single and double link.
- NIC_DRIVER_TO_SETUP_ARG = {
- u"vfio-pci": u"le link | performance",
- u"avf": u"le link | performance_avf",
- u"rdma-core": u"le link | performance_rdma",
- }
+ FORBIDDEN_SUITE_PREFIX_LIST += [
+ prefix for prefix in DPDK_NIC_DRIVER_TO_SUITE_PREFIX.values() if prefix
+ ]
# TODO CSIT-1481: Crypto HW should be read from topology file instead.
NIC_NAME_TO_CRYPTO_HW = {
u"Intel-XL710": u"HW_DH895xcc",
}
+ DEVICE_TYPE_TO_KEYWORD = {
+ u"scapy": None
+ }
+
PERF_TYPE_TO_KEYWORD = {
u"mrr": u"Traffic should pass with maximum rate",
u"ndrpdr": u"Find NDR and PDR intervals using optimized search",
PERF_TYPE_TO_SUITE_DOC_VER = {
u"mrr": u'''fication:* In MaxReceivedRate tests TG sends traffic\\
| ... | at line rate and reports total received packets over trial period.\\''',
- # TODO: Figure out how to include the full "*[Ver] TG verification:*"
- # while keeping this readable and without breaking line length limit.
u"ndrpdr": u'''ication:* TG finds and reports throughput NDR (Non Drop\\
| ... | Rate) with zero packet loss tolerance and throughput PDR (Partial Drop\\
| ... | Rate) with non-zero packet loss tolerance (LT) expressed in percentage\\