X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=resources%2Flibraries%2Fbash%2Ffunction%2Fcommon.sh;h=e0ae3f2fceaa5f92ffe78958dc5c8475e2bb2130;hb=cbb459aa51ef6e7fb0584bc2cd5a6a7621d7b4c2;hp=ed3b2044d3d6d018e95c2260a2e501c34efa6bc1;hpb=6da5a6920171682bd5bf6a77517bedfef91cbd0e;p=csit.git diff --git a/resources/libraries/bash/function/common.sh b/resources/libraries/bash/function/common.sh index ed3b2044d3..e0ae3f2fce 100644 --- a/resources/libraries/bash/function/common.sh +++ b/resources/libraries/bash/function/common.sh @@ -1,5 +1,5 @@ -# Copyright (c) 2021 Cisco and/or its affiliates. -# Copyright (c) 2021 PANTHEON.tech and/or its affiliates. +# Copyright (c) 2023 Cisco and/or its affiliates. +# Copyright (c) 2023 PANTHEON.tech and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: @@ -45,7 +45,7 @@ function activate_docker_topology () { device_image="$(< ${CSIT_DIR}/${IMAGE_VER_FILE})" case_text="${NODENESS}_${FLAVOR}" case "${case_text}" in - "1n_skx" | "1n_tx2") + "1n_skx" | "1n_tx2" | "1n_spr") # We execute reservation over csit-shim-dcr (ssh) which runs sourced # script's functions. Env variables are read from ssh output # back to localhost for further processing. @@ -79,12 +79,14 @@ function activate_docker_topology () { die "Trap attempt failed, please cleanup manually. Aborting!" } + parse_env_variables || die "Parse of environment variables failed!" + # Replace all variables in template with those in environment. source <(echo 'cat <topo.yml'; cat ${TOPOLOGIES[0]}; echo EOF;) || { die "Topology file create failed!" } - WORKING_TOPOLOGY="/tmp/topology.yaml" + WORKING_TOPOLOGY="${CSIT_DIR}/topologies/available/vpp_device.yaml" mv topo.yml "${WORKING_TOPOLOGY}" || { die "Topology move failed!" } @@ -116,7 +118,7 @@ function activate_virtualenv () { env_dir="${root_path}/env" req_path=${2-$CSIT_DIR/requirements.txt} rm -rf "${env_dir}" || die "Failed to clean previous virtualenv." - pip3 install virtualenv==20.0.20 || { + pip3 install virtualenv==20.15.1 || { die "Virtualenv package install failed." } virtualenv --no-download --python=$(which python3) "${env_dir}" || { @@ -135,19 +137,19 @@ function activate_virtualenv () { function archive_tests () { - # Create .tar.xz of generated/tests for archiving. + # Create .tar.gz of generated/tests for archiving. # To be run after generate_tests, kept separate to offer more flexibility. # Directory read: # - ${GENERATED_DIR}/tests - Tree of executed suites to archive. # File rewriten: - # - ${ARCHIVE_DIR}/tests.tar.xz - Archive of generated tests. + # - ${ARCHIVE_DIR}/generated_tests.tar.gz - Archive of generated tests. set -exuo pipefail - tar c "${GENERATED_DIR}/tests" | xz -3 > "${ARCHIVE_DIR}/tests.tar.xz" || { - die "Error creating archive of generated tests." - } + pushd "${ARCHIVE_DIR}" || die + tar czf "generated_tests.tar.gz" "${GENERATED_DIR}/tests" || true + popd || die } @@ -156,7 +158,7 @@ function check_download_dir () { # Fail if there are no files visible in ${DOWNLOAD_DIR}. # # Variables read: - # - DOWNLOAD_DIR - Path to directory pybot takes the build to test from. + # - DOWNLOAD_DIR - Path to directory robot takes the build to test from. # Directories read: # - ${DOWNLOAD_DIR} - Has to be non-empty to proceed. # Functions called: @@ -225,7 +227,7 @@ function common_dirs () { TOPOLOGIES_DIR=$(readlink -e "${CSIT_DIR}/topologies/available") || { die "Readlink failed." } - JOB_SPECS_DIR=$(readlink -e "${CSIT_DIR}/docs/job_specs") || { + JOB_SPECS_DIR=$(readlink -e "${CSIT_DIR}/resources/job_specs") || { die "Readlink failed." } RESOURCES_DIR=$(readlink -e "${CSIT_DIR}/resources") || { @@ -234,9 +236,6 @@ function common_dirs () { TOOLS_DIR=$(readlink -e "${RESOURCES_DIR}/tools") || { die "Readlink failed." } - DOC_GEN_DIR=$(readlink -e "${TOOLS_DIR}/doc_gen") || { - die "Readlink failed." - } PYTHON_SCRIPTS_DIR=$(readlink -e "${TOOLS_DIR}/scripts") || { die "Readlink failed." } @@ -256,7 +255,7 @@ function common_dirs () { } -function compose_pybot_arguments () { +function compose_robot_arguments () { # Variables read: # - WORKING_TOPOLOGY - Path to topology yaml file of the reserved testbed. @@ -266,21 +265,21 @@ function compose_pybot_arguments () { # - TEST_CODE - The test selection string from environment or argument. # - SELECTION_MODE - Selection criteria [test, suite, include, exclude]. # Variables set: - # - PYBOT_ARGS - String holding part of all arguments for pybot. - # - EXPANDED_TAGS - Array of strings pybot arguments compiled from tags. + # - ROBOT_ARGS - String holding part of all arguments for robot. + # - EXPANDED_TAGS - Array of strings robot arguments compiled from tags. set -exuo pipefail # No explicit check needed with "set -u". - PYBOT_ARGS=("--loglevel" "TRACE") - PYBOT_ARGS+=("--variable" "TOPOLOGY_PATH:${WORKING_TOPOLOGY}") + ROBOT_ARGS=("--loglevel" "TRACE") + ROBOT_ARGS+=("--variable" "TOPOLOGY_PATH:${WORKING_TOPOLOGY}") case "${TEST_CODE}" in *"device"*) - PYBOT_ARGS+=("--suite" "tests.${DUT}.device") + ROBOT_ARGS+=("--suite" "tests.${DUT}.device") ;; *"perf"*) - PYBOT_ARGS+=("--suite" "tests.${DUT}.perf") + ROBOT_ARGS+=("--suite" "tests.${DUT}.perf") ;; *) die "Unknown specification: ${TEST_CODE}" @@ -317,7 +316,7 @@ function deactivate_docker_topology () { case_text="${NODENESS}_${FLAVOR}" case "${case_text}" in - "1n_skx" | "1n_tx2") + "1n_skx" | "1n_tx2" | "1n_spr") ssh="ssh root@172.17.0.1 -p 6022" env_vars=$(env | grep CSIT_ | tr '\n' ' ' ) || die # The "declare -f" output is long and boring. @@ -358,19 +357,19 @@ function die () { } -function die_on_pybot_error () { +function die_on_robot_error () { # Source this fragment if you want to abort on any failed test case. # # Variables read: - # - PYBOT_EXIT_STATUS - Set by a pybot running fragment. + # - ROBOT_EXIT_STATUS - Set by a robot running fragment. # Functions called: # - die - Print to stderr and exit. set -exuo pipefail - if [[ "${PYBOT_EXIT_STATUS}" != "0" ]]; then - die "Test failures are present!" "${PYBOT_EXIT_STATUS}" + if [[ "${ROBOT_EXIT_STATUS}" != "0" ]]; then + die "Test failures are present!" "${ROBOT_EXIT_STATUS}" fi } @@ -383,7 +382,7 @@ function generate_tests () { # within any subdirectory after copying. # This is a separate function, because this code is called - # both by autogen checker and entries calling run_pybot. + # both by autogen checker and entries calling run_robot. # Directories read: # - ${CSIT_DIR}/tests - Used as templates for the generated tests. @@ -437,10 +436,18 @@ function get_test_code () { NODENESS="1n" FLAVOR="skx" ;; - *"1n-tx2"*) + *"1n-spr"*) + NODENESS="1n" + FLAVOR="spr" + ;; + *"1n-tx2"*) NODENESS="1n" FLAVOR="tx2" ;; + *"1n-aws"*) + NODENESS="1n" + FLAVOR="aws" + ;; *"2n-aws"*) NODENESS="2n" FLAVOR="aws" @@ -449,29 +456,61 @@ function get_test_code () { NODENESS="3n" FLAVOR="aws" ;; - *"2n-skx"*) + *"1n-c6gn"*) + NODENESS="1n" + FLAVOR="c6gn" + ;; + *"2n-c6gn"*) NODENESS="2n" - FLAVOR="skx" + FLAVOR="c6gn" ;; - *"2n-zn2"*) + *"3n-c6gn"*) + NODENESS="3n" + FLAVOR="c6gn" + ;; + *"1n-c6in"*) + NODENESS="1n" + FLAVOR="c6in" + ;; + *"2n-c6in"*) NODENESS="2n" - FLAVOR="zn2" + FLAVOR="c6in" ;; - *"3n-skx"*) + *"3n-c6in"*) NODENESS="3n" - FLAVOR="skx" + FLAVOR="c6in" + ;; + *"2n-zn2"*) + NODENESS="2n" + FLAVOR="zn2" ;; *"2n-clx"*) NODENESS="2n" FLAVOR="clx" ;; - *"2n-dnv"*) + *"2n-icx"*) + NODENESS="2n" + FLAVOR="icx" + ;; + *"2n-spr"*) NODENESS="2n" - FLAVOR="dnv" + FLAVOR="spr" + ;; + *"3n-icx"*) + NODENESS="3n" + FLAVOR="icx" ;; - *"3n-dnv"*) + *"3na-spr"*) + NODENESS="3na" + FLAVOR="spr" + ;; + *"3nb-spr"*) + NODENESS="3nb" + FLAVOR="spr" + ;; + *"3n-snr"*) NODENESS="3n" - FLAVOR="dnv" + FLAVOR="snr" ;; *"2n-tx2"*) NODENESS="2n" @@ -481,10 +520,9 @@ function get_test_code () { NODENESS="3n" FLAVOR="tsh" ;; - *) - # Fallback to 3-node Haswell by default (backward compatibility) + *"3n-alt"*) NODENESS="3n" - FLAVOR="hsw" + FLAVOR="alt" ;; esac } @@ -499,6 +537,8 @@ function get_test_tag_string () { # Variables set: # - TEST_TAG_STRING - The string following trigger word in gerrit comment. # May be empty, or even not set on event types not adding comment. + # Variables exported optionally: + # - GRAPH_NODE_VARIANT - Node variant to test with, set if found in trigger. # TODO: ci-management scripts no longer need to perform this. @@ -539,9 +579,6 @@ function get_test_tag_string () { elif [[ "${test_tag_array[0]}" == "skx" ]]; then export GRAPH_NODE_VARIANT="skx" TEST_TAG_STRING="${test_tag_array[@]:1}" || true - elif [[ "${test_tag_array[0]}" == "hsw" ]]; then - export GRAPH_NODE_VARIANT="hsw" - TEST_TAG_STRING="${test_tag_array[@]:1}" || true fi fi fi @@ -607,6 +644,56 @@ function move_archives () { } +function prepare_topology () { + + # Prepare virtual testbed topology if needed based on flavor. + + # Variables read: + # - TEST_CODE - String affecting test selection, usually jenkins job name. + # - NODENESS - Node multiplicity of testbed, either "2n" or "3n". + # - FLAVOR - Node flavor string, e.g. "clx" or "skx". + # Variables set: + # - TERRAFORM_MODULE_DIR - Terraform module directory. + # Functions called: + # - die - Print to stderr and exit. + # - terraform_init - Terraform init topology. + # - terraform_apply - Terraform apply topology. + + set -exuo pipefail + + case_text="${NODENESS}_${FLAVOR}" + case "${case_text}" in + "1n_aws" | "2n_aws" | "3n_aws") + export TF_VAR_testbed_name="${TEST_CODE}" + TERRAFORM_MODULE_DIR="terraform-aws-${NODENESS}-${FLAVOR}-c5n" + terraform_init || die "Failed to call terraform init." + trap "terraform_destroy" ERR EXIT || { + die "Trap attempt failed, please cleanup manually. Aborting!" + } + terraform_apply || die "Failed to call terraform apply." + ;; + "1n_c6gn" | "2n_c6gn" | "3n_c6gn") + export TF_VAR_testbed_name="${TEST_CODE}" + TERRAFORM_MODULE_DIR="terraform-aws-${NODENESS}-c6gn" + terraform_init || die "Failed to call terraform init." + trap "terraform_destroy" ERR EXIT || { + die "Trap attempt failed, please cleanup manually. Aborting!" + } + terraform_apply || die "Failed to call terraform apply." + ;; + "1n_c6in" | "2n_c6in" | "3n_c6in") + export TF_VAR_testbed_name="${TEST_CODE}" + TERRAFORM_MODULE_DIR="terraform-aws-${NODENESS}-c6in" + terraform_init || die "Failed to call terraform init." + trap "terraform_destroy" ERR EXIT || { + die "Trap attempt failed, please cleanup manually. Aborting!" + } + terraform_apply || die "Failed to call terraform apply." + ;; + esac +} + + function reserve_and_cleanup_testbed () { # Reserve physical testbed, perform cleanup, register trap to unreserve. @@ -684,39 +771,31 @@ function reserve_and_cleanup_testbed () { } -function run_pybot () { +function run_robot () { - # Run pybot with options based on input variables. Create output_info.xml + # Run robot with options based on input variables. # # Variables read: # - CSIT_DIR - Path to existing root of local CSIT git repository. # - ARCHIVE_DIR - Path to store robot result files in. - # - PYBOT_ARGS, EXPANDED_TAGS - See compose_pybot_arguments.sh + # - ROBOT_ARGS, EXPANDED_TAGS - See compose_robot_arguments.sh # - GENERATED_DIR - Tests are assumed to be generated under there. # Variables set: - # - PYBOT_EXIT_STATUS - Exit status of most recent pybot invocation. + # - ROBOT_EXIT_STATUS - Exit status of most recent robot invocation. # Functions called: # - die - Print to stderr and exit. set -exuo pipefail - all_options=("--outputdir" "${ARCHIVE_DIR}" "${PYBOT_ARGS[@]}") - all_options+=("--noncritical" "EXPECTED_FAILING") + all_options=("--outputdir" "${ARCHIVE_DIR}" "${ROBOT_ARGS[@]}") all_options+=("${EXPANDED_TAGS[@]}") pushd "${CSIT_DIR}" || die "Change directory operation failed." set +e robot "${all_options[@]}" "${GENERATED_DIR}/tests/" - PYBOT_EXIT_STATUS="$?" + ROBOT_EXIT_STATUS="$?" set -e - # Generate INFO level output_info.xml for post-processing. - all_options=("--loglevel" "INFO") - all_options+=("--log" "none") - all_options+=("--report" "none") - all_options+=("--output" "${ARCHIVE_DIR}/output_info.xml") - all_options+=("${ARCHIVE_DIR}/output.xml") - rebot "${all_options[@]}" || true popd || die "Change directory operation failed." } @@ -737,9 +816,9 @@ function select_arch_os () { case "${ID}" in "ubuntu"*) case "${VERSION}" in - *"LTS (Focal Fossa)"*) - IMAGE_VER_FILE="VPP_DEVICE_IMAGE_UBUNTU" - VPP_VER_FILE="VPP_STABLE_VER_UBUNTU_FOCAL" + *"LTS (Jammy Jellyfish)"*) + IMAGE_VER_FILE="VPP_DEVICE_IMAGE_UBUNTU_JAMMY" + VPP_VER_FILE="VPP_STABLE_VER_UBUNTU_JAMMY" PKG_SUFFIX="deb" ;; *) @@ -783,38 +862,63 @@ function select_tags () { set -exuo pipefail # NIC SELECTION - start_pattern='^ TG:' + case "${TEST_CODE}" in + *"1n-aws"* | *"1n-c6gn"* | *"1n-c6in"*) + start_pattern='^ SUT:' + ;; + *) + start_pattern='^ TG:' + ;; + esac end_pattern='^ \? \?[A-Za-z0-9]\+:' - # Remove the TG section from topology file + # Remove the sections from topology file sed_command="/${start_pattern}/,/${end_pattern}/d" - # All topologies DUT NICs + # All topologies NICs available=$(sed "${sed_command}" "${TOPOLOGIES_DIR}"/* \ | grep -hoP "model: \K.*" | sort -u) - # Selected topology DUT NICs + # Selected topology NICs reserved=$(sed "${sed_command}" "${WORKING_TOPOLOGY}" \ | grep -hoP "model: \K.*" | sort -u) - # All topologies DUT NICs - Selected topology DUT NICs + # All topologies NICs - Selected topology NICs exclude_nics=($(comm -13 <(echo "${reserved}") <(echo "${available}"))) || { die "Computation of excluded NICs failed." } # Select default NIC tag. case "${TEST_CODE}" in - *"3n-dnv"* | *"2n-dnv"*) - default_nic="nic_intel-x553" + *"3n-snr"*) + default_nic="nic_intel-e822cq" ;; *"3n-tsh"*) default_nic="nic_intel-x520-da2" ;; - *"3n-skx"* | *"2n-skx"* | *"2n-clx"* | *"2n-zn2"*) + *"3n-icx"* | *"2n-icx"*) + default_nic="nic_intel-e810cq" + ;; + *"3na-spr"*) + default_nic="nic_mellanox-cx7veat" + ;; + *"3nb-spr"*) + default_nic="nic_intel-e810cq" + ;; + *"2n-spr"*) + default_nic="nic_intel-e810cq" + ;; + *"2n-clx"* | *"2n-zn2"*) default_nic="nic_intel-xxv710" ;; - *"3n-hsw"* | *"2n-tx2"* | *"mrr-daily-master") + *"2n-tx2"* | *"3n-alt"*) default_nic="nic_intel-xl710" ;; - *"2n-aws"* | *"3n-aws"*) + *"1n-aws"* | *"2n-aws"* | *"3n-aws"*) default_nic="nic_amazon-nitro-50g" ;; + *"1n-c6gn"* | *"2n-c6gn"* | *"3n-c6gn"*) + default_nic="nic_amazon-nitro-100g" + ;; + *"1n-c6in"* | *"2n-c6in"* | *"3n-c6in"*) + default_nic="nic_amazon-nitro-200g" + ;; *) default_nic="nic_intel-x710" ;; @@ -826,23 +930,43 @@ function select_tags () { awk_nics_sub_cmd+='gsub("x710","10ge2p1x710");' awk_nics_sub_cmd+='gsub("xl710","40ge2p1xl710");' awk_nics_sub_cmd+='gsub("x520-da2","10ge2p1x520");' - awk_nics_sub_cmd+='gsub("x553","10ge2p1x553");' awk_nics_sub_cmd+='gsub("cx556a","100ge2p1cx556a");' + awk_nics_sub_cmd+='gsub("cx7veat","200ge2p1cx7veat");' + awk_nics_sub_cmd+='gsub("cx6dx","100ge2p1cx6dx");' + awk_nics_sub_cmd+='gsub("e810cq","100ge2p1e810cq");' awk_nics_sub_cmd+='gsub("vic1227","10ge2p1vic1227");' awk_nics_sub_cmd+='gsub("vic1385","40ge2p1vic1385");' awk_nics_sub_cmd+='gsub("nitro-50g","50ge1p1ENA");' + awk_nics_sub_cmd+='gsub("nitro-100g","100ge1p1ENA");' + awk_nics_sub_cmd+='gsub("nitro-200g","200ge1p1ENA");' awk_nics_sub_cmd+='if ($9 =="drv_avf") drv="avf-";' awk_nics_sub_cmd+='else if ($9 =="drv_rdma_core") drv ="rdma-";' + awk_nics_sub_cmd+='else if ($9 =="drv_mlx5_core") drv ="mlx5-";' + awk_nics_sub_cmd+='else if ($9 =="drv_af_xdp") drv ="af-xdp-";' awk_nics_sub_cmd+='else drv="";' - awk_nics_sub_cmd+='print "*"$7"-" drv $11"-"$5"."$3"-"$1"-" drv $11"-"$5' + awk_nics_sub_cmd+='if ($1 =="-") cores="";' + awk_nics_sub_cmd+='else cores=$1;' + awk_nics_sub_cmd+='print "*"$7"-" drv $11"-"$5"."$3"-" cores "-" drv $11"-"$5' # Tag file directory shorthand. tfd="${JOB_SPECS_DIR}" case "${TEST_CODE}" in # Select specific performance tests based on jenkins job type variable. + *"device"* ) + readarray -t test_tag_array <<< $(grep -v "#" \ + ${tfd}/vpp_device/${DUT}-${NODENESS}-${FLAVOR}.md | + awk {"$awk_nics_sub_cmd"} || echo "devicetest") || die + SELECTION_MODE="--test" + ;; + *"hoststack-daily"* ) + readarray -t test_tag_array <<< $(grep -v "#" \ + ${tfd}/hoststack_daily/${DUT}-${NODENESS}-${FLAVOR}.md | + awk {"$awk_nics_sub_cmd"} || echo "perftest") || die + SELECTION_MODE="--test" + ;; *"ndrpdr-weekly"* ) readarray -t test_tag_array <<< $(grep -v "#" \ - ${tfd}/mlr_weekly/${DUT}-${NODENESS}-${FLAVOR}.md | + ${tfd}/ndrpdr_weekly/${DUT}-${NODENESS}-${FLAVOR}.md | awk {"$awk_nics_sub_cmd"} || echo "perftest") || die SELECTION_MODE="--test" ;; @@ -880,11 +1004,11 @@ function select_tags () { if [[ -z "${TEST_TAG_STRING-}" ]]; then # If nothing is specified, we will run pre-selected tests by # following tags. - test_tag_array=("mrrAND${default_nic}AND1cAND64bANDip4base" - "mrrAND${default_nic}AND1cAND78bANDip6base" - "mrrAND${default_nic}AND1cAND64bANDl2bdbase" - "mrrAND${default_nic}AND1cAND64bANDl2xcbase" - "!dot1q" "!drv_avf") + test_tag_array=("mrrAND${default_nic}AND1cAND64bANDethip4-ip4base" + "mrrAND${default_nic}AND1cAND78bANDethip6-ip6base" + "mrrAND${default_nic}AND1cAND64bANDeth-l2bdbasemaclrn" + "mrrAND${default_nic}AND1cAND64bANDeth-l2xcbase" + "!drv_af_xdp" "!drv_avf") else # If trigger contains tags, split them into array. test_tag_array=(${TEST_TAG_STRING//:/ }) @@ -897,59 +1021,57 @@ function select_tags () { # # Reasons for blacklisting: # - ipsechw - Blacklisted on testbeds without crypto hardware accelerator. - # TODO: Add missing reasons here (if general) or where used (if specific). case "${TEST_CODE}" in - *"2n-skx"*) + *"1n-vbox"*) + test_tag_array+=("!avf") + test_tag_array+=("!vhost") + test_tag_array+=("!flow") + ;; + *"1n_tx2"*) + test_tag_array+=("!flow") + ;; + *"2n-clx"*) test_tag_array+=("!ipsechw") ;; - *"3n-skx"*) + *"2n-icx"*) test_tag_array+=("!ipsechw") - # Not enough nic_intel-xxv710 to support double link tests. - test_tag_array+=("!3_node_double_link_topoANDnic_intel-xxv710") ;; - *"2n-clx"*) + *"2n-spr"*) + test_tag_array+=("!ipsechw") + ;; + *"2n-tx2"*) test_tag_array+=("!ipsechw") ;; *"2n-zn2"*) test_tag_array+=("!ipsechw") ;; - *"2n-dnv"*) + *"3n-alt"*) test_tag_array+=("!ipsechw") - test_tag_array+=("!memif") - test_tag_array+=("!srv6_proxy") - test_tag_array+=("!vhost") - test_tag_array+=("!vts") - test_tag_array+=("!drv_avf") ;; - *"2n-tx2"*) + *"3n-icx"*) test_tag_array+=("!ipsechw") + test_tag_array+=("!3_node_double_link_topoANDnic_intel-xxv710") ;; - *"3n-dnv"*) - test_tag_array+=("!memif") - test_tag_array+=("!srv6_proxy") - test_tag_array+=("!vhost") - test_tag_array+=("!vts") - test_tag_array+=("!drv_avf") + *"3n-snr"*) + ;; + *"3na-spr"*) + test_tag_array+=("!ipsechw") + ;; + *"3nb-spr"*) + test_tag_array+=("!ipsechw") ;; *"3n-tsh"*) - # 3n-tsh only has x520 NICs which don't work with AVF test_tag_array+=("!drv_avf") test_tag_array+=("!ipsechw") ;; - *"3n-hsw"*) - test_tag_array+=("!drv_avf") - # All cards have access to QAT. But only one card (xl710) - # resides in same NUMA as QAT. Other cards must go over QPI - # which we do not want to even run. - test_tag_array+=("!ipsechwNOTnic_intel-xl710") + *"1n-aws"* | *"2n-aws"* | *"3n-aws"*) + test_tag_array+=("!ipsechw") ;; - *"2n-aws"* | *"3n-aws"*) + *"1n-c6gn"* | *"2n-c6gn"* | *"3n-c6gn"*) test_tag_array+=("!ipsechw") ;; - *) - # Default to 3n-hsw due to compatibility. - test_tag_array+=("!drv_avf") - test_tag_array+=("!ipsechwNOTnic_intel-xl710") + *"1n-c6in"* | *"2n-c6in"* | *"3n-c6in"*) + test_tag_array+=("!ipsechw") ;; esac @@ -958,17 +1080,15 @@ function select_tags () { TAGS=() prefix="" - - set +x if [[ "${TEST_CODE}" == "vpp-"* ]]; then - # Automatic prefixing for VPP jobs to limit the NIC used and - # traffic evaluation to MRR. - if [[ "${TEST_TAG_STRING-}" == *"nic_"* ]]; then - prefix="${prefix}mrrAND" - else - prefix="${prefix}mrrAND${default_nic}AND" + if [[ "${TEST_CODE}" != *"device"* ]]; then + # Automatic prefixing for VPP perf jobs to limit the NIC used. + if [[ "${TEST_TAG_STRING-}" != *"nic_"* ]]; then + prefix="${default_nic}AND" + fi fi fi + set +x for tag in "${test_tag_array[@]}"; do if [[ "${tag}" == "!"* ]]; then # Exclude tags are not prefixed. @@ -1000,7 +1120,7 @@ function select_topology () { # Variables read: # - NODENESS - Node multiplicity of testbed, either "2n" or "3n". - # - FLAVOR - Node flavor string, currently either "hsw" or "skx". + # - FLAVOR - Node flavor string, e.g. "clx" or "skx". # - CSIT_DIR - Path to existing root of local CSIT git repository. # - TOPOLOGIES_DIR - Path to existing directory with available topologies. # Variables set: @@ -1013,13 +1133,11 @@ function select_topology () { case_text="${NODENESS}_${FLAVOR}" case "${case_text}" in - # TODO: Move tags to "# Blacklisting certain tags per topology" section. - # TODO: Double link availability depends on NIC used. "1n_vbox") TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*vpp_device*.template ) TOPOLOGIES_TAGS="2_node_single_link_topo" ;; - "1n_skx" | "1n_tx2") + "1n_skx" | "1n_tx2" | "1n_spr") TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*vpp_device*.template ) TOPOLOGIES_TAGS="2_node_single_link_topo" ;; @@ -1035,20 +1153,32 @@ function select_topology () { TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_skx*.yaml ) TOPOLOGIES_TAGS="3_node_*_link_topo" ;; + "3n_icx") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_icx*.yaml ) + TOPOLOGIES_TAGS="3_node_*_link_topo" + ;; + "3na_spr") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3na_spr*.yaml ) + TOPOLOGIES_TAGS="3_node_*_link_topo" + ;; + "3nb_spr") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3nb_spr*.yaml ) + TOPOLOGIES_TAGS="3_node_*_link_topo" + ;; "2n_clx") TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_clx*.yaml ) TOPOLOGIES_TAGS="2_node_*_link_topo" ;; - "2n_dnv") - TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_dnv*.yaml ) - TOPOLOGIES_TAGS="2_node_single_link_topo" + "2n_icx") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_icx*.yaml ) + TOPOLOGIES_TAGS="2_node_*_link_topo" ;; - "3n_dnv") - TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_dnv*.yaml ) - TOPOLOGIES_TAGS="3_node_single_link_topo" + "2n_spr") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_spr*.yaml ) + TOPOLOGIES_TAGS="2_node_*_link_topo" ;; - "3n_hsw") - TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_hsw*.yaml ) + "3n_snr") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_snr*.yaml ) TOPOLOGIES_TAGS="3_node_single_link_topo" ;; "3n_tsh") @@ -1059,16 +1189,48 @@ function select_topology () { TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_tx2*.yaml ) TOPOLOGIES_TAGS="2_node_single_link_topo" ;; + "3n_alt") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_alt*.yaml ) + TOPOLOGIES_TAGS="3_node_single_link_topo" + ;; + "1n_aws") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*1n-aws*.yaml ) + TOPOLOGIES_TAGS="1_node_single_link_topo" + ;; "2n_aws") - TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_aws*.yaml ) + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n-aws*.yaml ) TOPOLOGIES_TAGS="2_node_single_link_topo" ;; "3n_aws") - TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_aws*.yaml ) + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n-aws*.yaml ) + TOPOLOGIES_TAGS="3_node_single_link_topo" + ;; + "1n_c6gn") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*1n-c6gn*.yaml ) + TOPOLOGIES_TAGS="1_node_single_link_topo" + ;; + "2n_c6gn") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n-c6gn*.yaml ) + TOPOLOGIES_TAGS="2_node_single_link_topo" + ;; + "3n_c6gn") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n-c6gn*.yaml ) + TOPOLOGIES_TAGS="3_node_single_link_topo" + ;; + "1n_c6in") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*1n-c6in*.yaml ) + TOPOLOGIES_TAGS="1_node_single_link_topo" + ;; + "2n_c6in") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n-c6in*.yaml ) + TOPOLOGIES_TAGS="2_node_single_link_topo" + ;; + "3n_c6in") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n-c6in*.yaml ) TOPOLOGIES_TAGS="3_node_single_link_topo" ;; *) - # No falling back to 3n_hsw default, that should have been done + # No falling back to default, that should have been done # by the function which has set NODENESS and FLAVOR. die "Unknown specification: ${case_text}" esac @@ -1079,67 +1241,6 @@ function select_topology () { } -function select_vpp_device_tags () { - - # Variables read: - # - TEST_CODE - String affecting test selection, usually jenkins job name. - # - TEST_TAG_STRING - String selecting tags, from gerrit comment. - # Can be unset. - # Variables set: - # - TAGS - Array of processed tag boolean expressions. - - set -exuo pipefail - - case "${TEST_CODE}" in - # Select specific device tests based on jenkins job type variable. - * ) - if [[ -z "${TEST_TAG_STRING-}" ]]; then - # If nothing is specified, we will run pre-selected tests by - # following tags. Items of array will be concatenated by OR - # in Robot Framework. - test_tag_array=() - else - # If trigger contains tags, split them into array. - test_tag_array=(${TEST_TAG_STRING//:/ }) - fi - SELECTION_MODE="--include" - ;; - esac - - # Blacklisting certain tags per topology. - # - # Reasons for blacklisting: - # - avf - AVF is not possible to run on enic driver of VirtualBox. - # - vhost - VirtualBox does not support nesting virtualization on Intel CPU. - case "${TEST_CODE}" in - *"1n-vbox"*) - test_tag_array+=("!avf") - test_tag_array+=("!vhost") - ;; - *) - ;; - esac - - TAGS=() - - # We will prefix with devicetest to prevent running other tests - # (e.g. Functional). - prefix="devicetestAND" - if [[ "${TEST_CODE}" == "vpp-"* ]]; then - # Automatic prefixing for VPP jobs to limit testing. - prefix="${prefix}" - fi - for tag in "${test_tag_array[@]}"; do - if [[ ${tag} == "!"* ]]; then - # Exclude tags are not prefixed. - TAGS+=("${tag}") - else - TAGS+=("${prefix}${tag}") - fi - done -} - - function set_environment_variables () { # Depending on testbed topology, overwrite defaults set in the @@ -1153,13 +1254,30 @@ function set_environment_variables () { set -exuo pipefail case "${TEST_CODE}" in - *"2n-aws"* | *"3n-aws"*) - # T-Rex 2.88 workaround for ENA NICs + *"1n-aws"* | *"2n-aws"* | *"3n-aws"*) export TREX_RX_DESCRIPTORS_COUNT=1024 export TREX_EXTRA_CMDLINE="--mbuf-factor 19" - # Settings to prevent duration stretching + export TREX_CORE_COUNT=6 + # Settings to prevent duration stretching. export PERF_TRIAL_STL_DELAY=0.1 ;; + *"1n-c6gn"* | *"2n-c6gn"* | *"3n-c6gn"*) + export TREX_RX_DESCRIPTORS_COUNT=1024 + export TREX_EXTRA_CMDLINE="--mbuf-factor 19" + export TREX_CORE_COUNT=6 + # Settings to prevent duration stretching. + export PERF_TRIAL_STL_DELAY=0.1 + ;; + *"1n-c6in"* | *"2n-c6in"* | *"3n-c6in"*) + export TREX_RX_DESCRIPTORS_COUNT=1024 + export TREX_EXTRA_CMDLINE="--mbuf-factor 19" + export TREX_CORE_COUNT=6 + # Settings to prevent duration stretching. + export PERF_TRIAL_STL_DELAY=0.1 + ;; + *"2n-zn2"*) + # Maciek's workaround for Zen2 with lower amount of cores. + export TREX_CORE_COUNT=14 esac } @@ -1178,7 +1296,8 @@ function untrap_and_unreserve_testbed () { # Variables read (by inner function): # - WORKING_TOPOLOGY - Path to topology yaml file of the reserved testbed. # - PYTHON_SCRIPTS_DIR - Path to directory holding Python scripts. - # Variables written: + # Variables set: + # - TERRAFORM_MODULE_DIR - Terraform module directory. # - WORKING_TOPOLOGY - Set to empty string on successful unreservation. # Trap unregistered: # - EXIT - Failure to untrap is reported, but ignored otherwise. @@ -1198,6 +1317,22 @@ function untrap_and_unreserve_testbed () { python3 "${PYTHON_SCRIPTS_DIR}/topo_reservation.py" -c -t "${wt}" || { die "${1:-FAILED TO UNRESERVE, FIX MANUALLY.}" 2 } + case "${TEST_CODE}" in + *"1n-aws"* | *"2n-aws"* | *"3n-aws"*) + TERRAFORM_MODULE_DIR="terraform-aws-${NODENESS}-${FLAVOR}-c5n" + terraform_destroy || die "Failed to call terraform destroy." + ;; + *"1n-c6gn"* | *"2n-c6gn"* | *"3n-c6gn"*) + TERRAFORM_MODULE_DIR="terraform-aws-${NODENESS}-${FLAVOR}" + terraform_destroy || die "Failed to call terraform destroy." + ;; + *"1n-c6in"* | *"2n-c6in"* | *"3n-c6in"*) + TERRAFORM_MODULE_DIR="terraform-aws-${NODENESS}-${FLAVOR}" + terraform_destroy || die "Failed to call terraform destroy." + ;; + *) + ;; + esac WORKING_TOPOLOGY="" set -eu fi