X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=resources%2Flibraries%2Fbash%2Ffunction%2Fcommon.sh;h=d768e9b66c86d97a52d6215b75826e17fd15eea1;hb=74aa13757057a4c93d62704ad0de7a71c5cf1357;hp=39b2be4608b4d0b420fa70f6d8c084d3679abbdb;hpb=75351e349449d6c397cb830828dc4f55883511a3;p=csit.git diff --git a/resources/libraries/bash/function/common.sh b/resources/libraries/bash/function/common.sh index 39b2be4608..d768e9b66c 100644 --- a/resources/libraries/bash/function/common.sh +++ b/resources/libraries/bash/function/common.sh @@ -1,5 +1,5 @@ -# Copyright (c) 2022 Cisco and/or its affiliates. -# Copyright (c) 2022 PANTHEON.tech and/or its affiliates. +# Copyright (c) 2023 Cisco and/or its affiliates. +# Copyright (c) 2023 PANTHEON.tech and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: @@ -118,7 +118,7 @@ function activate_virtualenv () { env_dir="${root_path}/env" req_path=${2-$CSIT_DIR/requirements.txt} rm -rf "${env_dir}" || die "Failed to clean previous virtualenv." - pip3 install virtualenv==20.0.20 || { + pip3 install virtualenv==20.15.1 || { die "Virtualenv package install failed." } virtualenv --no-download --python=$(which python3) "${env_dir}" || { @@ -227,7 +227,7 @@ function common_dirs () { TOPOLOGIES_DIR=$(readlink -e "${CSIT_DIR}/topologies/available") || { die "Readlink failed." } - JOB_SPECS_DIR=$(readlink -e "${CSIT_DIR}/docs/job_specs") || { + JOB_SPECS_DIR=$(readlink -e "${CSIT_DIR}/resources/job_specs") || { die "Readlink failed." } RESOURCES_DIR=$(readlink -e "${CSIT_DIR}/resources") || { @@ -236,9 +236,6 @@ function common_dirs () { TOOLS_DIR=$(readlink -e "${RESOURCES_DIR}/tools") || { die "Readlink failed." } - DOC_GEN_DIR=$(readlink -e "${TOOLS_DIR}/doc_gen") || { - die "Readlink failed." - } PYTHON_SCRIPTS_DIR=$(readlink -e "${TOOLS_DIR}/scripts") || { die "Readlink failed." } @@ -439,10 +436,14 @@ function get_test_code () { NODENESS="1n" FLAVOR="skx" ;; - *"1n-tx2"*) + *"1n-tx2"*) NODENESS="1n" FLAVOR="tx2" ;; + *"1n-aws"*) + NODENESS="1n" + FLAVOR="aws" + ;; *"2n-aws"*) NODENESS="2n" FLAVOR="aws" @@ -451,14 +452,6 @@ function get_test_code () { NODENESS="3n" FLAVOR="aws" ;; - *"2n-skx"*) - NODENESS="2n" - FLAVOR="skx" - ;; - *"3n-skx"*) - NODENESS="3n" - FLAVOR="skx" - ;; *"2n-zn2"*) NODENESS="2n" FLAVOR="zn2" @@ -471,17 +464,25 @@ function get_test_code () { NODENESS="2n" FLAVOR="icx" ;; + *"2n-spr"*) + NODENESS="2n" + FLAVOR="spr" + ;; *"3n-icx"*) NODENESS="3n" FLAVOR="icx" ;; - *"2n-dnv"*) - NODENESS="2n" - FLAVOR="dnv" + *"3na-spr"*) + NODENESS="3na" + FLAVOR="spr" ;; - *"3n-dnv"*) + *"3nb-spr"*) + NODENESS="3nb" + FLAVOR="spr" + ;; + *"3n-snr"*) NODENESS="3n" - FLAVOR="dnv" + FLAVOR="snr" ;; *"2n-tx2"*) NODENESS="2n" @@ -491,6 +492,10 @@ function get_test_code () { NODENESS="3n" FLAVOR="tsh" ;; + *"3n-alt"*) + NODENESS="3n" + FLAVOR="alt" + ;; esac } @@ -612,29 +617,12 @@ function move_archives () { function post_process_robot_outputs () { # Generate INFO level output_info.xml by rebot. - # Archive UTI raw json outputs. # # Variables read: # - ARCHIVE_DIR - Path to post-processed files. set -exuo pipefail - # Compress raw json outputs, as they will never be post-processed. - pushd "${ARCHIVE_DIR}" || die - if [ -d "tests" ]; then - # Use deterministic order. - options+=("--sort=name") - # We are keeping info outputs where they are. - # Assuming we want to move anything but info files (and dirs). - options+=("--exclude=*.info.json") - tar czf "generated_output_raw.tar.gz" "${options[@]}" "tests" || true - # Tar can remove when archiving, but chokes (not deterministically) - # on attempting to remove dirs (not empty as info files are there). - # So we need to delete the raw files manually. - find "tests" -type f -name "*.raw.json" -delete || true - fi - popd || die - # Generate INFO level output_info.xml for post-processing. all_options=("--loglevel" "INFO") all_options+=("--log" "none") @@ -653,6 +641,8 @@ function prepare_topology () { # - TEST_CODE - String affecting test selection, usually jenkins job name. # - NODENESS - Node multiplicity of testbed, either "2n" or "3n". # - FLAVOR - Node flavor string, e.g. "clx" or "skx". + # Variables set: + # - TERRAFORM_MODULE_DIR - Terraform module directory. # Functions called: # - die - Print to stderr and exit. # - terraform_init - Terraform init topology. @@ -662,14 +652,13 @@ function prepare_topology () { case_text="${NODENESS}_${FLAVOR}" case "${case_text}" in - "2n_aws") - export TF_VAR_testbed_name="${TEST_CODE}" - terraform_init || die "Failed to call terraform init." - terraform_apply || die "Failed to call terraform apply." - ;; - "3n_aws") + "1n_aws" | "2n_aws" | "3n_aws") export TF_VAR_testbed_name="${TEST_CODE}" + TERRAFORM_MODULE_DIR="terraform-aws-${NODENESS}-${FLAVOR}-c5n" terraform_init || die "Failed to call terraform init." + trap 'terraform_destroy' ERR || { + die "Trap attempt failed, please cleanup manually. Aborting!" + } terraform_apply || die "Failed to call terraform apply." ;; esac @@ -757,7 +746,6 @@ function run_pybot () { # Run pybot with options based on input variables. # Generate INFO level output_info.xml by rebot. - # Archive UTI raw json outputs. # # Variables read: # - CSIT_DIR - Path to existing root of local CSIT git repository. @@ -772,7 +760,6 @@ function run_pybot () { set -exuo pipefail all_options=("--outputdir" "${ARCHIVE_DIR}" "${PYBOT_ARGS[@]}") - all_options+=("--noncritical" "EXPECTED_FAILING") all_options+=("${EXPANDED_TAGS[@]}") pushd "${CSIT_DIR}" || die "Change directory operation failed." @@ -803,9 +790,9 @@ function select_arch_os () { case "${ID}" in "ubuntu"*) case "${VERSION}" in - *"LTS (Focal Fossa)"*) - IMAGE_VER_FILE="VPP_DEVICE_IMAGE_UBUNTU" - VPP_VER_FILE="VPP_STABLE_VER_UBUNTU_FOCAL" + *"LTS (Jammy Jellyfish)"*) + IMAGE_VER_FILE="VPP_DEVICE_IMAGE_UBUNTU_JAMMY" + VPP_VER_FILE="VPP_STABLE_VER_UBUNTU_JAMMY" PKG_SUFFIX="deb" ;; *) @@ -849,25 +836,32 @@ function select_tags () { set -exuo pipefail # NIC SELECTION - start_pattern='^ TG:' + case "${TEST_CODE}" in + *"1n-aws"*) + start_pattern='^ SUT:' + ;; + *) + start_pattern='^ TG:' + ;; + esac end_pattern='^ \? \?[A-Za-z0-9]\+:' - # Remove the TG section from topology file + # Remove the sections from topology file sed_command="/${start_pattern}/,/${end_pattern}/d" - # All topologies DUT NICs + # All topologies NICs available=$(sed "${sed_command}" "${TOPOLOGIES_DIR}"/* \ | grep -hoP "model: \K.*" | sort -u) - # Selected topology DUT NICs + # Selected topology NICs reserved=$(sed "${sed_command}" "${WORKING_TOPOLOGY}" \ | grep -hoP "model: \K.*" | sort -u) - # All topologies DUT NICs - Selected topology DUT NICs + # All topologies NICs - Selected topology NICs exclude_nics=($(comm -13 <(echo "${reserved}") <(echo "${available}"))) || { die "Computation of excluded NICs failed." } # Select default NIC tag. case "${TEST_CODE}" in - *"3n-dnv"* | *"2n-dnv"*) - default_nic="nic_intel-x553" + *"3n-snr"*) + default_nic="nic_intel-e822cq" ;; *"3n-tsh"*) default_nic="nic_intel-x520-da2" @@ -875,13 +869,22 @@ function select_tags () { *"3n-icx"* | *"2n-icx"*) default_nic="nic_intel-xxv710" ;; - *"3n-skx"* | *"2n-skx"* | *"2n-clx"* | *"2n-zn2"*) + *"3na-spr"*) + default_nic="nic_mellanox-cx7veat" + ;; + *"3nb-spr"*) + default_nic="nic_intel-e810cq" + ;; + *"2n-spr"*) + default_nic="nic_intel-e810cq" + ;; + *"2n-clx"* | *"2n-zn2"*) default_nic="nic_intel-xxv710" ;; - *"2n-tx2"* | *"mrr-daily-master") + *"2n-tx2"* | *"3n-alt"*) default_nic="nic_intel-xl710" ;; - *"2n-aws"* | *"3n-aws"*) + *"1n-aws"* | *"2n-aws"* | *"3n-aws"*) default_nic="nic_amazon-nitro-50g" ;; *) @@ -895,14 +898,15 @@ function select_tags () { awk_nics_sub_cmd+='gsub("x710","10ge2p1x710");' awk_nics_sub_cmd+='gsub("xl710","40ge2p1xl710");' awk_nics_sub_cmd+='gsub("x520-da2","10ge2p1x520");' - awk_nics_sub_cmd+='gsub("x553","10ge2p1x553");' awk_nics_sub_cmd+='gsub("cx556a","100ge2p1cx556a");' + awk_nics_sub_cmd+='gsub("cx7veat","200ge2p1cx7veat");' awk_nics_sub_cmd+='gsub("e810cq","100ge2p1e810cq");' awk_nics_sub_cmd+='gsub("vic1227","10ge2p1vic1227");' awk_nics_sub_cmd+='gsub("vic1385","40ge2p1vic1385");' awk_nics_sub_cmd+='gsub("nitro-50g","50ge1p1ENA");' awk_nics_sub_cmd+='if ($9 =="drv_avf") drv="avf-";' awk_nics_sub_cmd+='else if ($9 =="drv_rdma_core") drv ="rdma-";' + awk_nics_sub_cmd+='else if ($9 =="drv_mlx5_core") drv ="mlx5-";' awk_nics_sub_cmd+='else if ($9 =="drv_af_xdp") drv ="af-xdp-";' awk_nics_sub_cmd+='else drv="";' awk_nics_sub_cmd+='if ($1 =="-") cores="";' @@ -919,9 +923,15 @@ function select_tags () { awk {"$awk_nics_sub_cmd"} || echo "devicetest") || die SELECTION_MODE="--test" ;; + *"hoststack-daily"* ) + readarray -t test_tag_array <<< $(grep -v "#" \ + ${tfd}/hoststack_daily/${DUT}-${NODENESS}-${FLAVOR}.md | + awk {"$awk_nics_sub_cmd"} || echo "perftest") || die + SELECTION_MODE="--test" + ;; *"ndrpdr-weekly"* ) readarray -t test_tag_array <<< $(grep -v "#" \ - ${tfd}/mlr_weekly/${DUT}-${NODENESS}-${FLAVOR}.md | + ${tfd}/ndrpdr_weekly/${DUT}-${NODENESS}-${FLAVOR}.md | awk {"$awk_nics_sub_cmd"} || echo "perftest") || die SELECTION_MODE="--test" ;; @@ -985,51 +995,41 @@ function select_tags () { *"1n_tx2"*) test_tag_array+=("!flow") ;; - *"2n-skx"*) + *"2n-clx"*) test_tag_array+=("!ipsechw") ;; - *"3n-skx"*) + *"2n-icx"*) test_tag_array+=("!ipsechw") - # Not enough nic_intel-xxv710 to support double link tests. - test_tag_array+=("!3_node_double_link_topoANDnic_intel-xxv710") ;; - *"2n-clx"*) + *"2n-spr"*) test_tag_array+=("!ipsechw") ;; - *"2n-icx"*) + *"2n-tx2"*) test_tag_array+=("!ipsechw") ;; - *"3n-icx"*) + *"2n-zn2"*) test_tag_array+=("!ipsechw") - # Not enough nic_intel-xxv710 to support double link tests. - test_tag_array+=("!3_node_double_link_topoANDnic_intel-xxv710") ;; - *"2n-zn2"*) + *"3n-alt"*) test_tag_array+=("!ipsechw") ;; - *"2n-dnv"*) - test_tag_array+=("!memif") - test_tag_array+=("!srv6_proxy") - test_tag_array+=("!vhost") - test_tag_array+=("!vts") - test_tag_array+=("!drv_avf") + *"3n-icx"*) + test_tag_array+=("!ipsechw") + test_tag_array+=("!3_node_double_link_topoANDnic_intel-xxv710") ;; - *"2n-tx2"*) + *"3n-snr"*) + ;; + *"3na-spr"*) test_tag_array+=("!ipsechw") ;; - *"3n-dnv"*) - test_tag_array+=("!memif") - test_tag_array+=("!srv6_proxy") - test_tag_array+=("!vhost") - test_tag_array+=("!vts") - test_tag_array+=("!drv_avf") + *"3nb-spr"*) + test_tag_array+=("!ipsechw") ;; *"3n-tsh"*) - # 3n-tsh only has x520 NICs which don't work with AVF test_tag_array+=("!drv_avf") test_tag_array+=("!ipsechw") ;; - *"2n-aws"* | *"3n-aws"*) + *"1n-aws"* | *"2n-aws"* | *"3n-aws"*) test_tag_array+=("!ipsechw") ;; esac @@ -1120,6 +1120,14 @@ function select_topology () { TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_icx*.yaml ) TOPOLOGIES_TAGS="3_node_*_link_topo" ;; + "3na_spr") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3na_spr*.yaml ) + TOPOLOGIES_TAGS="3_node_*_link_topo" + ;; + "3nb_spr") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3nb_spr*.yaml ) + TOPOLOGIES_TAGS="3_node_*_link_topo" + ;; "2n_clx") TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_clx*.yaml ) TOPOLOGIES_TAGS="2_node_*_link_topo" @@ -1128,12 +1136,12 @@ function select_topology () { TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_icx*.yaml ) TOPOLOGIES_TAGS="2_node_*_link_topo" ;; - "2n_dnv") - TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_dnv*.yaml ) - TOPOLOGIES_TAGS="2_node_single_link_topo" + "2n_spr") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_spr*.yaml ) + TOPOLOGIES_TAGS="2_node_*_link_topo" ;; - "3n_dnv") - TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_dnv*.yaml ) + "3n_snr") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_snr*.yaml ) TOPOLOGIES_TAGS="3_node_single_link_topo" ;; "3n_tsh") @@ -1144,12 +1152,20 @@ function select_topology () { TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_tx2*.yaml ) TOPOLOGIES_TAGS="2_node_single_link_topo" ;; + "3n_alt") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_alt*.yaml ) + TOPOLOGIES_TAGS="3_node_single_link_topo" + ;; + "1n_aws") + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*1n-aws*.yaml ) + TOPOLOGIES_TAGS="1_node_single_link_topo" + ;; "2n_aws") - TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_aws*.yaml ) + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n-aws*.yaml ) TOPOLOGIES_TAGS="2_node_single_link_topo" ;; "3n_aws") - TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_aws*.yaml ) + TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n-aws*.yaml ) TOPOLOGIES_TAGS="3_node_single_link_topo" ;; *) @@ -1177,14 +1193,17 @@ function set_environment_variables () { set -exuo pipefail case "${TEST_CODE}" in - *"2n-aws"* | *"3n-aws"*) - # T-Rex 2.88 workaround for ENA NICs + *"1n-aws"* | *"2n-aws"* | *"3n-aws"*) + # T-Rex 2.88+ workaround for ENA NICs. export TREX_RX_DESCRIPTORS_COUNT=1024 export TREX_EXTRA_CMDLINE="--mbuf-factor 19" export TREX_CORE_COUNT=6 - # Settings to prevent duration stretching + # Settings to prevent duration stretching. export PERF_TRIAL_STL_DELAY=0.1 ;; + *"2n-zn2"*) + # Maciek's workaround for Zen2 with lower amount of cores. + export TREX_CORE_COUNT=14 esac } @@ -1203,7 +1222,8 @@ function untrap_and_unreserve_testbed () { # Variables read (by inner function): # - WORKING_TOPOLOGY - Path to topology yaml file of the reserved testbed. # - PYTHON_SCRIPTS_DIR - Path to directory holding Python scripts. - # Variables written: + # Variables set: + # - TERRAFORM_MODULE_DIR - Terraform module directory. # - WORKING_TOPOLOGY - Set to empty string on successful unreservation. # Trap unregistered: # - EXIT - Failure to untrap is reported, but ignored otherwise. @@ -1224,7 +1244,8 @@ function untrap_and_unreserve_testbed () { die "${1:-FAILED TO UNRESERVE, FIX MANUALLY.}" 2 } case "${TEST_CODE}" in - *"2n-aws"* | *"3n-aws"*) + *"1n-aws"* | *"2n-aws"* | *"3n-aws"*) + TERRAFORM_MODULE_DIR="terraform-aws-${NODENESS}-${FLAVOR}-c5n" terraform_destroy || die "Failed to call terraform destroy." ;; *)