-# Copyright (c) 2021 Cisco and/or its affiliates.
-# Copyright (c) 2021 PANTHEON.tech and/or its affiliates.
+# Copyright (c) 2022 Cisco and/or its affiliates.
+# Copyright (c) 2022 PANTHEON.tech and/or its affiliates.
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at:
function archive_tests () {
- # Create .tar.xz of generated/tests for archiving.
+ # Create .tar.gz of generated/tests for archiving.
# To be run after generate_tests, kept separate to offer more flexibility.
# Directory read:
# - ${GENERATED_DIR}/tests - Tree of executed suites to archive.
# File rewriten:
- # - ${ARCHIVE_DIR}/tests.tar.xz - Archive of generated tests.
+ # - ${ARCHIVE_DIR}/generated_tests.tar.gz - Archive of generated tests.
set -exuo pipefail
- filename="generated_robot_files.tar.xz"
- tar c "${GENERATED_DIR}/tests" | xz -3 > "${ARCHIVE_DIR}/${filename}" || {
- die "Error creating archive of generated tests."
- }
+ pushd "${ARCHIVE_DIR}" || die
+ tar czf "generated_tests.tar.gz" "${GENERATED_DIR}/tests" || true
+ popd || die
}
NODENESS="1n"
FLAVOR="skx"
;;
- *"1n-tx2"*)
+ *"1n-tx2"*)
NODENESS="1n"
FLAVOR="tx2"
;;
+ *"1n-aws"*)
+ NODENESS="1n"
+ FLAVOR="aws"
+ ;;
*"2n-aws"*)
NODENESS="2n"
FLAVOR="aws"
NODENESS="2n"
FLAVOR="skx"
;;
- *"2n-zn2"*)
- NODENESS="2n"
- FLAVOR="zn2"
- ;;
*"3n-skx"*)
NODENESS="3n"
FLAVOR="skx"
;;
+ *"2n-zn2"*)
+ NODENESS="2n"
+ FLAVOR="zn2"
+ ;;
*"2n-clx"*)
NODENESS="2n"
FLAVOR="clx"
;;
+ *"2n-icx"*)
+ NODENESS="2n"
+ FLAVOR="icx"
+ ;;
+ *"3n-icx"*)
+ NODENESS="3n"
+ FLAVOR="icx"
+ ;;
*"2n-dnv"*)
NODENESS="2n"
FLAVOR="dnv"
NODENESS="3n"
FLAVOR="dnv"
;;
+ *"3n-snr"*)
+ NODENESS="3n"
+ FLAVOR="snr"
+ ;;
*"2n-tx2"*)
NODENESS="2n"
FLAVOR="tx2"
NODENESS="3n"
FLAVOR="tsh"
;;
+ *"3n-alt"*)
+ NODENESS="3n"
+ FLAVOR="alt"
+ ;;
esac
}
}
+function post_process_robot_outputs () {
+
+ # Generate INFO level output_info.xml by rebot.
+ # Archive UTI raw json outputs.
+ #
+ # Variables read:
+ # - ARCHIVE_DIR - Path to post-processed files.
+
+ set -exuo pipefail
+
+ # Compress raw json outputs, as they will never be post-processed.
+ pushd "${ARCHIVE_DIR}" || die
+ if [ -d "tests" ]; then
+ # Use deterministic order.
+ options+=("--sort=name")
+ # We are keeping info outputs where they are.
+ # Assuming we want to move anything but info files (and dirs).
+ options+=("--exclude=*.info.json")
+ tar czf "generated_output_raw.tar.gz" "${options[@]}" "tests" || true
+ # Tar can remove when archiving, but chokes (not deterministically)
+ # on attempting to remove dirs (not empty as info files are there).
+ # So we need to delete the raw files manually.
+ find "tests" -type f -name "*.raw.json" -delete || true
+ fi
+ popd || die
+
+ # Generate INFO level output_info.xml for post-processing.
+ all_options=("--loglevel" "INFO")
+ all_options+=("--log" "none")
+ all_options+=("--report" "none")
+ all_options+=("--output" "${ARCHIVE_DIR}/output_info.xml")
+ all_options+=("${ARCHIVE_DIR}/output.xml")
+ rebot "${all_options[@]}" || true
+}
+
+
function prepare_topology () {
# Prepare virtual testbed topology if needed based on flavor.
case_text="${NODENESS}_${FLAVOR}"
case "${case_text}" in
- "2n_aws")
- export TF_VAR_testbed_name="${TEST_CODE}"
- terraform_init || die "Failed to call terraform init."
- terraform_apply || die "Failed to call terraform apply."
- ;;
- "3n_aws")
+ "1n_aws" | "2n_aws" | "3n_aws")
export TF_VAR_testbed_name="${TEST_CODE}"
terraform_init || die "Failed to call terraform init."
terraform_apply || die "Failed to call terraform apply."
function run_pybot () {
- # Run pybot with options based on input variables. Create output_info.xml
- #
- # Also, .info.json files are moved into an archive to speed up PAL.
+ # Run pybot with options based on input variables.
+ # Generate INFO level output_info.xml by rebot.
+ # Archive UTI raw json outputs.
#
# Variables read:
# - CSIT_DIR - Path to existing root of local CSIT git repository.
PYBOT_EXIT_STATUS="$?"
set -e
- # Compress raw json outputs, if any.
- pushd "${ARCHIVE_DIR}" || die
- if [ -d "tests" ]; then
- # Use deterministic order.
- options+=("--sort=name")
- # We are keeping info outputs where they are.
- # Assuming we want to move anything but info files (and dirs).
- options+=("--exclude=*.info.json")
- # There may be other unforeseen errors,
- # we still want to execute subsequent commands, so disable set -e.
- set +e
- tar cvf "tests_output_raw.tar" "${options[@]}" "tests"
- # If compression fails, it leaves an uncompressed .tar,
- # we still want to archive that to investigate why compression failed.
- time xz -9e "tests_output_raw.tar"
- # Tar can remove when archiving, but chokes (not deterministically)
- # on attempting to remove dirs (not empty as info files are there).
- # So we need to delete the raw files manually.
- find "tests" -type f -name "*.raw.json" -delete
- set -e
- fi
- popd || die
+ post_process_robot_outputs || die
- # Generate INFO level output_info.xml for post-processing.
- # This comes last, as it is slowest, and sometimes users abort here.
- all_options=("--loglevel" "INFO")
- all_options+=("--log" "none")
- all_options+=("--report" "none")
- all_options+=("--output" "${ARCHIVE_DIR}/output_info.xml")
- all_options+=("${ARCHIVE_DIR}/output.xml")
- rebot "${all_options[@]}" || true
popd || die "Change directory operation failed."
}
set -exuo pipefail
# NIC SELECTION
- start_pattern='^ TG:'
+ case "${TEST_CODE}" in
+ *"1n-aws"*)
+ start_pattern='^ SUT:'
+ ;;
+ *)
+ start_pattern='^ TG:'
+ ;;
+ esac
end_pattern='^ \? \?[A-Za-z0-9]\+:'
- # Remove the TG section from topology file
+ # Remove the sections from topology file
sed_command="/${start_pattern}/,/${end_pattern}/d"
- # All topologies DUT NICs
+ # All topologies NICs
available=$(sed "${sed_command}" "${TOPOLOGIES_DIR}"/* \
| grep -hoP "model: \K.*" | sort -u)
- # Selected topology DUT NICs
+ # Selected topology NICs
reserved=$(sed "${sed_command}" "${WORKING_TOPOLOGY}" \
| grep -hoP "model: \K.*" | sort -u)
- # All topologies DUT NICs - Selected topology DUT NICs
+ # All topologies NICs - Selected topology NICs
exclude_nics=($(comm -13 <(echo "${reserved}") <(echo "${available}"))) || {
die "Computation of excluded NICs failed."
}
*"3n-dnv"* | *"2n-dnv"*)
default_nic="nic_intel-x553"
;;
+ *"3n-snr"*)
+ default_nic="nic_intel-e810xxv"
+ ;;
*"3n-tsh"*)
default_nic="nic_intel-x520-da2"
;;
+ *"3n-icx"* | *"2n-icx"*)
+ default_nic="nic_intel-xxv710"
+ ;;
*"3n-skx"* | *"2n-skx"* | *"2n-clx"* | *"2n-zn2"*)
default_nic="nic_intel-xxv710"
;;
- *"2n-tx2"* | *"mrr-daily-master")
+ *"2n-tx2"* | *"3n-alt"* | *"mrr-daily-master")
default_nic="nic_intel-xl710"
;;
- *"2n-aws"* | *"3n-aws"*)
+ *"1n-aws"* | *"2n-aws"* | *"3n-aws"*)
default_nic="nic_amazon-nitro-50g"
;;
*)
*"2n-clx"*)
test_tag_array+=("!ipsechw")
;;
+ *"2n-icx"*)
+ test_tag_array+=("!ipsechw")
+ ;;
+ *"3n-icx"*)
+ test_tag_array+=("!ipsechw")
+ # Not enough nic_intel-xxv710 to support double link tests.
+ test_tag_array+=("!3_node_double_link_topoANDnic_intel-xxv710")
+ ;;
*"2n-zn2"*)
test_tag_array+=("!ipsechw")
;;
test_tag_array+=("!vts")
test_tag_array+=("!drv_avf")
;;
- *"2n-tx2"*)
+ *"2n-tx2"* | *"3n-alt"*)
test_tag_array+=("!ipsechw")
;;
*"3n-dnv"*)
test_tag_array+=("!vts")
test_tag_array+=("!drv_avf")
;;
+ *"3n-snr"*)
+ ;;
*"3n-tsh"*)
# 3n-tsh only has x520 NICs which don't work with AVF
test_tag_array+=("!drv_avf")
test_tag_array+=("!ipsechw")
;;
- *"2n-aws"* | *"3n-aws"*)
+ *"1n-aws"* | *"2n-aws"* | *"3n-aws"*)
test_tag_array+=("!ipsechw")
;;
esac
TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_skx*.yaml )
TOPOLOGIES_TAGS="3_node_*_link_topo"
;;
+ "3n_icx")
+ TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_icx*.yaml )
+ TOPOLOGIES_TAGS="3_node_*_link_topo"
+ ;;
"2n_clx")
TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_clx*.yaml )
TOPOLOGIES_TAGS="2_node_*_link_topo"
;;
+ "2n_icx")
+ TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_icx*.yaml )
+ TOPOLOGIES_TAGS="2_node_*_link_topo"
+ ;;
"2n_dnv")
TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_dnv*.yaml )
TOPOLOGIES_TAGS="2_node_single_link_topo"
TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_tx2*.yaml )
TOPOLOGIES_TAGS="2_node_single_link_topo"
;;
+ "3n_alt")
+ TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_alt*.yaml )
+ TOPOLOGIES_TAGS="3_node_single_link_topo"
+ ;;
+ "1n_aws")
+ TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*1n-aws*.yaml )
+ TOPOLOGIES_TAGS="1_node_single_link_topo"
+ ;;
"2n_aws")
- TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n_aws*.yaml )
+ TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*2n-aws*.yaml )
TOPOLOGIES_TAGS="2_node_single_link_topo"
;;
"3n_aws")
- TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n_aws*.yaml )
+ TOPOLOGIES=( "${TOPOLOGIES_DIR}"/*3n-aws*.yaml )
TOPOLOGIES_TAGS="3_node_single_link_topo"
;;
*)
set -exuo pipefail
case "${TEST_CODE}" in
- *"2n-aws"* | *"3n-aws"*)
- # T-Rex 2.88 workaround for ENA NICs
+ *"1n-aws"* | *"2n-aws"* | *"3n-aws"*)
+ # T-Rex 2.88+ workaround for ENA NICs.
export TREX_RX_DESCRIPTORS_COUNT=1024
export TREX_EXTRA_CMDLINE="--mbuf-factor 19"
export TREX_CORE_COUNT=6
- # Settings to prevent duration stretching
+ # Settings to prevent duration stretching.
export PERF_TRIAL_STL_DELAY=0.1
;;
+ *"2n-zn2"*)
+ # Maciek's workaround for Zen2 with lower amount of cores.
+ export TREX_CORE_COUNT=14
esac
}
die "${1:-FAILED TO UNRESERVE, FIX MANUALLY.}" 2
}
case "${TEST_CODE}" in
- *"2n-aws"* | *"3n-aws"*)
+ *"1n-aws"* | *"2n-aws"* | *"3n-aws"*)
terraform_destroy || die "Failed to call terraform destroy."
;;
*)