X-Git-Url: https://gerrit.fd.io/r/gitweb?p=csit.git;a=blobdiff_plain;f=bootstrap-verify-perf.sh;h=9e1d43ac70a7f8163782548d3f74df0e327633ee;hp=de478e5f11af3f45b9feb332a2bbd2fd3cb6b524;hb=faadf83599b8640c9235c38a4ab57c7adfb9eb96;hpb=824fb1d6a1d94636d7a73ad6b1fc69fabf8efad7 diff --git a/bootstrap-verify-perf.sh b/bootstrap-verify-perf.sh index de478e5f11..9e1d43ac70 100755 --- a/bootstrap-verify-perf.sh +++ b/bootstrap-verify-perf.sh @@ -1,5 +1,5 @@ #!/bin/bash -# Copyright (c) 2016 Cisco and/or its affiliates. +# Copyright (c) 2018 Cisco and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: @@ -12,64 +12,113 @@ # See the License for the specific language governing permissions and # limitations under the License. -set -x +set -xo pipefail +# TOPOLOGY # Space separated list of available testbeds, described by topology files -TOPOLOGIES="topologies/available/lf_testbed2-710-520.yaml" -VPP_STABLE_VER="1.0.0-437~g8f15e92_amd64" -VPP_REPO_URL="https://nexus.fd.io/service/local/repositories/fd.io.dev/content/io/fd/vpp" - -# Reservation dir +TOPOLOGIES_3N_HSW="topologies/available/lf_3n_hsw_testbed1.yaml \ + topologies/available/lf_3n_hsw_testbed2.yaml \ + topologies/available/lf_3n_hsw_testbed3.yaml" +TOPOLOGIES_2N_SKX="topologies/available/lf_2n_skx_testbed21.yaml \ + topologies/available/lf_2n_skx_testbed24.yaml" +TOPOLOGIES_3N_SKX="topologies/available/lf_3n_skx_testbed31.yaml \ + topologies/available/lf_3n_skx_testbed32.yaml" + +# SYSTEM +SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )" +export PYTHONPATH=${SCRIPT_DIR} +export DEBIAN_FRONTEND=noninteractive + +# RESERVATION RESERVATION_DIR="/tmp/reservation_dir" INSTALLATION_DIR="/tmp/install_dir" -PYBOT_ARGS="--noncritical MULTI_THREAD" - -# If we run this script from CSIT jobs we want to use stable vpp version -if [[ ${JOB_NAME} == csit-* ]] ; -then - mkdir vpp_download - cd vpp_download - #download vpp build from nexus and set VPP_DEBS variable - wget -q "${VPP_REPO_URL}/vpp/${VPP_STABLE_VER}/vpp-${VPP_STABLE_VER}.deb" || exit - wget -q "${VPP_REPO_URL}/vpp-dbg/${VPP_STABLE_VER}/vpp-dbg-${VPP_STABLE_VER}.deb" || exit - wget -q "${VPP_REPO_URL}/vpp-dev/${VPP_STABLE_VER}/vpp-dev-${VPP_STABLE_VER}.deb" || exit - wget -q "${VPP_REPO_URL}/vpp-dpdk-dev/${VPP_STABLE_VER}/vpp-dpdk-dev-${VPP_STABLE_VER}.deb" || exit - wget -q "${VPP_REPO_URL}/vpp-dpdk-dkms/${VPP_STABLE_VER}/vpp-dpdk-dkms-${VPP_STABLE_VER}.deb" || exit - wget -q "${VPP_REPO_URL}/vpp-lib/${VPP_STABLE_VER}/vpp-lib-${VPP_STABLE_VER}.deb" || exit - VPP_DEBS="$( readlink -f *.deb | tr '\n' ' ' )" - PYBOT_ARGS="${PYBOT_ARGS} --exitonfailure" - cd .. - -# If we run this script from vpp project we want to use local build -elif [[ ${JOB_NAME} == vpp-* ]] ; -then - #use local packages provided as argument list - # Jenkins VPP deb paths (convert to full path) - VPP_DEBS="$( readlink -f $@ | tr '\n' ' ' )" -else - echo "Unable to identify job type based on JOB_NAME variable: ${JOB_NAME}" - exit 1 -fi - -CUR_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )" -WORKING_TOPOLOGY="" -export PYTHONPATH=${CUR_DIR} - -sudo apt-get -y update -sudo apt-get -y install libpython2.7-dev python-virtualenv +# ARCHIVE +JOB_ARCHIVE_ARTIFACTS=(log.html output.xml report.html) +LOG_ARCHIVE_ARTIFACTS=(log.html output.xml report.html) +JOB_ARCHIVE_DIR="archive" +LOG_ARCHIVE_DIR="$WORKSPACE/archives" +mkdir -p ${JOB_ARCHIVE_DIR} +mkdir -p ${LOG_ARCHIVE_DIR} + +# JOB SETTING +case ${JOB_NAME} in + *2n-skx*) + TOPOLOGIES=$TOPOLOGIES_2N_SKX + TOPOLOGIES_TAGS="2_node_*_link_topo" + ;; + *3n-skx*) + TOPOLOGIES=$TOPOLOGIES_3N_SKX + TOPOLOGIES_TAGS="3_node_*_link_topo" + ;; + *) + TOPOLOGIES=$TOPOLOGIES_3N_HSW + TOPOLOGIES_TAGS="3_node_*_link_topo" + ;; +esac +case ${JOB_NAME} in + *hc2vpp*) + DUT="hc2vpp" + ;; + *vpp*) + DUT="vpp" + + case ${JOB_NAME} in + csit-vpp-*) + # Use downloaded packages with specific version + if [[ ${TEST_TAG} == *DAILY ]] || \ + [[ ${TEST_TAG} == *WEEKLY ]]; + then + echo Downloading latest VPP packages from NEXUS... + bash ${SCRIPT_DIR}/resources/tools/scripts/download_install_vpp_pkgs.sh \ + --skip-install + else + echo Downloading VPP packages of specific version from NEXUS... + DPDK_STABLE_VER=$(cat ${SCRIPT_DIR}/DPDK_STABLE_VER) + VPP_STABLE_VER=$(cat ${SCRIPT_DIR}/VPP_STABLE_VER_UBUNTU) + bash ${SCRIPT_DIR}/resources/tools/scripts/download_install_vpp_pkgs.sh \ + --skip-install --vpp ${VPP_STABLE_VER} --dkms ${DPDK_STABLE_VER} + fi + # Jenkins VPP deb paths (convert to full path) + DUT_PKGS="$( readlink -f ${DUT}*.deb | tr '\n' ' ' )" + ;; + vpp-csit-*) + # Use local packages provided as argument list + # Jenkins VPP deb paths (convert to full path) + DUT_PKGS="$( readlink -f $@ | tr '\n' ' ' )" + ;; + *) + echo "Unable to identify job type based on JOB_NAME variable: ${JOB_NAME}" + exit 1 + ;; + esac + ;; + *ligato*) + DUT="kubernetes" + ;; + *dpdk*) + DUT="dpdk" + ;; + *) + echo "Unable to identify dut type based on JOB_NAME variable: ${JOB_NAME}" + exit 1 + ;; +esac -virtualenv env +# ENVIRONMENT PREPARATION +virtualenv --system-site-packages env . env/bin/activate - -echo pip install pip install -r requirements.txt +if [ -z "${TOPOLOGIES}" ]; then + echo "No applicable topology found!" + exit 1 +fi # We iterate over available topologies and wait until we reserve topology while :; do for TOPOLOGY in ${TOPOLOGIES}; do - python ${CUR_DIR}/resources/tools/topo_reservation.py -t ${TOPOLOGY} + python ${SCRIPT_DIR}/resources/tools/scripts/topo_reservation.py -t ${TOPOLOGY} if [ $? -eq 0 ]; then WORKING_TOPOLOGY=${TOPOLOGY} echo "Reserved: ${WORKING_TOPOLOGY}" @@ -89,74 +138,116 @@ while :; do done function cancel_all { - python ${CUR_DIR}/resources/tools/topo_installation.py -c -d ${INSTALLATION_DIR} -t $1 - python ${CUR_DIR}/resources/tools/topo_reservation.py -c -t $1 + python ${SCRIPT_DIR}/resources/tools/scripts/topo_installation.py -c -d ${INSTALLATION_DIR} -t $1 + python ${SCRIPT_DIR}/resources/tools/scripts/topo_reservation.py -c -t $1 } # On script exit we cancel the reservation and installation and delete all vpp # packages trap "cancel_all ${WORKING_TOPOLOGY}" EXIT -python ${CUR_DIR}/resources/tools/topo_installation.py -t ${WORKING_TOPOLOGY} \ - -d ${INSTALLATION_DIR} \ - -p ${VPP_DEBS} +python ${SCRIPT_DIR}/resources/tools/scripts/topo_installation.py \ + -t ${WORKING_TOPOLOGY} -d ${INSTALLATION_DIR} -p ${DUT_PKGS} if [ $? -eq 0 ]; then - echo "VPP Installed on hosts from: ${WORKING_TOPOLOGY}" + echo "DUT installed on hosts from: ${WORKING_TOPOLOGY}" else - echo "Failed to copy vpp deb files to DUTs" + echo "Failed to copy DUT packages files to hosts from: ${WORKING_TOPOLOGY}" exit 1 fi +# CSIT EXECUTION +PYBOT_ARGS="--consolewidth 100 \ + --loglevel TRACE \ + --variable TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ + --suite tests.${DUT}.perf" + case "$TEST_TAG" in - # run specific performance tests based on jenkins job type variable - PERFTEST_LONG ) - pybot ${PYBOT_ARGS} \ - -L TRACE \ - -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ - -i perftest_long \ - tests/ + # select specific performance tests based on jenkins job type variable + PERFTEST_DAILY ) + TAGS=('ndrdiscANDnic_intel-x520-da2AND1c' + 'ndrdiscANDnic_intel-x520-da2AND2c' + 'ndrpdrANDnic_intel-x520-da2AND1c' + 'ndrpdrANDnic_intel-x520-da2AND2c' + 'ndrdiscAND1cANDipsec' + 'ndrdiscAND2cANDipsec') ;; - PERFTEST_SHORT ) - pybot ${PYBOT_ARGS} \ - -L TRACE \ - -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ - -i perftest_short \ - tests/ + PERFTEST_SEMI_WEEKLY ) + TAGS=('ndrdiscANDnic_intel-x710AND1c' + 'ndrdiscANDnic_intel-x710AND2c' + 'ndrdiscANDnic_intel-xl710AND1c' + 'ndrdiscANDnic_intel-xl710AND2c') ;; - PERFTEST_LONG_BRIDGE ) - pybot ${PYBOT_ARGS} \ - -L TRACE \ - -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ - -s "performance.Long_Bridge_Domain*" \ - tests/ + PERFTEST_MRR_DAILY ) + TAGS=('mrrAND64bAND1c' + 'mrrAND64bAND2c' + 'mrrAND64bAND4c' + 'mrrAND78bAND1c' + 'mrrAND78bAND2c' + 'mrrAND78bAND4c' + 'mrrANDimixAND1cANDvhost' + 'mrrANDimixAND2cANDvhost' + 'mrrANDimixAND4cANDvhost' + 'mrrANDimixAND1cANDmemif' + 'mrrANDimixAND2cANDmemif' + 'mrrANDimixAND4cANDmemif') ;; - PERFTEST_LONG_IPV4 ) - pybot ${PYBOT_ARGS} \ - -L TRACE \ - -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ - -s "performance.Long_IPv4*" \ - tests/ - ;; - PERFTEST_LONG_IPV6 ) - pybot ${PYBOT_ARGS} \ - -L TRACE \ - -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ - -s "performance.Long_IPv6*" \ - tests/ - ;; - PERFTEST_LONG_XCONNECT ) - pybot ${PYBOT_ARGS} \ - -L TRACE \ - -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ - -s "performance.Long_Xconnect*" \ - tests/ + VERIFY-PERF-PATCH ) + if [[ -z "$TEST_TAG_STRING" ]]; then + # If nothing is specified, we will run pre-selected tests by + # following tags. Items of array will be concatenated by OR in Robot + # Framework. + TEST_TAG_ARRAY=('mrrANDnic_intel-x710AND1cAND64bANDip4base' + 'mrrANDnic_intel-x710AND1cAND78bANDip6base' + 'mrrANDnic_intel-x710AND1cAND64bANDl2bdbase') + else + # If trigger contains tags, split them into array. + TEST_TAG_ARRAY=(${TEST_TAG_STRING//:/ }) + fi + + TAGS=() + + for TAG in "${TEST_TAG_ARRAY[@]}"; do + if [[ ${TAG} == "!"* ]]; then + # Exclude tags are not prefixed. + TAGS+=("${TAG}") + else + # We will prefix with perftest to prevent running other tests + # (e.g. Functional). + prefix="perftestAND" + if [[ ${JOB_NAME} == vpp-* ]] ; then + # Automatic prefixing for VPP jobs to limit the NIC used and + # traffic evaluation to MRR. + prefix="${prefix}mrrANDnic_intel-x710AND" + fi + TAGS+=("$prefix${TAG}") + fi + done ;; * ) - # run full performance test suite and exit on fail - pybot ${PYBOT_ARGS} \ - -L TRACE \ - -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ - -s performance \ - tests/ + TAGS=('perftest') esac +# Catenate TAG selections +EXPANDED_TAGS=() +for TAG in "${TAGS[@]}"; do + if [[ ${TAG} == "!"* ]]; then + EXPANDED_TAGS+=(" --exclude ${TAG#$"!"} ") + else + EXPANDED_TAGS+=(" --include ${TOPOLOGIES_TAGS}AND${TAG} ") + fi +done + +# Execute the test +pybot ${PYBOT_ARGS}${EXPANDED_TAGS[@]} tests/ +RETURN_STATUS=$(echo $?) + +# Archive JOB artifacts in jenkins +for i in ${JOB_ARCHIVE_ARTIFACTS[@]}; do + cp $( readlink -f ${i} | tr '\n' ' ' ) ${JOB_ARCHIVE_DIR}/ +done +# Archive JOB artifacts to logs.fd.io +for i in ${LOG_ARCHIVE_ARTIFACTS[@]}; do + cp $( readlink -f ${i} | tr '\n' ' ' ) ${LOG_ARCHIVE_DIR}/ +done + +exit ${RETURN_STATUS}