X-Git-Url: https://gerrit.fd.io/r/gitweb?p=csit.git;a=blobdiff_plain;f=bootstrap-verify-perf.sh;h=56e603709086b5784b07f46711bafb0ea640a12d;hp=746d1a0a584503320c27976d52081dfb97cf48c5;hb=b8bf181cafb0f4e8a317c308cfe83a3e022ce7c5;hpb=4ae37074a3e74bd6fdcd606d1553b99131372c84 diff --git a/bootstrap-verify-perf.sh b/bootstrap-verify-perf.sh index 746d1a0a58..56e6037090 100755 --- a/bootstrap-verify-perf.sh +++ b/bootstrap-verify-perf.sh @@ -1,5 +1,5 @@ #!/bin/bash -# Copyright (c) 2016 Cisco and/or its affiliates. +# Copyright (c) 2018 Cisco and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: @@ -12,25 +12,66 @@ # See the License for the specific language governing permissions and # limitations under the License. -set -x +set -xo pipefail # Space separated list of available testbeds, described by topology files -TOPOLOGIES="topologies/available/lf_testbed2-710-520.yaml" +TOPOLOGIES="topologies/available/lf_3n_hsw_testbed1.yaml \ + topologies/available/lf_3n_hsw_testbed2.yaml \ + topologies/available/lf_3n_hsw_testbed3.yaml" + +SCRIPT_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )" +export PYTHONPATH=${SCRIPT_DIR} +export DEBIAN_FRONTEND=noninteractive # Reservation dir RESERVATION_DIR="/tmp/reservation_dir" +INSTALLATION_DIR="/tmp/install_dir" + +JOB_ARCHIVE_ARTIFACTS=(log.html output.xml report.html) +LOG_ARCHIVE_ARTIFACTS=(log.html output.xml report.html) +JOB_ARCHIVE_DIR="archive" +LOG_ARCHIVE_DIR="$WORKSPACE/archives" +mkdir -p ${JOB_ARCHIVE_DIR} +mkdir -p ${LOG_ARCHIVE_DIR} + +# If we run this script from CSIT jobs we want to use stable vpp version +if [[ ${JOB_NAME} == csit-* ]] ; +then + if [[ ${TEST_TAG} == *DAILY ]] || \ + [[ ${TEST_TAG} == *WEEKLY ]]; + then + echo Downloading latest VPP packages from NEXUS... + bash ${SCRIPT_DIR}/resources/tools/scripts/download_install_vpp_pkgs.sh \ + --skip-install + else + echo Downloading VPP packages of specific version from NEXUS... + DPDK_STABLE_VER=$(cat ${SCRIPT_DIR}/DPDK_STABLE_VER) + VPP_STABLE_VER=$(cat ${SCRIPT_DIR}/VPP_STABLE_VER_UBUNTU) + #Temporary if arch will not be removed from VPP_STABLE_VER_UBUNTU + #VPP_STABLE_VER=${VPP_STABLE_VER%_amd64} + bash ${SCRIPT_DIR}/resources/tools/scripts/download_install_vpp_pkgs.sh \ + --skip-install --vpp ${VPP_STABLE_VER} --dkms ${DPDK_STABLE_VER} + fi + # Jenkins VPP deb paths (convert to full path) + VPP_DEBS="$( readlink -f vpp*.deb | tr '\n' ' ' )" -# Jenkins VPP deb paths (convert to full path) -VPP_DEBS="$( readlink -f $@ | tr '\n' ' ' )" +# If we run this script from vpp project we want to use local build +elif [[ ${JOB_NAME} == vpp-* ]] ; +then + # Use local packages provided as argument list + # Jenkins VPP deb paths (convert to full path) + VPP_DEBS="$( readlink -f $@ | tr '\n' ' ' )" +else + echo "Unable to identify job type based on JOB_NAME variable: ${JOB_NAME}" + exit 1 +fi -CUR_DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )" WORKING_TOPOLOGY="" -export PYTHONPATH=${CUR_DIR} sudo apt-get -y update sudo apt-get -y install libpython2.7-dev python-virtualenv -virtualenv env +virtualenv --system-site-packages env . env/bin/activate echo pip install @@ -40,7 +81,7 @@ pip install -r requirements.txt while :; do for TOPOLOGY in ${TOPOLOGIES}; do - python ${CUR_DIR}/resources/tools/topo_reservation.py -t ${TOPOLOGY} + python ${SCRIPT_DIR}/resources/tools/scripts/topo_reservation.py -t ${TOPOLOGY} if [ $? -eq 0 ]; then WORKING_TOPOLOGY=${TOPOLOGY} echo "Reserved: ${WORKING_TOPOLOGY}" @@ -53,30 +94,232 @@ while :; do break fi - # Wait 10 - 30 sec. before next try - SLEEP_TIME=$[ ( $RANDOM % 20 ) + 10 ]s + # Wait ~3minutes before next try + SLEEP_TIME=$[ ( $RANDOM % 20 ) + 180 ]s echo "Sleeping ${SLEEP_TIME}" sleep ${SLEEP_TIME} done -python ${CUR_DIR}/resources/tools/topo_installation.py -t ${WORKING_TOPOLOGY} \ - -d ${RESERVATION_DIR} \ - -p ${VPP_DEBS} +function cancel_all { + python ${SCRIPT_DIR}/resources/tools/scripts/topo_installation.py -c -d ${INSTALLATION_DIR} -t $1 + python ${SCRIPT_DIR}/resources/tools/scripts/topo_reservation.py -c -t $1 +} + +# On script exit we cancel the reservation and installation and delete all vpp +# packages +trap "cancel_all ${WORKING_TOPOLOGY}" EXIT + +python ${SCRIPT_DIR}/resources/tools/scripts/topo_installation.py \ + -t ${WORKING_TOPOLOGY} -d ${INSTALLATION_DIR} -p ${VPP_DEBS} if [ $? -eq 0 ]; then echo "VPP Installed on hosts from: ${WORKING_TOPOLOGY}" else echo "Failed to copy vpp deb files to DUTs" - exit $? + exit 1 fi -function cancel_reservation { - python ${CUR_DIR}/resources/tools/topo_reservation.py -c -t $1 -} +# Based on job we will identify DUT +if [[ ${JOB_NAME} == *hc2vpp* ]] ; +then + DUT="hc2vpp" +elif [[ ${JOB_NAME} == *vpp* ]] ; +then + DUT="vpp" +elif [[ ${JOB_NAME} == *ligato* ]] ; +then + DUT="kubernetes" +elif [[ ${JOB_NAME} == *dpdk* ]] ; +then + DUT="dpdk" +else + echo "Unable to identify dut type based on JOB_NAME variable: ${JOB_NAME}" + exit 1 +fi -# On script exit we cancel the reservation and delete all vpp packages -trap "cancel_reservation ${WORKING_TOPOLOGY}" EXIT +PYBOT_ARGS="--consolewidth 120 --loglevel TRACE --variable TOPOLOGY_PATH:${WORKING_TOPOLOGY} --suite tests.${DUT}.perf" + +case "$TEST_TAG" in + # select specific performance tests based on jenkins job type variable + PERFTEST_DAILY ) + TAGS=('ndrdiscANDnic_intel-x520-da2AND1t1c' + 'ndrdiscANDnic_intel-x520-da2AND2t2c' + 'ndrpdrANDnic_intel-x520-da2AND1t1c' + 'ndrpdrANDnic_intel-x520-da2AND2t2c' + 'ndrdiscAND1t1cANDipsec' + 'ndrdiscAND2t2cANDipsec') + ;; + PERFTEST_SEMI_WEEKLY ) + TAGS=('ndrdiscANDnic_intel-x710AND1t1c' + 'ndrdiscANDnic_intel-x710AND2t2c' + 'ndrdiscANDnic_intel-xl710AND1t1c' + 'ndrdiscANDnic_intel-xl710AND2t2c') + ;; + PERFTEST_MRR_DAILY ) + TAGS=('mrrAND64bAND1t1c' + 'mrrAND64bAND2t2c' + 'mrrAND64bAND4t4c' + 'mrrAND78bAND1t1c' + 'mrrAND78bAND2t2c' + 'mrrAND78bAND4t4c' + 'mrrANDimixAND1t1cANDvhost' + 'mrrANDimixAND2t2cANDvhost' + 'mrrANDimixAND4t4cANDvhost' + 'mrrANDimixAND1t1cANDmemif' + 'mrrANDimixAND2t2cANDmemif' + 'mrrANDimixAND4t4cANDmemif') + ;; + VERIFY-PERF-NDRDISC ) + TAGS=('ndrdiscAND1t1c' + 'ndrdiscAND2t2c') + ;; + VERIFY-PERF-PDRDISC ) + TAGS=('pdrdiscAND1t1c' + 'pdrdiscAND2t2c') + ;; + VERIFY-PERF-MRR ) + TAGS=('mrrAND1t1c' + 'mrrAND2t2c') + ;; + VERIFY-PERF-IP4 ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDip4base' + 'mrrANDnic_intel-x520-da2AND1t1cANDip4fwdANDfib_2m') + ;; + VERIFY-PERF-IP6 ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDip6base' + 'mrrANDnic_intel-x520-da2AND1t1cANDip6fwdANDfib_2m') + ;; + VERIFY-PERF-L2 ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDl2xcbase' + 'mrrANDnic_intel-x520-da2AND1t1cANDl2bdbase' + '!lbond_dpdk') + ;; + VERIFY-PERF-LISP ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDlisp') + ;; + VERIFY-PERF-VXLAN ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDvxlan') + ;; + VERIFY-PERF-VHOST ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDvhost' + '!lbond_dpdk') + ;; + VERIFY-PERF-MEMIF ) + TAGS=('pdrdiscANDnic_intel-x520-da2AND1t1cANDmemif' + 'pdrdiscANDnic_intel-x520-da2AND2t2cANDmemif' + 'mrrANDnic_intel-x520-da2AND1t1cANDmemif' + 'mrrANDnic_intel-x520-da2AND2t2cANDmemif') + ;; + VERIFY-PERF-IPSECHW ) + TAGS=('pdrdiscANDnic_intel-xl710AND1t1cANDipsechw' + 'pdrdiscANDnic_intel-xl710AND2t2cANDipsechw' + 'mrrANDnic_intel-xl710AND1t1cANDipsechw' + 'mrrANDnic_intel-xl710AND2t2cANDipsechw') + ;; + VERIFY-PERF-SRV6 ) + TAGS=('mrrANDsrv6AND1t1c' + 'mrrANDsrv6AND2t2c') + ;; + VPP-VERIFY-PERF-IP4 ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDip4base' + 'mrrANDnic_intel-x520-da2AND1t1cANDip4fwdANDfib_2m') + ;; + VPP-VERIFY-PERF-IP6 ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDip6base' + 'mrrANDnic_intel-x520-da2AND1t1cANDip6fwdANDfib_2m') + ;; + VPP-VERIFY-PERF-L2 ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDl2xcbase' + 'mrrANDnic_intel-x520-da2AND1t1cANDl2bdbase' + 'mrrANDnic_intel-x520-da2AND1t1cANDdot1q' + '!lbond_dpdk') + ;; + VPP-VERIFY-PERF-LISP ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDlisp') + ;; + VPP-VERIFY-PERF-VXLAN ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDvxlan') + ;; + VPP-VERIFY-PERF-VHOST ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDvhost' + '!lbond_dpdk') + ;; + VPP-VERIFY-PERF-MEMIF ) + TAGS=('pdrdiscANDnic_intel-x520-da2AND1t1cANDmemif' + 'pdrdiscANDnic_intel-x520-da2AND2t2cANDmemif' + 'mrrANDnic_intel-x520-da2AND1t1cANDmemif' + 'mrrANDnic_intel-x520-da2AND2t2cANDmemif') + ;; + VPP-VERIFY-PERF-ACL ) + TAGS=('mrrANDnic_intel-x520-da2AND1t1cANDacl' + 'mrrANDnic_intel-x520-da2AND2t2cANDacl') + ;; + VPP-VERIFY-PERF-IPSECHW ) + TAGS=('pdrdiscANDnic_intel-xl710AND1t1cANDipsechw' + 'pdrdiscANDnic_intel-xl710AND2t2cANDipsechw' + 'mrrANDnic_intel-xl710AND1t1cANDipsechw' + 'mrrANDnic_intel-xl710AND2t2cANDipsechw') + ;; + VPP-VERIFY-PERF-SRV6 ) + TAGS=('mrrANDsrv6AND1t1c' + 'mrrANDsrv6AND2t2c') + ;; + VERIFY-PERF-PATCH ) + if [[ -z "$TEST_TAG_STRING" ]]; then + # If nothing is specified, we will run pre-selected tests by + # following tags. Items of array will be concatenated by OR in Robot + # Framework. + TEST_TAG_ARRAY=('mrrANDnic_intel-x710AND1t1cAND64bANDip4base' + 'mrrANDnic_intel-x710AND1t1cAND78bANDip6base' + 'mrrANDnic_intel-x710AND1t1cAND64bANDl2bdbase') + else + # If trigger contains tags, split them into array. + TEST_TAG_ARRAY=(${TEST_TAG_STRING//:/ }) + fi + + TAGS=() + + for TAG in "${TEST_TAG_ARRAY[@]}"; do + if [[ ${TAG} == "!"* ]]; then + # Exclude tags are not prefixed. + TAGS+=("${TAG}") + else + # We will prefix with perftest to prevent running other tests + # (e.g. Functional). + prefix="perftestAND" + if [[ ${JOB_NAME} == vpp-* ]] ; then + # Automatic prefixing for VPP jobs to limit the NIC used and + # traffic evaluation to MRR. + prefix="${prefix}mrrANDnic_intel-x710AND" + fi + TAGS+=("$prefix${TAG}") + fi + done + ;; + * ) + TAGS=('perftest') +esac + +# Catenate TAG selections +EXPANDED_TAGS=() +for TAG in "${TAGS[@]}"; do + if [[ ${TAG} == "!"* ]]; then + EXPANDED_TAGS+=(" --exclude ${TAG#$"!"} ") + else + EXPANDED_TAGS+=(" --include ${TAG} ") + fi +done + +# Execute the test +pybot ${PYBOT_ARGS}${EXPANDED_TAGS[@]} tests/ +RETURN_STATUS=$(echo $?) + +# Archive JOB artifacts in jenkins +for i in ${JOB_ARCHIVE_ARTIFACTS[@]}; do + cp $( readlink -f ${i} | tr '\n' ' ' ) ${JOB_ARCHIVE_DIR}/ +done +# Archive JOB artifacts to logs.fd.io +for i in ${LOG_ARCHIVE_ARTIFACTS[@]}; do + cp $( readlink -f ${i} | tr '\n' ' ' ) ${LOG_ARCHIVE_DIR}/ +done -# run performance test suite -pybot -L TRACE \ - -v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \ - -s performance tests/ +exit ${RETURN_STATUS}