ARCHIVE_ARTIFACTS=(log.html output.xml report.html output_perf_data.xml)
+# If we run this script from CSIT jobs we want to use stable vpp version
+if [[ ${JOB_NAME} == csit-* ]] ;
+then
+ mkdir -p vpp/build-root
+ cd vpp/build-root
+
+ if [[ ${TEST_TAG} == *NIGHTLY ]] || \
+ [[ ${TEST_TAG} == *DAILY ]] || \
+ [[ ${TEST_TAG} == *WEEKLY ]];
+ then
+ # Download the latest VPP build .deb install packages
+ echo Downloading VPP packages...
+ bash ${SCRIPT_DIR}/resources/tools/scripts/download_install_vpp_pkgs.sh --skip-install
+
+ VPP_DEBS="$( readlink -f *.deb | tr '\n' ' ' )"
+ # Take vpp package and get the vpp version
+ VPP_STABLE_VER="$( expr match $(ls *.deb | head -n 1) 'vpp-\(.*\)-deb.deb' )"
+ else
+ DPDK_STABLE_VER=$(cat ${SCRIPT_DIR}/DPDK_STABLE_VER)_amd64
+ VPP_REPO_URL=$(cat ${SCRIPT_DIR}/VPP_REPO_URL_UBUNTU)
+ VPP_STABLE_VER=$(cat ${SCRIPT_DIR}/VPP_STABLE_VER_UBUNTU)
+ VPP_CLASSIFIER="-deb"
+ # Download vpp build from nexus and set VPP_DEBS variable
+ wget -q "${VPP_REPO_URL}/vpp/${VPP_STABLE_VER}/vpp-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit
+ wget -q "${VPP_REPO_URL}/vpp-dbg/${VPP_STABLE_VER}/vpp-dbg-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit
+ wget -q "${VPP_REPO_URL}/vpp-dev/${VPP_STABLE_VER}/vpp-dev-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit
+ # Temporary disable using dpdk
+ # wget -q "${VPP_REPO_URL}/vpp-dpdk-dkms/${DPDK_STABLE_VER}/vpp-dpdk-dkms-${DPDK_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit
+ wget -q "${VPP_REPO_URL}/vpp-lib/${VPP_STABLE_VER}/vpp-lib-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit
+ wget -q "${VPP_REPO_URL}/vpp-plugins/${VPP_STABLE_VER}/vpp-plugins-${VPP_STABLE_VER}${VPP_CLASSIFIER}.deb" || exit
+ VPP_DEBS="$( readlink -f *.deb | tr '\n' ' ' )"
+ fi
+
+ # Temporary workaround as ligato docker file requires specific file name
+ rename -v 's/^(.*)-(\d.*)-deb.deb/$1_$2.deb/' *.deb
+ cd ${SCRIPT_DIR}
+
+# If we run this script from vpp project we want to use local build
+elif [[ ${JOB_NAME} == vpp-* ]] ;
+then
+ mkdir -p vpp/build-root
+ # Use local packages provided as argument list
+ # Jenkins VPP deb paths (convert to full path)
+ VPP_DEBS="$( readlink -f $@ | tr '\n' ' ' )"
+ # Take vpp package and get the vpp version
+ VPP_STABLE_VER="$( expr match $1 'vpp-\(.*\)-deb.deb' )"
+ # Move files to build-root for packing
+ for deb in ${VPP_DEBS}; do mv ${deb} vpp/build-root/; done
+else
+ echo "Unable to identify job type based on JOB_NAME variable: ${JOB_NAME}"
+ exit 1
+fi
+dpkg -x vpp/build-root/vpp_${VPP_STABLE_VER}.deb /tmp/vpp
+
+# Compress all VPP debs and remove temporary directory
+tar -zcvf ${SCRIPT_DIR}/vpp.tar.gz vpp/* && rm -R vpp
+
LIGATO_REPO_URL=$(cat ${SCRIPT_DIR}/LIGATO_REPO_URL)
-LIGATO_STABLE_VER=$(cat ${SCRIPT_DIR}/LIGATO_STABLE_VER)
-VPP_COMMIT=$1
-VPP_BUILD=$1
-DOCKER_DEB="docker-ce_17.06.2~ce-0~ubuntu_amd64.deb"
+VPP_AGENT_STABLE_VER=$(cat ${SCRIPT_DIR}/VPP_AGENT_STABLE_VER)
+VPP_AGENT_STABLE_COMMIT="$( expr match `cat VPP_AGENT_STABLE_VER` '.*g\(.*\)' )"
+DOCKER_DEB="docker-ce_17.09.0~ce-0~ubuntu_amd64.deb"
# Clone & checkout stable vnf-agent
cd .. && git clone ${LIGATO_REPO_URL}/vpp-agent
echo "Failed to run: git clone --depth 1 ${LIGATO_REPO_URL}/vpp-agent"
exit 1
fi
-cd vpp-agent && git checkout ${LIGATO_STABLE_VER}
+cd vpp-agent && git checkout b99e43a
# If the git checkout fails, complain clearly and exit
if [ $? != 0 ]; then
- echo "Failed to run: git checkout ${LIGATO_STABLE_VER}"
+ echo "Failed to run: git checkout ${VPP_AGENT_STABLE_VER}"
exit 1
fi
exit 1
fi
-# Compile vnf-agent docker image
-cd ${SCRIPT_DIR}/../vpp-agent/docker/dev_vpp_agent/ &&\
- ./build.sh --agent ${LIGATO_STABLE_VER} --vpp ${VPP_COMMIT} &&\
- ./shrink.sh
+# Pull ligato/dev_vpp_agent docker image and re-tag as local
+if [[ ${VPP_AGENT_STABLE_VER} == g* ]] ;
+then
+ sudo docker pull ligato/dev-vpp-agent:${VPP_AGENT_STABLE_COMMIT}
+ sudo docker tag ligato/dev-vpp-agent:${VPP_AGENT_STABLE_COMMIT}\
+ dev_vpp_agent:latest
+else
+ sudo docker pull ligato/dev-vpp-agent:${VPP_AGENT_STABLE_VER}
+ sudo docker tag ligato/dev-vpp-agent:${VPP_AGENT_STABLE_VER}\
+ dev_vpp_agent:latest
+fi
+sudo docker images
+# Start dev_vpp_agent container as daemon
+sudo docker run --rm -itd --name agentcnt dev_vpp_agent bash
+# Copy latest vpp api into running container
+sudo docker cp /tmp/vpp/usr/share/vpp/api agentcnt:/usr/share/vpp
+# Recompile vpp-agent
+sudo docker exec -i agentcnt \
+ script -qec '. ~/.bashrc; cd /root/go/src/github.com/ligato/vpp-agent && make generate && make install'
+if [ $? != 0 ]; then
+ echo "Failed to build vpp-agent in Docker image."
+ exit 1
+fi
+# Extract vpp-agent
+rm -rf agent
+mkdir -p agent
+sudo docker cp agentcnt:/root/go/bin/vpp-agent agent/
+sudo docker cp agentcnt:/root/go/bin/vpp-agent-ctl agent/
+sudo docker cp agentcnt:/root/go/bin/agentctl agent/
+tar -zcvf ${SCRIPT_DIR}/../vpp-agent/docker/prod_vpp_agent/agent.tar.gz agent
+# Kill running container
+sudo docker rm -f agentcnt
+
+# Build prod_vpp_agent docker image
cd ${SCRIPT_DIR}/../vpp-agent/docker/prod_vpp_agent/ &&\
- ./build.sh &&\
- ./shrink.sh
+ mv ${SCRIPT_DIR}/vpp.tar.gz . &&\
+ sudo docker build -t prod_vpp_agent --no-cache .
# Export Docker image
-sudo docker save prod_vpp_agent_shrink | gzip > prod_vpp_agent_shrink.tar.gz
+sudo docker save prod_vpp_agent | gzip > prod_vpp_agent.tar.gz
# If image build fails, complain clearly and exit
if [ $? != 0 ]; then
echo "Failed to build vpp-agent Docker image."
exit 1
fi
-DOCKER_IMAGE="$( readlink -f prod_vpp_agent_shrink.tar.gz | tr '\n' ' ' )"
+DOCKER_IMAGE="$( readlink -f prod_vpp_agent.tar.gz | tr '\n' ' ' )"
cd ${SCRIPT_DIR}
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include ndrdiscANDnic_intel-x520-da2AND1t1cORndrdiscANDnic_intel-x520-da2AND2t2c \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include ndrdiscANDnic_intel-x710AND1t1cORndrdiscANDnic_intel-x710AND2t2cORndrdiscANDnic_intel-xl710AND1t1cORndrdiscANDnic_intel-xl710AND2t2c \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include ndrdiscAND1t1cORndrdiscAND2t2c \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include pdrdiscAND1t1cORpdrdiscAND2t2c \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include ndrchkAND1t1cORndrchkAND2t2c \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include ndrchkAND1t1cORndrchkAND2t2c \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include ndrdiscANDnic_intel-x520-da2AND1t1cANDip4baseORndrdiscANDnic_intel-x520-da2AND1t1cANDip4fwdANDfib_2m \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include ndrdiscANDnic_intel-x520-da2AND1t1cANDip6baseORndrdiscANDnic_intel-x520-da2AND1t1cANDip6fwdANDfib_2m \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include ndrdiscANDnic_intel-x520-da2AND1t1cANDl2xcbaseORndrdiscANDnic_intel-x520-da2AND1t1cANDl2bdbase \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include ndrdiscANDnic_intel-x520-da2AND1t1cANDlisp \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include ndrdiscANDnic_intel-x520-da2AND1t1cANDvxlan \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include ndrdiscANDnic_intel-x520-da2AND1t1cANDvhost \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include pdrchkANDnic_intel-x520-da2AND1t1cANDip4baseORpdrchkANDnic_intel-x520-da2AND1t1cANDip4fwdANDfib_2m \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include pdrchkANDnic_intel-x520-da2AND1t1cANDip6baseORpdrchkANDnic_intel-x520-da2AND1t1cANDip6fwdANDfib_2m \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include pdrchkANDnic_intel-x520-da2AND1t1cANDl2xcbaseORpdrchkANDnic_intel-x520-da2AND1t1cANDl2bdbase \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include pdrchkANDnic_intel-x520-da2AND1t1cANDlisp \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include pdrchkANDnic_intel-x520-da2AND1t1cANDvxlan \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include pdrdiscANDnic_intel-x520-da2AND1t1cANDvhost \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--include pdrdiscANDnic_intel-x520-da2AND1t1cANDacl \
--include pdrdiscANDnic_intel-x520-da2AND2t2cANDacl \
tests/
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
--exclude SKIP_PATCH \
-i NDRPDRDISC \
tests/
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
-i NDRCHK \
tests/
RETURN_STATUS=$(echo $?)
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
tests/
RETURN_STATUS=$(echo $?)
;;
pybot ${PYBOT_ARGS} \
-v TOPOLOGY_PATH:${WORKING_TOPOLOGY} \
-v DPDK_TEST:True \
- -s "tests.ligato.perf" \
+ -s "tests.kubernetes.perf" \
tests/
RETURN_STATUS=$(echo $?)
esac
python ${SCRIPT_DIR}/resources/tools/scripts/robot_output_parser.py \
-i ${SCRIPT_DIR}/output.xml \
-o ${SCRIPT_DIR}/output_perf_data.xml \
- -v ${VPP_BUILD}
+ -v ${VPP_STABLE_VER}
if [ ! $? -eq 0 ]; then
echo "Parsing ${SCRIPT_DIR}/output.xml failed"
fi