X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=resources%2Flibraries%2Fpython%2FSetupFramework.py;h=9aff15278e6ecda3a234706a6569c410f41f268e;hb=91237b65fb3aa3409456ff6ad57caab08a3041cc;hp=b81a3065c68535ef52ba2eec2ffd0f22be26d5e3;hpb=5eb99d868051556dce3d509545d130971d74e1fa;p=csit.git diff --git a/resources/libraries/python/SetupFramework.py b/resources/libraries/python/SetupFramework.py index b81a3065c6..9aff15278e 100644 --- a/resources/libraries/python/SetupFramework.py +++ b/resources/libraries/python/SetupFramework.py @@ -1,4 +1,4 @@ -# Copyright (c) 2016 Cisco and/or its affiliates. +# Copyright (c) 2018 Cisco and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: @@ -16,40 +16,56 @@ nodes. All tasks required to be run before the actual tests are started is supposed to end up here. """ +import datetime +from os import environ, remove +from os.path import basename from shlex import split -from subprocess import Popen, PIPE, call -from multiprocessing import Pool +from subprocess import Popen, PIPE from tempfile import NamedTemporaryFile -from os.path import basename +import threading from robot.api import logger -from robot.libraries.BuiltIn import BuiltIn from resources.libraries.python.ssh import SSH -from resources.libraries.python.constants import Constants as con +from resources.libraries.python.Constants import Constants as con from resources.libraries.python.topology import NodeType __all__ = ["SetupFramework"] def pack_framework_dir(): - """Pack the testing WS into temp file, return its name.""" + """Pack the testing WS into temp file, return its name. + + :returns: Tarball file name. + :rtype: str + :raises Exception: When failed to pack testing framework. + """ + + try: + directory = environ["TMPDIR"] + except KeyError: + directory = None - tmpfile = NamedTemporaryFile(suffix=".tgz", prefix="openvpp-testing-") + if directory is not None: + tmpfile = NamedTemporaryFile(suffix=".tgz", prefix="csit-testing-", + dir="{0}".format(directory)) + else: + tmpfile = NamedTemporaryFile(suffix=".tgz", prefix="csit-testing-") file_name = tmpfile.name tmpfile.close() proc = Popen( - split("tar --exclude-vcs -zcf {0} .".format(file_name)), - stdout=PIPE, stderr=PIPE) + split("tar --sparse --exclude-vcs --exclude=output*.xml " + "--exclude=./tmp -zcf {0} ." + .format(file_name)), stdout=PIPE, stderr=PIPE) (stdout, stderr) = proc.communicate() logger.debug(stdout) logger.debug(stderr) return_code = proc.wait() - if 0 != return_code: - raise Exception("Could not pack testing framework.") + if return_code != 0: + raise RuntimeError("Could not pack testing framework.") return file_name @@ -61,86 +77,152 @@ def copy_tarball_to_node(tarball, node): :param node: Dictionary created from topology. :type tarball: str :type node: dict - :return: nothing + :returns: nothing """ logger.console('Copying tarball to {0}'.format(node['host'])) ssh = SSH() ssh.connect(node) ssh.scp(tarball, "/tmp/") + logger.console('Copying tarball to {0} done'.format(node['host'])) def extract_tarball_at_node(tarball, node): """Extract tarball at given node. - Extracts tarball using tar on given node to specific CSIT loocation. + Extracts tarball using tar on given node to specific CSIT location. :param tarball: Path to tarball to upload. :param node: Dictionary created from topology. :type tarball: str :type node: dict - :return: nothing + :returns: nothing + :raises RuntimeError: When failed to unpack tarball. """ - logger.console('Extracting tarball to {0} on {1}'.format( - con.REMOTE_FW_DIR, node['host'])) + logger.console('Extracting tarball to {0} on {1}' + .format(con.REMOTE_FW_DIR, node['host'])) ssh = SSH() ssh.connect(node) - - cmd = 'sudo rm -rf {1}; mkdir {1} ; tar -zxf {0} -C {1}; ' \ - 'rm -f {0}'.format(tarball, con.REMOTE_FW_DIR) - (ret_code, _, stderr) = ssh.exec_command(cmd, timeout=30) - if 0 != ret_code: - logger.error('Unpack error: {0}'.format(stderr)) - raise Exception('Failed to unpack {0} at node {1}'.format( - tarball, node['host'])) + (ret_code, _, _) = ssh.exec_command( + 'sudo rm -rf {1}; mkdir {1} ; tar -zxf {0} -C {1}; rm -f {0}' + .format(tarball, con.REMOTE_FW_DIR), timeout=30) + if ret_code != 0: + raise RuntimeError('Failed to extract {0} at node {1}' + .format(tarball, node['host'])) + logger.console('Extracting tarball to {0} on {1} done' + .format(con.REMOTE_FW_DIR, node['host'])) def create_env_directory_at_node(node): - """Create fresh virtualenv to a directory, install pip requirements.""" - logger.console('Extracting virtualenv, installing requirements.txt ' - 'on {0}'.format(node['host'])) + """Create fresh virtualenv to a directory, install pip requirements. + + :param node: Node to create virtualenv on. + :type node: dict + :returns: nothing + :raises RuntimeError: When failed to setup virtualenv. + """ + logger.console('Virtualenv setup including requirements.txt on {0}' + .format(node['host'])) ssh = SSH() ssh.connect(node) - (ret_code, stdout, stderr) = ssh.exec_command( - 'cd {0} && rm -rf env && virtualenv --system-site-packages env && . env/bin/activate && ' + (ret_code, _, _) = ssh.exec_command( + 'cd {0} && rm -rf env && ' + 'virtualenv --system-site-packages --never-download env && ' + '. env/bin/activate && ' 'pip install -r requirements.txt' .format(con.REMOTE_FW_DIR), timeout=100) - if 0 != ret_code: - logger.error('Virtualenv creation error: {0}'.format(stdout + stderr)) - raise Exception('Virtualenv setup failed') - else: - logger.console('Virtualenv created on {0}'.format(node['host'])) + if ret_code != 0: + raise RuntimeError('Virtualenv setup including requirements.txt on {0}' + .format(node['host'])) + logger.console('Virtualenv on {0} created'.format(node['host'])) -def setup_node(args): - """Run all set-up methods for a node. - This method is used as map_async parameter. It receives tuple with all - parameters as passed to map_async function. +def setup_node(node, tarball, remote_tarball, results=None): + """Copy a tarball to a node and extract it. - :param args: All parameters needed to setup one node. - :type args: tuple - :return: nothing + :param node: A node where the tarball will be copied and extracted. + :param tarball: Local path of tarball to be copied. + :param remote_tarball: Remote path of the tarball. + :param results: A list where to store the result of node setup, optional. + :type node: dict + :type tarball: str + :type remote_tarball: str + :type results: list + :returns: True - success, False - error + :rtype: bool """ - tarball, remote_tarball, node = args - copy_tarball_to_node(tarball, node) - extract_tarball_at_node(remote_tarball, node) - if node['type'] == NodeType.TG: - create_env_directory_at_node(node) - logger.console('Setup of node {0} done'.format(node['host'])) + try: + copy_tarball_to_node(tarball, node) + extract_tarball_at_node(remote_tarball, node) + if node['type'] == NodeType.TG: + create_env_directory_at_node(node) + except RuntimeError as exc: + logger.error("Node {0} setup failed, error:'{1}'" + .format(node['host'], exc.message)) + result = False + else: + logger.console('Setup of node {0} done'.format(node['host'])) + result = True + + if isinstance(results, list): + results.append(result) + return result def delete_local_tarball(tarball): """Delete local tarball to prevent disk pollution. - :param tarball: Path to tarball to upload. + :param tarball: Path of local tarball to delete. :type tarball: str - :return: nothing + :returns: nothing """ - call(split('sh -c "rm {0} > /dev/null 2>&1"'.format(tarball))) + remove(tarball) + + +def delete_framework_dir(node): + """Delete framework directory in /tmp/ on given node. + :param node: Node to delete framework directory on. + :type node: dict + """ + logger.console('Deleting framework directory on {0}' + .format(node['host'])) + ssh = SSH() + ssh.connect(node) + (ret_code, _, _) = ssh.exec_command( + 'sudo rm -rf {0}' + .format(con.REMOTE_FW_DIR), timeout=100) + if ret_code != 0: + raise RuntimeError('Deleting framework directory on {0} failed' + .format(node)) -class SetupFramework(object): # pylint: disable=too-few-public-methods + +def cleanup_node(node, results=None): + """Delete a tarball from a node. + + :param node: A node where the tarball will be delete. + :param results: A list where to store the result of node cleanup, optional. + :type node: dict + :type results: list + :returns: True - success, False - error + :rtype: bool + """ + try: + delete_framework_dir(node) + except RuntimeError: + logger.error("Cleanup of node {0} failed".format(node['host'])) + result = False + else: + logger.console('Cleanup of node {0} done'.format(node['host'])) + result = True + + if isinstance(results, list): + results.append(result) + return result + + +class SetupFramework(object): """Setup suite run on topology nodes. Many VAT/CLI based tests need the scripts at remote hosts before executing @@ -150,7 +232,12 @@ class SetupFramework(object): # pylint: disable=too-few-public-methods @staticmethod def setup_framework(nodes): - """Pack the whole directory and extract in temp on each node.""" + """Pack the whole directory and extract in temp on each node. + + :param nodes: Topology nodes. + :type nodes: dict + :raises RuntimeError: If setup framework failed. + """ tarball = pack_framework_dir() msg = 'Framework packed to {0}'.format(tarball) @@ -158,23 +245,62 @@ class SetupFramework(object): # pylint: disable=too-few-public-methods logger.trace(msg) remote_tarball = "/tmp/{0}".format(basename(tarball)) - # Turn off logging since we use multiprocessing - log_level = BuiltIn().set_log_level('NONE') - params = ((tarball, remote_tarball, node) for node in nodes.values()) - pool = Pool(processes=len(nodes)) - result = pool.map_async(setup_node, params) - pool.close() - pool.join() + results = [] + threads = [] + + for node in nodes.values(): + thread = threading.Thread(target=setup_node, args=(node, + tarball, + remote_tarball, + results)) + thread.start() + threads.append(thread) logger.info( - 'Executed node setups in parallel, waiting for processes to end') - result.wait() + 'Executing node setups in parallel, waiting for threads to end') + + for thread in threads: + thread.join() - logger.info('Results: {0}'.format(result.get())) + logger.info('Results: {0}'.format(results)) - # Turn on logging - BuiltIn().set_log_level(log_level) - logger.trace('Test framework copied to all topology nodes') delete_local_tarball(tarball) - logger.console('All nodes are ready') + if all(results): + logger.console('All nodes are ready') + else: + raise RuntimeError('Failed to setup framework') + + +class CleanupFramework(object): + """Clean up suite run on topology nodes.""" + + @staticmethod + def cleanup_framework(nodes): + """Perform cleanup on each node. + + :param nodes: Topology nodes. + :type nodes: dict + :raises RuntimeError: If cleanup framework failed. + """ + + results = [] + threads = [] + + for node in nodes.values(): + thread = threading.Thread(target=cleanup_node, + args=(node, results)) + thread.start() + threads.append(thread) + + logger.info( + 'Executing node cleanups in parallel, waiting for threads to end') + + for thread in threads: + thread.join() + + logger.info('Results: {0}'.format(results)) + if all(results): + logger.console('All nodes cleaned up') + else: + raise RuntimeError('Failed to cleaned up framework')