X-Git-Url: https://gerrit.fd.io/r/gitweb?p=csit.git;a=blobdiff_plain;f=resources%2Flibraries%2Fpython%2FDUTSetup.py;h=e2fba066d3308f0b35789da10d7b3e358b9a5e2a;hp=4fc0e6fc9c97eca921f50b7c1c144fb78be973fa;hb=e001fdea995835f1ef75a5e21607ba02d78e4068;hpb=8fa4cadbb4b3c6a95dab9e7715e5e6435225673f diff --git a/resources/libraries/python/DUTSetup.py b/resources/libraries/python/DUTSetup.py index 4fc0e6fc9c..e2fba066d3 100644 --- a/resources/libraries/python/DUTSetup.py +++ b/resources/libraries/python/DUTSetup.py @@ -15,7 +15,7 @@ from robot.api import logger -from resources.libraries.python.constants import Constants +from resources.libraries.python.Constants import Constants from resources.libraries.python.ssh import SSH, exec_cmd_no_error from resources.libraries.python.topology import NodeType, Topology @@ -33,7 +33,8 @@ class DUTSetup(object): :type service: str """ if DUTSetup.running_in_container(node): - command = 'echo $(< /var/log/supervisord.log)' + command = ('echo $(< /var/log/supervisord.log);' + 'echo $(< /tmp/*supervisor*.log)') else: command = ('journalctl --no-pager --unit={name} ' '--since="$(echo `systemctl show -p ' @@ -59,6 +60,39 @@ class DUTSetup(object): if node['type'] == NodeType.DUT: DUTSetup.get_service_logs(node, service) + @staticmethod + def restart_service(node, service): + """Restarts the named service on node. + + :param node: Node in the topology. + :param service: Service unit name. + :type node: dict + :type service: str + """ + if DUTSetup.running_in_container(node): + command = 'supervisorctl restart {name}'.format(name=service) + else: + command = 'service {name} restart'.format(name=service) + message = 'Node {host} failed to restart service {name}'.\ + format(host=node['host'], name=service) + + exec_cmd_no_error(node, command, timeout=30, sudo=True, message=message) + + DUTSetup.get_service_logs(node, service) + + @staticmethod + def restart_service_on_all_duts(nodes, service): + """Retarts the named service on all DUTs. + + :param node: Nodes in the topology. + :param service: Service unit name. + :type node: dict + :type service: str + """ + for node in nodes.values(): + if node['type'] == NodeType.DUT: + DUTSetup.restart_service(node, service) + @staticmethod def start_service(node, service): """Start up the named service on node. @@ -68,6 +102,7 @@ class DUTSetup(object): :type node: dict :type service: str """ + # TODO: change command to start once all parent function updated. if DUTSetup.running_in_container(node): command = 'supervisorctl restart {name}'.format(name=service) else: @@ -175,19 +210,17 @@ class DUTSetup(object): 'on node: {0}\n {1}'. format(node['host'], stdout + stderr)) - if len(stdout.splitlines()) == 1: + pid_list = stdout.split() + if len(pid_list) == 1: return int(stdout) - elif not stdout.splitlines(): + elif not pid_list: logger.debug("No VPP PID found on node {0}". format(node['host'])) continue else: logger.debug("More then one VPP PID found on node {0}". format(node['host'])) - ret_list = list() - for line in stdout.splitlines(): - ret_list.append(int(line)) - return ret_list + return [int(pid) for pid in pid_list] return None @@ -207,17 +240,19 @@ class DUTSetup(object): return pids @staticmethod - def crypto_device_verify(node, force_init=False, numvfs=32): + def crypto_device_verify(node, crypto_type, numvfs, force_init=False): """Verify if Crypto QAT device virtual functions are initialized on all DUTs. If parameter force initialization is set to True, then try to initialize or remove VFs on QAT. :param node: DUT node. - :param force_init: If True then try to initialize to specific value. + :crypto_type: Crypto device type - HW_DH895xcc or HW_C3xxx. :param numvfs: Number of VFs to initialize, 0 - disable the VFs. + :param force_init: If True then try to initialize to specific value. :type node: dict - :type force_init: bool + :type crypto_type: string :type numvfs: int + :type force_init: bool :returns: nothing :raises RuntimeError: If QAT VFs are not created and force init is set to False. @@ -228,26 +263,38 @@ class DUTSetup(object): if sriov_numvfs != numvfs: if force_init: # QAT is not initialized and we want to initialize with numvfs - DUTSetup.crypto_device_init(node, numvfs) + DUTSetup.crypto_device_init(node, crypto_type, numvfs) else: raise RuntimeError('QAT device failed to create VFs on {host}'. format(host=node['host'])) @staticmethod - def crypto_device_init(node, numvfs): + def crypto_device_init(node, crypto_type, numvfs): """Init Crypto QAT device virtual functions on DUT. :param node: DUT node. + :crypto_type: Crypto device type - HW_DH895xcc or HW_C3xxx. :param numvfs: Number of VFs to initialize, 0 - disable the VFs. :type node: dict + :type crypto_type: string :type numvfs: int :returns: nothing :raises RuntimeError: If failed to stop VPP or QAT failed to initialize. """ + if crypto_type == "HW_DH895xcc": + kernel_mod = "qat_dh895xcc" + kernel_drv = "dh895xcc" + elif crypto_type == "HW_C3xxx": + kernel_mod = "qat_c3xxx" + kernel_drv = "c3xxx" + else: + raise RuntimeError('Unsupported crypto device type on {host}'. + format(host=node['host'])) + pci_addr = Topology.get_cryptodev(node) # QAT device must be re-bound to kernel driver before initialization. - DUTSetup.verify_kernel_module(node, 'qat_dh895xcc', force_load=True) + DUTSetup.verify_kernel_module(node, kernel_mod, force_load=True) # Stop VPP to prevent deadlock. DUTSetup.stop_service(node, Constants.VPP_UNIT) @@ -258,7 +305,7 @@ class DUTSetup(object): DUTSetup.pci_driver_unbind(node, pci_addr) # Bind to kernel driver. - DUTSetup.pci_driver_bind(node, pci_addr, 'dh895xcc') + DUTSetup.pci_driver_bind(node, pci_addr, kernel_drv) # Initialize QAT VFs. if numvfs > 0: @@ -595,73 +642,43 @@ class DUTSetup(object): exec_cmd_no_error(node, command, timeout=30, sudo=True, message=message) @staticmethod - def install_vpp_on_all_duts(nodes, vpp_pkg_dir, vpp_rpm_pkgs, vpp_deb_pkgs): - """Install VPP on all DUT nodes. + def install_vpp_on_all_duts(nodes, vpp_pkg_dir): + """Install VPP on all DUT nodes. Start the VPP service in case of + systemd is not available or does not support autostart. :param nodes: Nodes in the topology. :param vpp_pkg_dir: Path to directory where VPP packages are stored. - :param vpp_rpm_pkgs: List of VPP rpm packages to be installed. - :param vpp_deb_pkgs: List of VPP deb packages to be installed. :type nodes: dict :type vpp_pkg_dir: str - :type vpp_rpm_pkgs: list - :type vpp_deb_pkgs: list :raises RuntimeError: If failed to remove or install VPP. """ for node in nodes.values(): + message = 'Failed to install VPP on host {host}!'.\ + format(host=node['host']) if node['type'] == NodeType.DUT: - logger.debug("Installing VPP on node {0}".format(node['host'])) - - ssh = SSH() - ssh.connect(node) - - cmd = "[[ -f /etc/redhat-release ]]" - return_code, _, _ = ssh.exec_command(cmd) - if not int(return_code): - # workaroud - uninstall existing vpp installation until - # start-testcase script is updated on all virl servers - rpm_pkgs_remove = "vpp*" - cmd_u = 'yum -y remove "{0}"'.format(rpm_pkgs_remove) - r_rcode, _, r_err = ssh.exec_command_sudo(cmd_u, timeout=90) - if int(r_rcode): - raise RuntimeError('Failed to remove previous VPP' - 'installation on host {0}:\n{1}' - .format(node['host'], r_err)) - - rpm_pkgs = "*.rpm ".join(str(vpp_pkg_dir + pkg) - for pkg in vpp_rpm_pkgs) + "*.rpm" - cmd_i = "rpm -ivh {0}".format(rpm_pkgs) - ret_code, _, err = ssh.exec_command_sudo(cmd_i, timeout=90) - if int(ret_code): - raise RuntimeError('Failed to install VPP on host {0}:' - '\n{1}'.format(node['host'], err)) - else: - ssh.exec_command_sudo("rpm -qai vpp*") - logger.info("VPP installed on node {0}". - format(node['host'])) + command = 'ln -s /dev/null /etc/sysctl.d/80-vpp.conf || true' + exec_cmd_no_error(node, command, sudo=True) + + command = '. /etc/lsb-release; echo "${DISTRIB_ID}"' + stdout, _ = exec_cmd_no_error(node, command) + + if stdout.strip() == 'Ubuntu': + exec_cmd_no_error(node, 'apt-get purge -y "*vpp*" || true', + timeout=120, sudo=True) + exec_cmd_no_error(node, 'dpkg -i --force-all {dir}*.deb'. + format(dir=vpp_pkg_dir), timeout=120, + sudo=True, message=message) + exec_cmd_no_error(node, 'dpkg -l | grep vpp', sudo=True) + if DUTSetup.running_in_container(node): + DUTSetup.restart_service(node, Constants.VPP_UNIT) else: - # workaroud - uninstall existing vpp installation until - # start-testcase script is updated on all virl servers - deb_pkgs_remove = "vpp*" - cmd_u = 'apt-get purge -y "{0}"'.format(deb_pkgs_remove) - r_rcode, _, r_err = ssh.exec_command_sudo(cmd_u, timeout=90) - if int(r_rcode): - raise RuntimeError('Failed to remove previous VPP' - 'installation on host {0}:\n{1}' - .format(node['host'], r_err)) - deb_pkgs = "*.deb ".join(str(vpp_pkg_dir + pkg) - for pkg in vpp_deb_pkgs) + "*.deb" - cmd_i = "dpkg -i --force-all {0}".format(deb_pkgs) - ret_code, _, err = ssh.exec_command_sudo(cmd_i, timeout=90) - if int(ret_code): - raise RuntimeError('Failed to install VPP on host {0}:' - '\n{1}'.format(node['host'], err)) - else: - ssh.exec_command_sudo("dpkg -l | grep vpp") - logger.info("VPP installed on node {0}". - format(node['host'])) - - ssh.disconnect(node) + exec_cmd_no_error(node, 'yum -y remove "*vpp*" || true', + timeout=120, sudo=True) + exec_cmd_no_error(node, 'rpm -ivh {dir}*.rpm'. + format(dir=vpp_pkg_dir), timeout=120, + sudo=True, message=message) + exec_cmd_no_error(node, 'rpm -qai *vpp*', sudo=True) + DUTSetup.restart_service(node, Constants.VPP_UNIT) @staticmethod def running_in_container(node): @@ -682,6 +699,26 @@ class DUTSetup(object): return False return True + @staticmethod + def get_docker_mergeddir(node, uuid): + """Get Docker overlay for MergedDir diff. + + :param node: DUT node. + :param uuid: Docker UUID. + :type node: dict + :type uuid: str + :returns: Docker container MergedDir. + :rtype: str + :raises RuntimeError: If getting output failed. + """ + command = "docker inspect --format='"\ + "{{{{.GraphDriver.Data.MergedDir}}}}' {uuid}".format(uuid=uuid) + message = 'Failed to get directory of {uuid} on host {host}'.\ + format(uuid=uuid, host=node['host']) + + stdout, _ = exec_cmd_no_error(node, command, sudo=True, message=message) + return stdout.strip() + @staticmethod def get_huge_page_size(node): """Get default size of huge pages in system.