X-Git-Url: https://gerrit.fd.io/r/gitweb?p=csit.git;a=blobdiff_plain;f=resources%2Ftools%2Fpresentation%2Finput_data_files.py;h=5bd6af42d6d43458674a11176a05c90585eb0275;hp=441a240d79d8e7b95dc99c37b46f760e29de14c3;hb=c763cfcb064e4f4acf6b8309b08d3800b9bd5331;hpb=9e476b274acf41b5ac358bc4dba3a0f1e16f04b2 diff --git a/resources/tools/presentation/input_data_files.py b/resources/tools/presentation/input_data_files.py index 441a240d79..5bd6af42d6 100644 --- a/resources/tools/presentation/input_data_files.py +++ b/resources/tools/presentation/input_data_files.py @@ -1,4 +1,4 @@ -# Copyright (c) 2018 Cisco and/or its affiliates. +# Copyright (c) 2021 Cisco and/or its affiliates. # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at: @@ -16,41 +16,45 @@ Download all data. """ import re -import requests import logging +import gzip from os import rename, mkdir from os.path import join +from http.client import responses, HTTPException from zipfile import ZipFile, is_zipfile, BadZipfile -from httplib import responses -from requests.adapters import HTTPAdapter -from requests.packages.urllib3.util.retry import Retry -from requests import codes, RequestException, Timeout, TooManyRedirects, \ - HTTPError, ConnectionError -from errors import PresentationError +import requests + +from requests.adapters import HTTPAdapter, Retry +from requests.exceptions import RequestException +from requests import codes + +from urllib3.exceptions import HTTPError # Chunk size used for file download CHUNK_SIZE = 512 # Separator used in file names -SEPARATOR = "__" +SEPARATOR = u"__" REGEX_RELEASE = re.compile(r'(\D*)(\d{4}|master)(\D*)') -def _download_file(url, file_name, log, arch=False): +def _download_file(url, file_name, arch=False, verify=True, repeat=1): """Download a file with input data. :param url: URL to the file to download. :param file_name: Name of file to download. - :param log: List of log messages. - :param arch: If True, also .gz file is downloaded + :param arch: If True, also .gz file is downloaded. + :param verify: If true, verify the certificate. + :param repeat: The number of attempts to download the file. :type url: str :type file_name: str - :type log: list of tuples (severity, msg) :type arch: bool + :type verify: bool + :type repeat: int :returns: True if the download was successful, otherwise False. :rtype: bool """ @@ -81,254 +85,228 @@ def _download_file(url, file_name, log, arch=False): ) adapter = HTTPAdapter(max_retries=retry) session = requests.Session() - session.mount('http://', adapter) - session.mount('https://', adapter) + session.mount(u"http://", adapter) + session.mount(u"https://", adapter) return session success = False - try: - log.append(("INFO", " Connecting to '{0}' ...".format(url))) - session = requests_retry_session() - response = session.get(url, stream=True) - code = response.status_code - log.append(("INFO", " {0}: {1}".format(code, responses[code]))) - - if code != codes["OK"]: - if session: - session.close() - url = url.replace("_info", "") - log.append(("INFO", " Connecting to '{0}' ...".format(url))) + while repeat: + repeat -= 1 + session = None + try: + logging.info(f" Connecting to {url} ...") session = requests_retry_session() - response = session.get(url, stream=True) + response = session.get(url, stream=True, verify=verify) code = response.status_code - log.append(("INFO", " {0}: {1}".format(code, responses[code]))) - if code != codes["OK"]: - return False, file_name - file_name = file_name.replace("_info", "") + logging.info(f" {code}: {responses[code]}") - dst_file_name = file_name.replace(".gz", "") - log.append(("INFO", " Downloading the file '{0}' to '{1}' ...". - format(url, dst_file_name))) - with open(dst_file_name, "wb") as file_handle: - for chunk in response.iter_content(chunk_size=CHUNK_SIZE): - if chunk: - file_handle.write(chunk) + if code != codes[u"OK"]: + if session: + session.close() + return False, file_name - if arch and ".gz" in file_name: + dst_file_name = file_name.replace(u".gz", u"") + logging.info(f" Downloading the file {url} to {dst_file_name}") + with open(dst_file_name, u"wb") as file_handle: + for chunk in response.iter_content(chunk_size=CHUNK_SIZE): + if chunk: + file_handle.write(chunk) + + if arch and u".gz" in file_name: + if session: + session.close() + logging.info(f" Downloading the file {url} to {file_name}") + session = requests_retry_session() + response = session.get(url, stream=True, verify=verify) + if response.status_code == codes[u"OK"]: + with open(file_name, u"wb") as file_handle: + file_handle.write(response.raw.read()) + else: + logging.error( + f"Not possible to download the file " + f"{url} to {file_name}" + ) + + success = True + repeat = 0 + except (HTTPException, HTTPError) as err: + logging.error(f"Connection broken:\n{repr(err)}") + except RequestException as err: + logging.error(f"HTTP Request exception:\n{repr(err)}") + except (IOError, ValueError, KeyError) as err: + logging.error(f"Download failed.\n{repr(err)}") + finally: if session: session.close() - log.append(("INFO", " Downloading the file '{0}' to '{1}' ...". - format(url, file_name))) - session = requests_retry_session() - response = session.get(url, stream=True) - if response.status_code == codes["OK"]: - with open(file_name, "wb") as file_handle: - file_handle.write(response.raw.read()) - else: - log.append(("ERROR", "Not possible to download the file '{0}' " - "to '{1}' ...".format(url, file_name))) - - success = True - except ConnectionError as err: - log.append(("ERROR", "Not possible to connect to '{0}'.".format(url))) - log.append(("DEBUG", repr(err))) - except HTTPError as err: - log.append(("ERROR", "Invalid HTTP response from '{0}'.".format(url))) - log.append(("DEBUG", repr(err))) - except TooManyRedirects as err: - log.append(("ERROR", "Request exceeded the configured number " - "of maximum re-directions.")) - log.append(("DEBUG", repr(err))) - except Timeout as err: - log.append(("ERROR", "Request timed out.")) - log.append(("DEBUG", repr(err))) - except RequestException as err: - log.append(("ERROR", "Unexpected HTTP request exception.")) - log.append(("DEBUG", repr(err))) - except (IOError, ValueError, KeyError) as err: - log.append(("ERROR", "Download failed.")) - log.append(("DEBUG", repr(err))) - finally: - if session: - session.close() - - log.append(("INFO", " Download finished.")) return success, file_name -def _unzip_file(spec, build, pid, log): +def _unzip_file(spec, build, pid): """Unzip downloaded source file. :param spec: Specification read form the specification file. :param build: Information about the build. - :param log: List of log messages. :type spec: Specification :type build: dict - :type log: list of tuples (severity, msg) :returns: True if the download was successful, otherwise False. :rtype: bool """ - file_name = build["file-name"] - if ".zip" in file_name: - data_file = spec.input["zip-extract"] - else: - data_file = spec.input["extract"] - - directory = spec.environment["paths"]["DIR[WORKING,DATA]"] + file_name = build[u"file-name"] + data_file = "robot-plugin/output.xml" + directory = spec.environment[u"paths"][u"DIR[WORKING,DATA]"] tmp_dir = join(directory, str(pid)) try: mkdir(tmp_dir) except OSError: pass - new_name = "{0}{1}{2}".format(file_name.rsplit('.')[-2], - SEPARATOR, - data_file.split("/")[-1]) + new_name = \ + f"{file_name.rsplit(u'.')[-2]}{SEPARATOR}{data_file.split(u'/')[-1]}" - log.append(("INFO", " Unzipping: '{0}' from '{1}'.". - format(data_file, file_name))) + logging.info(f" Unzipping: {data_file} from {file_name}.") try: - with ZipFile(file_name, 'r') as zip_file: + with ZipFile(file_name, u'r') as zip_file: zip_file.extract(data_file, tmp_dir) - log.append(("INFO", " Renaming the file '{0}' to '{1}'". - format(join(tmp_dir, data_file), new_name))) + logging.info( + f" Renaming the file {join(tmp_dir, data_file)} to {new_name}" + ) rename(join(tmp_dir, data_file), new_name) - build["file-name"] = new_name + build[u"file-name"] = new_name return True except (BadZipfile, RuntimeError) as err: - log.append(("ERROR", "Failed to unzip the file '{0}': {1}.". - format(file_name, str(err)))) + logging.error(f"Failed to unzip the file {file_name}: {repr(err)}.") return False except OSError as err: - log.append(("ERROR", "Failed to rename the file '{0}': {1}.". - format(data_file, str(err)))) + logging.error(f"Failed to rename the file {data_file}: {repr(err)}.") return False -def download_and_unzip_data_file(spec, job, build, pid, log): +def _download_xml(source, job, build, w_dir, arch): + """ + + :param source: + :param job: + :param build: + :param w_dir: Path to working directory + :param arch: + :return: + """ + + file_name = source.get(u"file-name", u"") + new_name = join( + w_dir, + f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}" + ) + url = u"{0}/{1}".format( + source.get(u"url", u""), + source.get(u"path", u"").format( + job=job, build=build[u'build'], filename=file_name + ) + ) + logging.info(f" Trying to download {url}") + success, downloaded_name = _download_file( + url, new_name, arch=arch, verify=(u"nginx" not in url), repeat=3 + ) + return success, downloaded_name + + +def _download_xml_docs(source, job, build, w_dir, arch): + """ + + :param source: + :param job: + :param build: + :param w_dir: Path to working directory + :param arch: + :return: + """ + + file_name = source.get(u"file-name", u"") + release = re.search(REGEX_RELEASE, job).group(2) + for rls in (release, u"master"): + try: + rls = f"rls{int(rls)}" + except ValueError: + pass # It is master + url = ( + f"{source.get(u'url', u'')}/" + f"{rls}/" + f"{source.get(u'path', u'')}/" + f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}" + ) + new_name = join( + w_dir, + f"{job}{SEPARATOR}{build[u'build']}{SEPARATOR}{file_name}" + ) + + logging.info(f" Trying to download {url}") + + success, downloaded_name = _download_file(url, new_name, arch=arch) + if success: + if file_name.endswith(u".gz"): + with gzip.open(downloaded_name[:-3], u"rb") as gzip_file: + file_content = gzip_file.read() + with open(downloaded_name[:-3], u"wb") as xml_file: + xml_file.write(file_content) + break + + return success, downloaded_name + + +def download_and_unzip_data_file(spec, job, build, pid): """Download and unzip a source file. :param spec: Specification read form the specification file. :param job: Name of the Jenkins job. :param build: Information about the build. :param pid: PID of the process executing this method. - :param log: List of log messages. :type spec: Specification :type job: str :type build: dict :type pid: int - :type log: list of tuples (severity, msg) :returns: True if the download was successful, otherwise False. :rtype: bool """ - # Try to download .gz from logs.fd.io - - file_name = spec.input["file-name"] - url = "{0}/{1}".format( - spec.environment["urls"]["URL[NEXUS,LOG]"], - spec.input["download-path"].format( - job=job, build=build["build"], filename=file_name)) - new_name = join(spec.environment["paths"]["DIR[WORKING,DATA]"], - "{job}{sep}{build}{sep}{name}".format( - job=job, sep=SEPARATOR, build=build["build"], - name=file_name)) - - logging.info("Trying to download {0}".format(url)) - - arch = True if spec.configuration.get("archive-inputs", True) else False - success, downloaded_name = _download_file(url, new_name, log, arch=arch) - - if not success: - - # Try to download .gz from docs.fd.io - - file_name = spec.input["file-name"] - url = "{0}/{1}".format( - spec.environment["urls"]["URL[NEXUS,DOC]"], - spec.input["download-path"].format( - job=job, build=build["build"], filename=file_name)) - new_name = join(spec.environment["paths"]["DIR[WORKING,DATA]"], - "{job}{sep}{build}{sep}{name}".format( - job=job, sep=SEPARATOR, build=build["build"], - name=file_name)) - - logging.info("Downloading {0}".format(url)) - - if spec.configuration.get("archive-inputs", True): - arch = True - success, downloaded_name = _download_file(url, new_name, log, arch=arch) - - if not success: - - # Try to download .zip from docs.fd.io - - file_name = spec.input["zip-file-name"] - new_name = join(spec.environment["paths"]["DIR[WORKING,DATA]"], - "{job}{sep}{build}{sep}{name}".format( - job=job, sep=SEPARATOR, build=build["build"], - name=file_name)) - release = re.search(REGEX_RELEASE, job).group(2) - for rls in (release, "master"): - nexus_file_name = "{job}{sep}{build}{sep}{name}". \ - format(job=job, sep=SEPARATOR, build=build["build"], - name=file_name) - try: - rls = "rls{0}".format(int(rls)) - except ValueError: - # It is 'master' - pass - url = "{url}/{release}/{dir}/{file}". \ - format(url=spec.environment["urls"]["URL[NEXUS,DOC]"], - release=rls, - dir=spec.environment["urls"]["DIR[NEXUS,DOC]"], - file=nexus_file_name) - - logging.info("Downloading {0}".format(url)) - - success, downloaded_name = _download_file(url, new_name, log) - if success: - break - - if not success: - - # Try to download .zip from jenkins.fd.io - - file_name = spec.input["zip-file-name"] - download_path = spec.input["zip-download-path"] - if job.startswith("csit-"): - url = spec.environment["urls"]["URL[JENKINS,CSIT]"] - elif job.startswith("hc2vpp-"): - url = spec.environment["urls"]["URL[JENKINS,HC]"] - else: - raise PresentationError( - "No url defined for the job '{}'.".format(job)) - - full_name = download_path.format( - job=job, build=build["build"], filename=file_name) - url = "{0}/{1}".format(url, full_name) - new_name = join(spec.environment["paths"]["DIR[WORKING,DATA]"], - "{job}{sep}{build}{sep}{name}". - format(job=job, sep=SEPARATOR, build=build["build"], - name=file_name)) - - logging.info("Downloading {0}".format(url)) - - success, downloaded_name = _download_file(url, new_name, log) - - if success and downloaded_name.endswith(".zip"): + download = { + "xml": _download_xml, + "xml-docs": _download_xml_docs + } + + success = False + downloaded_name = u"" + arch = bool(spec.environment.get(u"archive-inputs", True)) + + for source in spec.environment.get(u"data-sources", tuple()): + if not source.get(u"enabled", False): + continue + download_type = source.get(u"type", None) + if not download_type: + continue + success, downloaded_name = download[download_type]( + source, + job, + build, + spec.environment[u"paths"][u"DIR[WORKING,DATA]"], + arch + ) + if success: + source[u"successful-downloads"] += 1 + build[u"source"] = source[u"type"] + break + + # TODO: Remove when only .gz is used. + if success and downloaded_name.endswith(u".zip"): if not is_zipfile(downloaded_name): - log.append(("ERROR", - "Zip file '{0}' is corrupted.".format(new_name))) + logging.error(f"Zip file {downloaded_name} is corrupted.") success = False if success: - build["file-name"] = downloaded_name - - if file_name.endswith(".gz"): - build["file-name"] = downloaded_name[:-3] - - if downloaded_name.endswith(".zip"): - success = _unzip_file(spec, build, pid, log) + if downloaded_name.endswith(u".gz"): + build[u"file-name"] = downloaded_name[:-3] + # TODO: Remove when only .gz is used. + elif downloaded_name.endswith(u".zip"): + build[u"file-name"] = downloaded_name + success = _unzip_file(spec, build, pid) return success