X-Git-Url: https://gerrit.fd.io/r/gitweb?p=csit.git;a=blobdiff_plain;f=resources%2Ftools%2Fpresentation%2Finput_data_parser.py;h=9c0e38073ca2f49e190f519e4b837c3a0d90c7a2;hp=20974f1f05f141827b14812fe1e885b3f4fa5594;hb=cee3ad0f9cc29ffc67d9c87c58920252671beb21;hpb=f2af716be043a2bb17a2580ed212200515da9ad0 diff --git a/resources/tools/presentation/input_data_parser.py b/resources/tools/presentation/input_data_parser.py index 20974f1f05..9c0e38073c 100644 --- a/resources/tools/presentation/input_data_parser.py +++ b/resources/tools/presentation/input_data_parser.py @@ -22,6 +22,7 @@ import multiprocessing import os import re +import resource import pandas as pd import logging @@ -1266,11 +1267,11 @@ class InputData(object): file_name = self._cfg.input["file-name"] full_name = join( self._cfg.environment["paths"]["DIR[WORKING,DATA]"], - "{job}{sep}{build}{sep}{name}". - format(job=job, - sep=SEPARATOR, - build=build["build"], - name=file_name)) + "{job}{sep}{build}{sep}{name}".format( + job=job, + sep=SEPARATOR, + build=build["build"], + name=file_name)) try: remove(full_name) logs.append(("INFO", @@ -1278,10 +1279,18 @@ class InputData(object): format(name=full_name))) except OSError as err: logs.append(("ERROR", - "Cannot remove the file '{0}': {1}". - format(full_name, repr(err)))) + "Cannot remove the file '{0}': {1}". + format(full_name, repr(err)))) logs.append(("INFO", " Done.")) + result = { + "data": data, + "state": state, + "job": job, + "build": build + } + data_queue.put(result) + for level, line in logs: if level == "INFO": logging.info(line) @@ -1294,13 +1303,8 @@ class InputData(object): elif level == "WARNING": logging.warning(line) - result = { - "data": data, - "state": state, - "job": job, - "build": build - } - data_queue.put(result) + logging.info("Memory allocation: {0:,d}MB".format( + resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1000)) def download_and_parse_data(self, repeat=1): """Download the input data files, parse input data from input files and @@ -1333,44 +1337,40 @@ class InputData(object): for build in builds: work_queue.put((job, build, repeat)) - work_queue.join() + work_queue.join() - logging.info("Done.") + logging.info("Done.") + logging.info("Collecting data:") - logging.info("Collecting data:") + while not data_queue.empty(): + result = data_queue.get() - while not data_queue.empty(): - try: - result = data_queue.get() - - job = result["job"] - build_nr = result["build"]["build"] - - logging.info(" {job}-{build}".format(job=job, - build=build_nr)) - if result["data"]: - data = result["data"] - build_data = pd.Series({ - "metadata": pd.Series( - data["metadata"].values(), - index=data["metadata"].keys()), - "suites": pd.Series(data["suites"].values(), - index=data["suites"].keys()), - "tests": pd.Series(data["tests"].values(), - index=data["tests"].keys())}) - - if self._input_data.get(job, None) is None: - self._input_data[job] = pd.Series() - self._input_data[job][str(build_nr)] = build_data - - self._cfg.set_input_file_name( - job, build_nr, result["build"]["file-name"]) - - self._cfg.set_input_state(job, build_nr, result["state"]) - - except (MemoryError, EOFError) as err: - logging.error(repr(err)) - raise + job = result["job"] + build_nr = result["build"]["build"] + logging.info(" {job}-{build}".format(job=job, build=build_nr)) + + if result["data"]: + data = result["data"] + build_data = pd.Series({ + "metadata": pd.Series( + data["metadata"].values(), + index=data["metadata"].keys()), + "suites": pd.Series(data["suites"].values(), + index=data["suites"].keys()), + "tests": pd.Series(data["tests"].values(), + index=data["tests"].keys())}) + + if self._input_data.get(job, None) is None: + self._input_data[job] = pd.Series() + self._input_data[job][str(build_nr)] = build_data + + self._cfg.set_input_file_name( + job, build_nr, result["build"]["file-name"]) + + self._cfg.set_input_state(job, build_nr, result["state"]) + + logging.info("Memory allocation: {0:,d}MB".format( + resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1000)) del data_queue