Code Review
/
csit.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
review
|
tree
raw
|
inline
| side by side
Trending: Add multiprocessing, remove archiving
[csit.git]
/
resources
/
tools
/
presentation
/
input_data_parser.py
diff --git
a/resources/tools/presentation/input_data_parser.py
b/resources/tools/presentation/input_data_parser.py
index
84245c4
..
9c0e380
100644
(file)
--- a/
resources/tools/presentation/input_data_parser.py
+++ b/
resources/tools/presentation/input_data_parser.py
@@
-22,6
+22,7
@@
import multiprocessing
import os
import re
import multiprocessing
import os
import re
+import resource
import pandas as pd
import logging
import pandas as pd
import logging
@@
-1266,11
+1267,11
@@
class InputData(object):
file_name = self._cfg.input["file-name"]
full_name = join(
self._cfg.environment["paths"]["DIR[WORKING,DATA]"],
file_name = self._cfg.input["file-name"]
full_name = join(
self._cfg.environment["paths"]["DIR[WORKING,DATA]"],
- "{job}{sep}{build}{sep}{name}".
-
format(
job=job,
-
sep=SEPARATOR,
-
build=build["build"],
-
name=file_name))
+ "{job}{sep}{build}{sep}{name}".
format(
+ job=job,
+ sep=SEPARATOR,
+ build=build["build"],
+ name=file_name))
try:
remove(full_name)
logs.append(("INFO",
try:
remove(full_name)
logs.append(("INFO",
@@
-1278,10
+1279,18
@@
class InputData(object):
format(name=full_name)))
except OSError as err:
logs.append(("ERROR",
format(name=full_name)))
except OSError as err:
logs.append(("ERROR",
- "Cannot remove the file '{0}': {1}".
- format(full_name, repr(err))))
+
"Cannot remove the file '{0}': {1}".
+
format(full_name, repr(err))))
logs.append(("INFO", " Done."))
logs.append(("INFO", " Done."))
+ result = {
+ "data": data,
+ "state": state,
+ "job": job,
+ "build": build
+ }
+ data_queue.put(result)
+
for level, line in logs:
if level == "INFO":
logging.info(line)
for level, line in logs:
if level == "INFO":
logging.info(line)
@@
-1294,13
+1303,8
@@
class InputData(object):
elif level == "WARNING":
logging.warning(line)
elif level == "WARNING":
logging.warning(line)
- result = {
- "data": data,
- "state": state,
- "job": job,
- "build": build
- }
- data_queue.put(result)
+ logging.info("Memory allocation: {0:,d}MB".format(
+ resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1000))
def download_and_parse_data(self, repeat=1):
"""Download the input data files, parse input data from input files and
def download_and_parse_data(self, repeat=1):
"""Download the input data files, parse input data from input files and
@@
-1336,18
+1340,21
@@
class InputData(object):
work_queue.join()
logging.info("Done.")
work_queue.join()
logging.info("Done.")
+ logging.info("Collecting data:")
while not data_queue.empty():
result = data_queue.get()
job = result["job"]
build_nr = result["build"]["build"]
while not data_queue.empty():
result = data_queue.get()
job = result["job"]
build_nr = result["build"]["build"]
+ logging.info(" {job}-{build}".format(job=job, build=build_nr))
if result["data"]:
data = result["data"]
build_data = pd.Series({
if result["data"]:
data = result["data"]
build_data = pd.Series({
- "metadata": pd.Series(data["metadata"].values(),
- index=data["metadata"].keys()),
+ "metadata": pd.Series(
+ data["metadata"].values(),
+ index=data["metadata"].keys()),
"suites": pd.Series(data["suites"].values(),
index=data["suites"].keys()),
"tests": pd.Series(data["tests"].values(),
"suites": pd.Series(data["suites"].values(),
index=data["suites"].keys()),
"tests": pd.Series(data["tests"].values(),
@@
-1357,11
+1364,14
@@
class InputData(object):
self._input_data[job] = pd.Series()
self._input_data[job][str(build_nr)] = build_data
self._input_data[job] = pd.Series()
self._input_data[job][str(build_nr)] = build_data
- self._cfg.set_input_file_name(
job, build_nr,
-
result["build"]["file-name"])
+ self._cfg.set_input_file_name(
+
job, build_nr,
result["build"]["file-name"])
self._cfg.set_input_state(job, build_nr, result["state"])
self._cfg.set_input_state(job, build_nr, result["state"])
+ logging.info("Memory allocation: {0:,d}MB".format(
+ resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1000))
+
del data_queue
# Terminate all workers
del data_queue
# Terminate all workers