import multiprocessing
import os
import re
+import resource
import pandas as pd
import logging
:type msg: Message
:returns: Nothing.
"""
- if msg.message.count("return STDOUT Thread "):
+ if msg.message.count("Thread 0 vpp_main"):
self._show_run_lookup_nr += 1
if self._lookup_kw_nr == 1 and self._show_run_lookup_nr == 1:
self._data["tests"][self._test_ID]["show-run"] = str()
if setup_kw.name.count("Show Vpp Version On All Duts") \
and not self._version:
self._msg_type = "vpp-version"
-
- elif setup_kw.name.count("Setup performance global Variables") \
+ elif setup_kw.name.count("Set Global Variable") \
and not self._timestamp:
self._msg_type = "timestamp"
elif setup_kw.name.count("Setup Framework") and not self._testbed:
logs = list()
- logging.info(" Processing the job/build: {0}: {1}".
- format(job, build["build"]))
-
logs.append(("INFO", " Processing the job/build: {0}: {1}".
format(job, build["build"])))
"'{build}', or it is damaged. Skipped.".
format(job=job, build=build["build"])))
if success:
- logs.append(("INFO", " Processing data from the build '{0}' ...".
+ logs.append(("INFO", " Processing data from the build '{0}' ...".
format(build["build"])))
data = self._parse_tests(job, build, logs)
if data is None:
file_name = self._cfg.input["file-name"]
full_name = join(
self._cfg.environment["paths"]["DIR[WORKING,DATA]"],
- "{job}{sep}{build}{sep}{name}".
- format(job=job,
- sep=SEPARATOR,
- build=build["build"],
- name=file_name))
+ "{job}{sep}{build}{sep}{name}".format(
+ job=job,
+ sep=SEPARATOR,
+ build=build["build"],
+ name=file_name))
try:
remove(full_name)
logs.append(("INFO",
format(name=full_name)))
except OSError as err:
logs.append(("ERROR",
- "Cannot remove the file '{0}': {1}".
- format(full_name, repr(err))))
-
+ "Cannot remove the file '{0}': {1}".
+ format(full_name, repr(err))))
logs.append(("INFO", " Done."))
result = {
"data": data,
"state": state,
"job": job,
- "build": build,
- "logs": logs
+ "build": build
}
data_queue.put(result)
+ for level, line in logs:
+ if level == "INFO":
+ logging.info(line)
+ elif level == "ERROR":
+ logging.error(line)
+ elif level == "DEBUG":
+ logging.debug(line)
+ elif level == "CRITICAL":
+ logging.critical(line)
+ elif level == "WARNING":
+ logging.warning(line)
+
+ logging.info("Memory allocation: {0:,d}MB".format(
+ resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1000))
+
def download_and_parse_data(self, repeat=1):
"""Download the input data files, parse input data from input files and
store in pandas' Series.
work_queue.join()
logging.info("Done.")
+ logging.info("Collecting data:")
while not data_queue.empty():
result = data_queue.get()
job = result["job"]
build_nr = result["build"]["build"]
+ logging.info(" {job}-{build}".format(job=job, build=build_nr))
if result["data"]:
data = result["data"]
build_data = pd.Series({
- "metadata": pd.Series(data["metadata"].values(),
- index=data["metadata"].keys()),
+ "metadata": pd.Series(
+ data["metadata"].values(),
+ index=data["metadata"].keys()),
"suites": pd.Series(data["suites"].values(),
index=data["suites"].keys()),
"tests": pd.Series(data["tests"].values(),
self._input_data[job] = pd.Series()
self._input_data[job][str(build_nr)] = build_data
- self._cfg.set_input_file_name(job, build_nr,
- result["build"]["file-name"])
+ self._cfg.set_input_file_name(
+ job, build_nr, result["build"]["file-name"])
self._cfg.set_input_state(job, build_nr, result["state"])
- for item in result["logs"]:
- if item[0] == "INFO":
- logging.info(item[1])
- elif item[0] == "ERROR":
- logging.error(item[1])
- elif item[0] == "DEBUG":
- logging.debug(item[1])
- elif item[0] == "CRITICAL":
- logging.critical(item[1])
- elif item[0] == "WARNING":
- logging.warning(item[1])
+ logging.info("Memory allocation: {0:,d}MB".format(
+ resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1000))
del data_queue