)
REGEX_VERSION_VPP = re.compile(
- r"(return STDOUT Version:\s*|"
- r"VPP Version:\s*|VPP version:\s*)(.*)"
+ r"(VPP Version:\s*|VPP version:\s*)(.*)"
)
REGEX_VERSION_DPDK = re.compile(
r"(DPDK version:\s*|DPDK Version:\s*)(.*)"
:returns: Nothing.
"""
- if msg.message.count(u"return STDOUT Version:") or \
- msg.message.count(u"VPP Version:") or \
- msg.message.count(u"VPP version:"):
+ if msg.message.count(u"VPP version:") or \
+ msg.message.count(u"VPP Version:"):
self._version = str(
re.search(self.REGEX_VERSION_VPP, msg.message).group(2)
)
self._data[u"metadata"][u"version"] = self._version
self._msg_type = None
+ logging.info(self._version)
def _get_dpdk_version(self, msg):
"""Called when extraction of DPDK version is required.
:type setup_kw: Keyword
:returns: Nothing.
"""
- for keyword in setup_kw.body:
+ for keyword in setup_kw.setup:
if self.start_setup_kw(keyword) is not False:
self.visit_setup_kw(keyword)
self.end_setup_kw(keyword)
self._for_output = for_output
# Data store:
- self._input_data = pd.Series(dtype="object")
+ self._input_data = pd.Series(dtype="float64")
@property
def data(self):
result.visit(checker)
checker.data[u"metadata"][u"tests_total"] = \
- result.statistics.total.all.total
+ result.statistics.total.total
checker.data[u"metadata"][u"tests_passed"] = \
- result.statistics.total.all.passed
+ result.statistics.total.passed
checker.data[u"metadata"][u"tests_failed"] = \
- result.statistics.total.all.failed
+ result.statistics.total.failed
checker.data[u"metadata"][u"elapsedtime"] = result.suite.elapsedtime
checker.data[u"metadata"][u"generated"] = result.suite.endtime[:14]
})
if self._input_data.get(job, None) is None:
- self._input_data[job] = pd.Series(dtype="object")
+ self._input_data[job] = pd.Series(dtype="float64")
self._input_data[job][str(build_nr)] = build_data
self._cfg.set_input_file_name(
job, build_nr, result[u"build"][u"file-name"]
})
if self._input_data.get(job, None) is None:
- self._input_data[job] = pd.Series(dtype="object")
+ self._input_data[job] = pd.Series(dtype="float64")
self._input_data[job][str(build_nr)] = build_data
self._cfg.set_input_state(job, build_nr, u"processed")
params.extend((u"type", u"status"))
data_to_filter = data if data else element[u"data"]
- data = pd.Series(dtype="object")
+ data = pd.Series(dtype="float64")
try:
for job, builds in data_to_filter.items():
- data[job] = pd.Series(dtype="object")
+ data[job] = pd.Series(dtype="float64")
for build in builds:
- data[job][str(build)] = pd.Series(dtype="object")
+ data[job][str(build)] = pd.Series(dtype="float64")
try:
data_dict = dict(
self.data[job][str(build)][data_set].items())
for test_id, test_data in data_dict.items():
if eval(cond, {u"tags": test_data.get(u"tags", u"")}):
data[job][str(build)][test_id] = \
- pd.Series(dtype="object")
+ pd.Series(dtype="float64")
if params is None:
for param, val in test_data.items():
data[job][str(build)][test_id][param] = val
else:
tests = include
- data = pd.Series(dtype="object")
+ data = pd.Series(dtype="float64")
try:
for job, builds in element[u"data"].items():
- data[job] = pd.Series(dtype="object")
+ data[job] = pd.Series(dtype="float64")
for build in builds:
- data[job][str(build)] = pd.Series(dtype="object")
+ data[job][str(build)] = pd.Series(dtype="float64")
for test in tests:
try:
reg_ex = re.compile(str(test).lower())
test_data = self.data[job][
str(build)][data_set][test_id]
data[job][str(build)][test_id] = \
- pd.Series(dtype="object")
+ pd.Series(dtype="float64")
if params is None:
for param, val in test_data.items():
data[job][str(build)][test_id]\
logging.info(u" Merging data ...")
- merged_data = pd.Series(dtype="object")
+ merged_data = pd.Series(dtype="float64")
for builds in data.values:
for item in builds.values:
for item_id, item_data in item.items():