-# Copyright (c) 2021 Cisco and/or its affiliates.
+# Copyright (c) 2022 Cisco and/or its affiliates.
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at:
r'hostname=\"(\d{1,3}.\d{1,3}.\d{1,3}.\d{1,3})\",hook=\"(.*)\"'
)
- def __init__(self, metadata, mapping, ignore, for_output):
+ def __init__(self, metadata, mapping, ignore, process_oper):
"""Initialisation.
:param metadata: Key-value pairs to be included in "metadata" part of
:param mapping: Mapping of the old names of test cases to the new
(actual) one.
:param ignore: List of TCs to be ignored.
- :param for_output: Output to be generated from downloaded data.
+ :param process_oper: If True, operational data (show run, telemetry) is
+ processed.
:type metadata: dict
:type mapping: dict
:type ignore: list
- :type for_output: str
+ :type process_oper: bool
"""
# Type of message to parse out from the test messages
# Ignore list
self._ignore = ignore
- self._for_output = for_output
+ self._process_oper = process_oper
# Number of PAPI History messages found:
# 0 - no message
)
if all(hdr_lat):
return hdr_lat
- else:
- hdr_lat = (
- in_list_1[0], in_list_1[1], in_list_1[2],
- in_list_2[0], in_list_2[1], in_list_2[2]
- )
- for item in hdr_lat:
- if item in (u"-1", u"4294967295", u"0"):
- return None
- return hdr_lat
- return None
+ hdr_lat = (
+ int(in_list_1[0]), int(in_list_1[1]), int(in_list_1[2]),
+ int(in_list_2[0]), int(in_list_2[1]), int(in_list_2[2])
+ )
+ for item in hdr_lat:
+ if item in (-1, 4294967295, 0):
+ return None
+ return hdr_lat
try:
out_msg = (
u"level": len(suite.longname.split(u"."))
}
- suite.keywords.visit(self)
+ suite.setup.visit(self)
def end_suite(self, suite):
"""Called when suite ends.
:returns: Nothing.
"""
if self.start_test(test) is not False:
- test.keywords.visit(self)
+ test.body.visit(self)
self.end_test(test)
def start_test(self, test):
if test.status == u"PASS":
test_result[u"throughput"], test_result[u"status"] = \
self._get_plr_throughput(test.message)
+ elif u"LDP_NGINX" in tags:
+ test_result[u"type"] = u"LDP_NGINX"
+ test_result[u"result"], test_result[u"status"] = \
+ self._get_vsap_data(test.message, tags)
elif u"HOSTSTACK" in tags:
test_result[u"type"] = u"HOSTSTACK"
if test.status == u"PASS":
test_result[u"result"], test_result[u"status"] = \
self._get_hoststack_data(test.message, tags)
- elif u"LDP_NGINX" in tags:
- test_result[u"type"] = u"LDP_NGINX"
- test_result[u"result"], test_result[u"status"] = \
- self._get_vsap_data(test.message, tags)
- # elif u"TCP" in tags: # This might be not used
- # test_result[u"type"] = u"TCP"
- # if test.status == u"PASS":
- # groups = re.search(self.REGEX_TCP, test.message)
- # test_result[u"result"] = int(groups.group(2))
elif u"RECONF" in tags:
test_result[u"type"] = u"RECONF"
if test.status == u"PASS":
:type test_kw: Keyword
:returns: Nothing.
"""
- for keyword in test_kw.keywords:
+ for keyword in test_kw.body:
if self.start_test_kw(keyword) is not False:
self.visit_test_kw(keyword)
self.end_test_kw(keyword)
:type test_kw: Keyword
:returns: Nothing.
"""
- if self._for_output == u"trending":
+ if not self._process_oper:
return
if test_kw.name.count(u"Run Telemetry On All Duts"):
:type setup_kw: Keyword
:returns: Nothing.
"""
- for keyword in setup_kw.keywords:
+ for keyword in setup_kw.body:
if self.start_setup_kw(keyword) is not False:
self.visit_setup_kw(keyword)
self.end_setup_kw(keyword)
:type teardown_kw: Keyword
:returns: Nothing.
"""
- for keyword in teardown_kw.keywords:
+ for keyword in teardown_kw.body:
if self.start_teardown_kw(keyword) is not False:
self.visit_teardown_kw(keyword)
self.end_teardown_kw(keyword)
self._for_output = for_output
# Data store:
- self._input_data = pd.Series()
+ self._input_data = pd.Series(dtype="object")
@property
def data(self):
f"Error occurred while parsing output.xml: {repr(err)}"
)
return None
+
+ process_oper = False
+ if u"-vpp-perf-report-coverage-" in job:
+ process_oper = True
+ # elif u"-vpp-perf-report-iterative-" in job:
+ # # Exceptions for TBs where we do not have coverage data:
+ # for item in (u"-2n-icx", ):
+ # if item in job:
+ # process_oper = True
+ # break
checker = ExecutionChecker(
- metadata, self._cfg.mapping, self._cfg.ignore, self._for_output
+ metadata, self._cfg.mapping, self._cfg.ignore, process_oper
)
result.visit(checker)
})
if self._input_data.get(job, None) is None:
- self._input_data[job] = pd.Series()
+ self._input_data[job] = pd.Series(dtype="object")
self._input_data[job][str(build_nr)] = build_data
self._cfg.set_input_file_name(
job, build_nr, result[u"build"][u"file-name"]
})
if self._input_data.get(job, None) is None:
- self._input_data[job] = pd.Series()
+ self._input_data[job] = pd.Series(dtype="object")
self._input_data[job][str(build_nr)] = build_data
self._cfg.set_input_state(job, build_nr, u"processed")
params.extend((u"type", u"status"))
data_to_filter = data if data else element[u"data"]
- data = pd.Series()
+ data = pd.Series(dtype="object")
try:
for job, builds in data_to_filter.items():
- data[job] = pd.Series()
+ data[job] = pd.Series(dtype="object")
for build in builds:
- data[job][str(build)] = pd.Series()
+ data[job][str(build)] = pd.Series(dtype="object")
try:
data_dict = dict(
self.data[job][str(build)][data_set].items())
for test_id, test_data in data_dict.items():
if eval(cond, {u"tags": test_data.get(u"tags", u"")}):
- data[job][str(build)][test_id] = pd.Series()
+ data[job][str(build)][test_id] = \
+ pd.Series(dtype="object")
if params is None:
for param, val in test_data.items():
data[job][str(build)][test_id][param] = val
else:
tests = include
- data = pd.Series()
+ data = pd.Series(dtype="object")
try:
for job, builds in element[u"data"].items():
- data[job] = pd.Series()
+ data[job] = pd.Series(dtype="object")
for build in builds:
- data[job][str(build)] = pd.Series()
+ data[job][str(build)] = pd.Series(dtype="object")
for test in tests:
try:
reg_ex = re.compile(str(test).lower())
if re.match(reg_ex, str(test_id).lower()):
test_data = self.data[job][
str(build)][data_set][test_id]
- data[job][str(build)][test_id] = pd.Series()
+ data[job][str(build)][test_id] = \
+ pd.Series(dtype="object")
if params is None:
for param, val in test_data.items():
data[job][str(build)][test_id]\
logging.info(u" Merging data ...")
- merged_data = pd.Series()
+ merged_data = pd.Series(dtype="object")
for builds in data.values:
for item in builds.values:
for item_id, item_data in item.items():