+ logging.warning(u"The output file is not defined.")
+ return
+
+
+def table_comparison(table, input_data):
+ """Generate the table(s) with algorithm: table_comparison
+ specified in the specification file.
+
+ :param table: Table to generate.
+ :param input_data: Data to process.
+ :type table: pandas.Series
+ :type input_data: InputData
+ """
+ logging.info(f" Generating the table {table.get(u'title', u'')} ...")
+
+ # Transform the data
+ logging.info(
+ f" Creating the data set for the {table.get(u'type', u'')} "
+ f"{table.get(u'title', u'')}."
+ )
+
+ columns = table.get(u"columns", None)
+ if not columns:
+ logging.error(
+ f"No columns specified for {table.get(u'title', u'')}. Skipping."
+ )
+ return
+
+ cols = list()
+ for idx, col in enumerate(columns):
+ if col.get(u"data-set", None) is None:
+ logging.warning(f"No data for column {col.get(u'title', u'')}")
+ continue
+ tag = col.get(u"tag", None)
+ data = input_data.filter_data(
+ table,
+ params=[
+ u"throughput",
+ u"result",
+ u"latency",
+ u"name",
+ u"parent",
+ u"tags"
+ ],
+ data=col[u"data-set"],
+ continue_on_error=True
+ )
+ col_data = {
+ u"title": col.get(u"title", f"Column{idx}"),
+ u"data": dict()
+ }
+ for builds in data.values:
+ for build in builds:
+ for tst_name, tst_data in build.items():
+ if tag and tag not in tst_data[u"tags"]:
+ continue
+ tst_name_mod = \
+ _tpc_modify_test_name(tst_name, ignore_nic=True).\
+ replace(u"2n1l-", u"")
+ if col_data[u"data"].get(tst_name_mod, None) is None:
+ name = tst_data[u'name'].rsplit(u'-', 1)[0]
+ if u"across testbeds" in table[u"title"].lower() or \
+ u"across topologies" in table[u"title"].lower():
+ name = _tpc_modify_displayed_test_name(name)
+ col_data[u"data"][tst_name_mod] = {
+ u"name": name,
+ u"replace": True,
+ u"data": list(),
+ u"mean": None,
+ u"stdev": None
+ }
+ _tpc_insert_data(
+ target=col_data[u"data"][tst_name_mod],
+ src=tst_data,
+ include_tests=table[u"include-tests"]
+ )
+
+ replacement = col.get(u"data-replacement", None)
+ if replacement:
+ rpl_data = input_data.filter_data(
+ table,
+ params=[
+ u"throughput",
+ u"result",
+ u"latency",
+ u"name",
+ u"parent",
+ u"tags"
+ ],
+ data=replacement,
+ continue_on_error=True
+ )
+ for builds in rpl_data.values:
+ for build in builds:
+ for tst_name, tst_data in build.items():
+ if tag and tag not in tst_data[u"tags"]:
+ continue
+ tst_name_mod = \
+ _tpc_modify_test_name(tst_name, ignore_nic=True).\
+ replace(u"2n1l-", u"")
+ if col_data[u"data"].get(tst_name_mod, None) is None:
+ name = tst_data[u'name'].rsplit(u'-', 1)[0]
+ if u"across testbeds" in table[u"title"].lower() \
+ or u"across topologies" in \
+ table[u"title"].lower():
+ name = _tpc_modify_displayed_test_name(name)
+ col_data[u"data"][tst_name_mod] = {
+ u"name": name,
+ u"replace": False,
+ u"data": list(),
+ u"mean": None,
+ u"stdev": None
+ }
+ if col_data[u"data"][tst_name_mod][u"replace"]:
+ col_data[u"data"][tst_name_mod][u"replace"] = False
+ col_data[u"data"][tst_name_mod][u"data"] = list()
+ _tpc_insert_data(
+ target=col_data[u"data"][tst_name_mod],
+ src=tst_data,
+ include_tests=table[u"include-tests"]
+ )
+
+ if table[u"include-tests"] in (u"NDR", u"PDR") or \
+ u"latency" in table[u"include-tests"]:
+ for tst_name, tst_data in col_data[u"data"].items():
+ if tst_data[u"data"]:
+ tst_data[u"mean"] = mean(tst_data[u"data"])
+ tst_data[u"stdev"] = stdev(tst_data[u"data"])
+
+ cols.append(col_data)
+
+ tbl_dict = dict()
+ for col in cols:
+ for tst_name, tst_data in col[u"data"].items():
+ if tbl_dict.get(tst_name, None) is None:
+ tbl_dict[tst_name] = {
+ "name": tst_data[u"name"]
+ }
+ tbl_dict[tst_name][col[u"title"]] = {
+ u"mean": tst_data[u"mean"],
+ u"stdev": tst_data[u"stdev"]
+ }
+
+ if not tbl_dict:
+ logging.warning(f"No data for table {table.get(u'title', u'')}!")
+ return
+
+ tbl_lst = list()
+ for tst_data in tbl_dict.values():
+ row = [tst_data[u"name"], ]
+ for col in cols:
+ row.append(tst_data.get(col[u"title"], None))
+ tbl_lst.append(row)
+
+ comparisons = table.get(u"comparisons", None)
+ rcas = list()
+ if comparisons and isinstance(comparisons, list):
+ for idx, comp in enumerate(comparisons):
+ try:
+ col_ref = int(comp[u"reference"])
+ col_cmp = int(comp[u"compare"])
+ except KeyError:
+ logging.warning(u"Comparison: No references defined! Skipping.")
+ comparisons.pop(idx)
+ continue
+ if not (0 < col_ref <= len(cols) and 0 < col_cmp <= len(cols) or
+ col_ref == col_cmp):
+ logging.warning(f"Wrong values of reference={col_ref} "
+ f"and/or compare={col_cmp}. Skipping.")
+ comparisons.pop(idx)
+ continue
+ rca_file_name = comp.get(u"rca-file", None)
+ if rca_file_name:
+ try:
+ with open(rca_file_name, u"r") as file_handler:
+ rcas.append(
+ {
+ u"title": f"RCA{idx + 1}",
+ u"data": load(file_handler, Loader=FullLoader)
+ }
+ )
+ except (YAMLError, IOError) as err:
+ logging.warning(
+ f"The RCA file {rca_file_name} does not exist or "
+ f"it is corrupted!"
+ )
+ logging.debug(repr(err))
+ rcas.append(None)
+ else:
+ rcas.append(None)
+ else:
+ comparisons = None
+
+ tbl_cmp_lst = list()
+ if comparisons:
+ for row in tbl_lst:
+ new_row = deepcopy(row)
+ for comp in comparisons:
+ ref_itm = row[int(comp[u"reference"])]
+ if ref_itm is None and \
+ comp.get(u"reference-alt", None) is not None:
+ ref_itm = row[int(comp[u"reference-alt"])]
+ cmp_itm = row[int(comp[u"compare"])]
+ if ref_itm is not None and cmp_itm is not None and \
+ ref_itm[u"mean"] is not None and \
+ cmp_itm[u"mean"] is not None and \
+ ref_itm[u"stdev"] is not None and \
+ cmp_itm[u"stdev"] is not None:
+ try:
+ delta, d_stdev = relative_change_stdev(
+ ref_itm[u"mean"], cmp_itm[u"mean"],
+ ref_itm[u"stdev"], cmp_itm[u"stdev"]
+ )
+ except ZeroDivisionError:
+ break
+ if delta is None or math.isnan(delta):
+ break
+ new_row.append({
+ u"mean": delta * 1e6,
+ u"stdev": d_stdev * 1e6
+ })
+ else:
+ break
+ else:
+ tbl_cmp_lst.append(new_row)
+
+ try:
+ tbl_cmp_lst.sort(key=lambda rel: rel[0], reverse=False)
+ tbl_cmp_lst.sort(key=lambda rel: rel[-1][u'mean'], reverse=True)
+ except TypeError as err:
+ logging.warning(f"Empty data element in table\n{tbl_cmp_lst}\n{err}")
+
+ tbl_for_csv = list()
+ for line in tbl_cmp_lst:
+ row = [line[0], ]
+ for idx, itm in enumerate(line[1:]):
+ if itm is None or not isinstance(itm, dict) or\
+ itm.get(u'mean', None) is None or \
+ itm.get(u'stdev', None) is None:
+ row.append(u"NT")
+ row.append(u"NT")
+ else:
+ row.append(round(float(itm[u'mean']) / 1e6, 3))
+ row.append(round(float(itm[u'stdev']) / 1e6, 3))
+ for rca in rcas:
+ if rca is None:
+ continue
+ rca_nr = rca[u"data"].get(row[0], u"-")
+ row.append(f"[{rca_nr}]" if rca_nr != u"-" else u"-")
+ tbl_for_csv.append(row)
+
+ header_csv = [u"Test Case", ]
+ for col in cols:
+ header_csv.append(f"Avg({col[u'title']})")
+ header_csv.append(f"Stdev({col[u'title']})")
+ for comp in comparisons:
+ header_csv.append(
+ f"Avg({comp.get(u'title', u'')})"
+ )
+ header_csv.append(
+ f"Stdev({comp.get(u'title', u'')})"
+ )
+ for rca in rcas:
+ if rca:
+ header_csv.append(rca[u"title"])
+
+ legend_lst = table.get(u"legend", None)
+ if legend_lst is None:
+ legend = u""
+ else:
+ legend = u"\n" + u"\n".join(legend_lst) + u"\n"
+
+ footnote = u""
+ if rcas and any(rcas):
+ footnote += u"\nRoot Cause Analysis:\n"
+ for rca in rcas:
+ if rca:
+ footnote += f"{rca[u'data'].get(u'footnote', u'')}\n"
+
+ csv_file_name = f"{table[u'output-file']}-csv.csv"
+ with open(csv_file_name, u"wt", encoding='utf-8') as file_handler:
+ file_handler.write(
+ u",".join([f'"{itm}"' for itm in header_csv]) + u"\n"
+ )
+ for test in tbl_for_csv:
+ file_handler.write(
+ u",".join([f'"{item}"' for item in test]) + u"\n"
+ )
+ if legend_lst:
+ for item in legend_lst:
+ file_handler.write(f'"{item}"\n')
+ if footnote:
+ for itm in footnote.split(u"\n"):
+ file_handler.write(f'"{itm}"\n')
+
+ tbl_tmp = list()
+ max_lens = [0, ] * len(tbl_cmp_lst[0])
+ for line in tbl_cmp_lst:
+ row = [line[0], ]
+ for idx, itm in enumerate(line[1:]):
+ if itm is None or not isinstance(itm, dict) or \
+ itm.get(u'mean', None) is None or \
+ itm.get(u'stdev', None) is None:
+ new_itm = u"NT"
+ else:
+ if idx < len(cols):
+ new_itm = (
+ f"{round(float(itm[u'mean']) / 1e6, 2)} "
+ f"\u00B1{round(float(itm[u'stdev']) / 1e6, 2)}".
+ replace(u"nan", u"NaN")
+ )
+ else:
+ new_itm = (
+ f"{round(float(itm[u'mean']) / 1e6, 2):+} "
+ f"\u00B1{round(float(itm[u'stdev']) / 1e6, 2)}".
+ replace(u"nan", u"NaN")
+ )
+ if len(new_itm.rsplit(u" ", 1)[-1]) > max_lens[idx]:
+ max_lens[idx] = len(new_itm.rsplit(u" ", 1)[-1])
+ row.append(new_itm)
+
+ tbl_tmp.append(row)
+
+ header = [u"Test Case", ]
+ header.extend([col[u"title"] for col in cols])
+ header.extend([comp.get(u"title", u"") for comp in comparisons])
+
+ tbl_final = list()
+ for line in tbl_tmp:
+ row = [line[0], ]
+ for idx, itm in enumerate(line[1:]):
+ if itm in (u"NT", u"NaN"):
+ row.append(itm)
+ continue
+ itm_lst = itm.rsplit(u"\u00B1", 1)
+ itm_lst[-1] = \
+ f"{u' ' * (max_lens[idx] - len(itm_lst[-1]))}{itm_lst[-1]}"
+ itm_str = u"\u00B1".join(itm_lst)
+
+ if idx >= len(cols):
+ # Diffs
+ rca = rcas[idx - len(cols)]
+ if rca:
+ # Add rcas to diffs
+ rca_nr = rca[u"data"].get(row[0], None)
+ if rca_nr:
+ hdr_len = len(header[idx + 1]) - 1
+ if hdr_len < 19:
+ hdr_len = 19
+ rca_nr = f"[{rca_nr}]"
+ itm_str = (
+ f"{u' ' * (4 - len(rca_nr))}{rca_nr}"
+ f"{u' ' * (hdr_len - 4 - len(itm_str))}"
+ f"{itm_str}"
+ )
+ row.append(itm_str)
+ tbl_final.append(row)
+
+ # Generate csv tables:
+ csv_file_name = f"{table[u'output-file']}.csv"
+ logging.info(f" Writing the file {csv_file_name}")
+ with open(csv_file_name, u"wt", encoding='utf-8') as file_handler:
+ file_handler.write(u";".join(header) + u"\n")
+ for test in tbl_final:
+ file_handler.write(u";".join([str(item) for item in test]) + u"\n")
+
+ # Generate txt table:
+ txt_file_name = f"{table[u'output-file']}.txt"
+ logging.info(f" Writing the file {txt_file_name}")
+ convert_csv_to_pretty_txt(csv_file_name, txt_file_name, delimiter=u";")
+
+ with open(txt_file_name, u'a', encoding='utf-8') as file_handler:
+ file_handler.write(legend)
+ file_handler.write(footnote)
+
+ # Generate html table:
+ _tpc_generate_html_table(
+ header,
+ tbl_final,
+ table[u'output-file'],
+ legend=legend,
+ footnote=footnote,
+ sort_data=False,
+ title=table.get(u"title", u"")
+ )
+
+
+def table_weekly_comparison(table, in_data):
+ """Generate the table(s) with algorithm: table_weekly_comparison
+ specified in the specification file.
+
+ :param table: Table to generate.
+ :param in_data: Data to process.
+ :type table: pandas.Series
+ :type in_data: InputData
+ """
+ logging.info(f" Generating the table {table.get(u'title', u'')} ...")
+
+ # Transform the data
+ logging.info(
+ f" Creating the data set for the {table.get(u'type', u'')} "
+ f"{table.get(u'title', u'')}."
+ )
+
+ incl_tests = table.get(u"include-tests", None)
+ if incl_tests not in (u"NDR", u"PDR"):
+ logging.error(f"Wrong tests to include specified ({incl_tests}).")