CSIT-1126: Detection and reporting of failed MRR tests 28/13028/20
authorTibor Frank <tifrank@cisco.com>
Wed, 13 Jun 2018 13:19:18 +0000 (15:19 +0200)
committerTibor Frank <tifrank@cisco.com>
Mon, 18 Jun 2018 13:12:10 +0000 (13:12 +0000)
Change-Id: Iafdd12a573b58ab3ba9c69699cd97c5a4e598601
Signed-off-by: Tibor Frank <tifrank@cisco.com>
docs/cpta/introduction/index.rst
docs/new/cpta/introduction/index.rst
resources/tools/presentation/generator_tables.py
resources/tools/presentation/input_data_parser.py
resources/tools/presentation/new/generator_tables.py
resources/tools/presentation/new/input_data_parser.py
resources/tools/presentation/new/specification_CPTA.yaml
resources/tools/presentation/new/utils.py
resources/tools/presentation/specification_CPTA.yaml
resources/tools/presentation/utils.py

index 016037b..991181a 100644 (file)
@@ -20,18 +20,40 @@ executed twice a day (target start: every 12 hrs, 02:00, 14:00 UTC). All
 trend and anomaly evaluation is based on a rolling window of <N=14> data
 samples, covering last 7 days.
 
-Legend to table:
+Failed tests
+------------
 
-    - **Test Case** : name of FD.io CSIT test case, naming convention
+The table lists the tests which failed over the <N=14> runs of the trending
+jobs.
+
+Legend to the table:
+
+    - **Test Case**: name of FD.io CSIT test case, naming convention
+      `here <https://wiki.fd.io/view/CSIT/csit-test-naming>`_.
+    - **Fails [#]**: number of fails of the tests over the period.
+    - **Last Fail [Date]**: the date and time when the test failed the last
+      time.
+    - **Last Fail [VPP Build]**: VPP build which was tested when the test failed
+      the last time.
+    - **Last Fail [CSIT Build]**: the last CSIT build where the test failed.
+
+.. include:: ../../../_build/_static/vpp/failed-tests.rst
+
+Dashboard
+---------
+
+Legend to the tables:
+
+    - **Test Case**: name of FD.io CSIT test case, naming convention
       `here <https://wiki.fd.io/view/CSIT/csit-test-naming>`_.
-    - **Trend [Mpps]** : last value of performance trend.
-    - **Short-Term Change [%]** : Relative change of last trend value
+    - **Trend [Mpps]**: last value of performance trend.
+    - **Short-Term Change [%]**: Relative change of last trend value
       vs. last week trend value.
-    - **Long-Term Change [%]** : Relative change of last trend value vs.
+    - **Long-Term Change [%]**: Relative change of last trend value vs.
       maximum of trend values over the last quarter except last week.
-    - **Regressions [#]** : Number of regressions detected.
-    - **Progressions [#]** : Number of progressions detected.
-    - **Outliers [#]** : Number of outliers detected.
+    - **Regressions [#]**: Number of regressions detected.
+    - **Progressions [#]**: Number of progressions detected.
+    - **Outliers [#]**: Number of outliers detected.
 
 Tested VPP worker-thread-core combinations (1t1c, 2t2c, 4t4c) are listed
 in separate tables in section 1.x. Followed by trending methodology in
@@ -39,16 +61,16 @@ section 2. and trendline graphs in sections 3.x. Performance test  data
 used for trendline graphs is provided in sections 4.x.
 
 VPP worker on 1t1c
-------------------
+``````````````````
 
 .. include:: ../../../_build/_static/vpp/performance-trending-dashboard-1t1c.rst
 
 VPP worker on 2t2c
-------------------
+``````````````````
 
 .. include:: ../../../_build/_static/vpp/performance-trending-dashboard-2t2c.rst
 
 VPP worker on 4t4c
-------------------
+``````````````````
 
 .. include:: ../../../_build/_static/vpp/performance-trending-dashboard-4t4c.rst
index 016037b..991181a 100644 (file)
@@ -20,18 +20,40 @@ executed twice a day (target start: every 12 hrs, 02:00, 14:00 UTC). All
 trend and anomaly evaluation is based on a rolling window of <N=14> data
 samples, covering last 7 days.
 
-Legend to table:
+Failed tests
+------------
 
-    - **Test Case** : name of FD.io CSIT test case, naming convention
+The table lists the tests which failed over the <N=14> runs of the trending
+jobs.
+
+Legend to the table:
+
+    - **Test Case**: name of FD.io CSIT test case, naming convention
+      `here <https://wiki.fd.io/view/CSIT/csit-test-naming>`_.
+    - **Fails [#]**: number of fails of the tests over the period.
+    - **Last Fail [Date]**: the date and time when the test failed the last
+      time.
+    - **Last Fail [VPP Build]**: VPP build which was tested when the test failed
+      the last time.
+    - **Last Fail [CSIT Build]**: the last CSIT build where the test failed.
+
+.. include:: ../../../_build/_static/vpp/failed-tests.rst
+
+Dashboard
+---------
+
+Legend to the tables:
+
+    - **Test Case**: name of FD.io CSIT test case, naming convention
       `here <https://wiki.fd.io/view/CSIT/csit-test-naming>`_.
-    - **Trend [Mpps]** : last value of performance trend.
-    - **Short-Term Change [%]** : Relative change of last trend value
+    - **Trend [Mpps]**: last value of performance trend.
+    - **Short-Term Change [%]**: Relative change of last trend value
       vs. last week trend value.
-    - **Long-Term Change [%]** : Relative change of last trend value vs.
+    - **Long-Term Change [%]**: Relative change of last trend value vs.
       maximum of trend values over the last quarter except last week.
-    - **Regressions [#]** : Number of regressions detected.
-    - **Progressions [#]** : Number of progressions detected.
-    - **Outliers [#]** : Number of outliers detected.
+    - **Regressions [#]**: Number of regressions detected.
+    - **Progressions [#]**: Number of progressions detected.
+    - **Outliers [#]**: Number of outliers detected.
 
 Tested VPP worker-thread-core combinations (1t1c, 2t2c, 4t4c) are listed
 in separate tables in section 1.x. Followed by trending methodology in
@@ -39,16 +61,16 @@ section 2. and trendline graphs in sections 3.x. Performance test  data
 used for trendline graphs is provided in sections 4.x.
 
 VPP worker on 1t1c
-------------------
+``````````````````
 
 .. include:: ../../../_build/_static/vpp/performance-trending-dashboard-1t1c.rst
 
 VPP worker on 2t2c
-------------------
+``````````````````
 
 .. include:: ../../../_build/_static/vpp/performance-trending-dashboard-2t2c.rst
 
 VPP worker on 4t4c
-------------------
+``````````````````
 
 .. include:: ../../../_build/_static/vpp/performance-trending-dashboard-4t4c.rst
index abece85..6aa57db 100644 (file)
@@ -17,7 +17,6 @@
 
 import logging
 import csv
-import prettytable
 import pandas as pd
 
 from string import replace
@@ -27,7 +26,7 @@ from xml.etree import ElementTree as ET
 
 from errors import PresentationError
 from utils import mean, stdev, relative_change, remove_outliers,\
-    split_outliers, classify_anomalies
+    split_outliers, classify_anomalies, convert_csv_to_pretty_txt
 
 
 def generate_tables(spec, data):
@@ -522,18 +521,8 @@ def table_performance_comparison(table, input_data):
                      ]
 
     for i, txt_name in enumerate(tbl_names_txt):
-        txt_table = None
         logging.info("      Writing file: '{0}'".format(txt_name))
-        with open(tbl_names[i], 'rb') as csv_file:
-            csv_content = csv.reader(csv_file, delimiter=',', quotechar='"')
-            for row in csv_content:
-                if txt_table is None:
-                    txt_table = prettytable.PrettyTable(row)
-                else:
-                    txt_table.add_row(row)
-            txt_table.align["Test case"] = "l"
-        with open(txt_name, "w") as txt_file:
-            txt_file.write(str(txt_table))
+        convert_csv_to_pretty_txt(tbl_names[i], txt_name)
 
     # Selected tests in csv:
     input_file = "{0}-ndr-1t1c-full{1}".format(table["output-file"],
@@ -711,18 +700,8 @@ def table_performance_comparison_mrr(table, input_data):
                      ]
 
     for i, txt_name in enumerate(tbl_names_txt):
-        txt_table = None
         logging.info("      Writing file: '{0}'".format(txt_name))
-        with open(tbl_names[i], 'rb') as csv_file:
-            csv_content = csv.reader(csv_file, delimiter=',', quotechar='"')
-            for row in csv_content:
-                if txt_table is None:
-                    txt_table = prettytable.PrettyTable(row)
-                else:
-                    txt_table.add_row(row)
-            txt_table.align["Test case"] = "l"
-        with open(txt_name, "w") as txt_file:
-            txt_file.write(str(txt_table))
+        convert_csv_to_pretty_txt(tbl_names[i], txt_name)
 
 
 def table_performance_trending_dashboard(table, input_data):
@@ -852,18 +831,92 @@ def table_performance_trending_dashboard(table, input_data):
             file_handler.write(",".join([str(item) for item in test]) + '\n')
 
     txt_file_name = "{0}.txt".format(table["output-file"])
-    txt_table = None
     logging.info("    Writing file: '{0}'".format(txt_file_name))
-    with open(file_name, 'rb') as csv_file:
-        csv_content = csv.reader(csv_file, delimiter=',', quotechar='"')
-        for row in csv_content:
-            if txt_table is None:
-                txt_table = prettytable.PrettyTable(row)
-            else:
-                txt_table.add_row(row)
-        txt_table.align["Test case"] = "l"
-    with open(txt_file_name, "w") as txt_file:
-        txt_file.write(str(txt_table))
+    convert_csv_to_pretty_txt(file_name, txt_file_name)
+
+
+def _generate_url(base, test_name):
+    """Generate URL to a trending plot from the name of the test case.
+
+    :param base: The base part of URL common to all test cases.
+    :param test_name: The name of the test case.
+    :type base: str
+    :type test_name: str
+    :returns: The URL to the plot with the trending data for the given test
+        case.
+    :rtype str
+    """
+
+    url = base
+    file_name = ""
+    anchor = "#"
+    feature = ""
+
+    if "lbdpdk" in test_name or "lbvpp" in test_name:
+        file_name = "link_bonding.html"
+
+    elif "testpmd" in test_name or "l3fwd" in test_name:
+        file_name = "dpdk.html"
+
+    elif "memif" in test_name:
+        file_name = "container_memif.html"
+
+    elif "srv6" in test_name:
+        file_name = "srv6.html"
+
+    elif "vhost" in test_name:
+        if "l2xcbase" in test_name or "l2bdbasemaclrn" in test_name:
+            file_name = "vm_vhost_l2.html"
+        elif "ip4base" in test_name:
+            file_name = "vm_vhost_ip4.html"
+
+    elif "ipsec" in test_name:
+        file_name = "ipsec.html"
+
+    elif "ethip4lispip" in test_name or "ethip4vxlan" in test_name:
+        file_name = "ip4_tunnels.html"
+
+    elif "ip4base" in test_name or "ip4scale" in test_name:
+        file_name = "ip4.html"
+        if "iacl" in test_name or "snat" in test_name or "cop" in test_name:
+            feature = "-features"
+
+    elif "ip6base" in test_name or "ip6scale" in test_name:
+        file_name = "ip6.html"
+
+    elif "l2xcbase" in test_name or "l2xcscale" in test_name \
+            or "l2bdbasemaclrn" in test_name or "l2bdscale" in test_name \
+            or "l2dbbasemaclrn" in test_name or "l2dbscale" in test_name:
+        file_name = "l2.html"
+        if "iacl" in test_name:
+            feature = "-features"
+
+    if "x520" in test_name:
+        anchor += "x520-"
+    elif "x710" in test_name:
+        anchor += "x710-"
+    elif "xl710" in test_name:
+        anchor += "xl710-"
+
+    if "64b" in test_name:
+        anchor += "64b-"
+    elif "78b" in test_name:
+        anchor += "78b-"
+    elif "imix" in test_name:
+        anchor += "imix-"
+    elif "9000b" in test_name:
+        anchor += "9000b-"
+    elif "1518" in test_name:
+        anchor += "1518b-"
+
+    if "1t1c" in test_name:
+        anchor += "1t1c"
+    elif "2t2c" in test_name:
+        anchor += "2t2c"
+    elif "4t4c" in test_name:
+        anchor += "4t4c"
+
+    return url + file_name + anchor + feature
 
 
 def table_performance_trending_dashboard_html(table, input_data):
@@ -924,83 +977,12 @@ def table_performance_trending_dashboard_html(table, input_data):
             alignment = "left" if c_idx == 0 else "center"
             td = ET.SubElement(tr, "td", attrib=dict(align=alignment))
             # Name:
-            url = "../trending/"
-            file_name = ""
-            anchor = "#"
-            feature = ""
             if c_idx == 0:
-                if "lbdpdk" in item or "lbvpp" in item:
-                    file_name = "link_bonding.html"
-
-                elif "testpmd" in item or "l3fwd" in item:
-                    file_name = "dpdk.html"
-
-                elif "memif" in item:
-                    file_name = "container_memif.html"
-
-                elif "srv6" in item:
-                    file_name = "srv6.html"
-
-                elif "vhost" in item:
-                    if "l2xcbase" in item or "l2bdbasemaclrn" in item:
-                        file_name = "vm_vhost_l2.html"
-                    elif "ip4base" in item:
-                        file_name = "vm_vhost_ip4.html"
-
-                elif "ipsec" in item:
-                    file_name = "ipsec.html"
-
-                elif "ethip4lispip" in item or "ethip4vxlan" in item:
-                    file_name = "ip4_tunnels.html"
-
-                elif "ip4base" in item or "ip4scale" in item:
-                    file_name = "ip4.html"
-                    if "iacl" in item or "snat" in item or "cop" in item:
-                        feature = "-features"
-
-                elif "ip6base" in item or "ip6scale" in item:
-                    file_name = "ip6.html"
-
-                elif "l2xcbase" in item or "l2xcscale" in item \
-                        or "l2bdbasemaclrn" in item or "l2bdscale" in item \
-                        or "l2dbbasemaclrn" in item or "l2dbscale" in item:
-                    file_name = "l2.html"
-                    if "iacl" in item:
-                        feature = "-features"
-
-                if "x520" in item:
-                    anchor += "x520-"
-                elif "x710" in item:
-                    anchor += "x710-"
-                elif "xl710" in item:
-                    anchor += "xl710-"
-
-                if "64b" in item:
-                    anchor += "64b-"
-                elif "78b" in item:
-                    anchor += "78b-"
-                elif "imix" in item:
-                    anchor += "imix-"
-                elif "9000b" in item:
-                    anchor += "9000b-"
-                elif "1518" in item:
-                    anchor += "1518b-"
-
-                if "1t1c" in item:
-                    anchor += "1t1c"
-                elif "2t2c" in item:
-                    anchor += "2t2c"
-                elif "4t4c" in item:
-                    anchor += "4t4c"
-
-                url = url + file_name + anchor + feature
-
+                url = _generate_url("../trending/", item)
                 ref = ET.SubElement(td, "a", attrib=dict(href=url))
                 ref.text = item
-
             else:
                 td.text = item
-
     try:
         with open(table["output-file"], 'w') as html_file:
             logging.info("    Writing file: '{0}'".format(table["output-file"]))
@@ -1010,3 +992,155 @@ def table_performance_trending_dashboard_html(table, input_data):
     except KeyError:
         logging.warning("The output file is not defined.")
         return
+
+
+def table_failed_tests(table, input_data):
+    """Generate the table(s) with algorithm: table_failed_tests
+    specified in the specification file.
+
+    :param table: Table to generate.
+    :param input_data: Data to process.
+    :type table: pandas.Series
+    :type input_data: InputData
+    """
+
+    logging.info("  Generating the table {0} ...".
+                 format(table.get("title", "")))
+
+    # Transform the data
+    logging.info("    Creating the data set for the {0} '{1}'.".
+                 format(table.get("type", ""), table.get("title", "")))
+    data = input_data.filter_data(table, continue_on_error=True)
+
+    # Prepare the header of the tables
+    header = ["Test Case",
+              "Fails [#]",
+              "Last Fail [Timestamp]",
+              "Last Fail [VPP Build]",
+              "Last Fail [CSIT Build]"]
+
+    # Generate the data for the table according to the model in the table
+    # specification
+    tbl_dict = dict()
+    for job, builds in table["data"].items():
+        for build in builds:
+            build = str(build)
+            for tst_name, tst_data in data[job][build].iteritems():
+                if tst_name.lower() in table["ignore-list"]:
+                    continue
+                if tbl_dict.get(tst_name, None) is None:
+                    name = "{0}-{1}".format(tst_data["parent"].split("-")[0],
+                                            "-".join(tst_data["name"].
+                                                     split("-")[1:]))
+                    tbl_dict[tst_name] = {"name": name,
+                                          "data": OrderedDict()}
+                try:
+                    tbl_dict[tst_name]["data"][build] = (
+                        tst_data["status"],
+                        input_data.metadata(job, build).get("generated", ""),
+                        input_data.metadata(job, build).get("version", ""),
+                        build)
+                except (TypeError, KeyError):
+                    pass  # No data in output.xml for this test
+
+    tbl_lst = list()
+    for tst_data in tbl_dict.values():
+        win_size = min(len(tst_data["data"]), table["window"])
+        fails_nr = 0
+        for val in tst_data["data"].values()[-win_size:]:
+            if val[0] == "FAIL":
+                fails_nr += 1
+                fails_last_date = val[1]
+                fails_last_vpp = val[2]
+                fails_last_csit = val[3]
+        if fails_nr:
+            tbl_lst.append([tst_data["name"],
+                            fails_nr,
+                            fails_last_date,
+                            fails_last_vpp,
+                            "mrr-daily-build-{0}".format(fails_last_csit)])
+
+    tbl_lst.sort(key=lambda rel: rel[2], reverse=True)
+    tbl_sorted = list()
+    for nrf in range(table["window"], -1, -1):
+        tbl_fails = [item for item in tbl_lst if item[1] == nrf]
+        tbl_sorted.extend(tbl_fails)
+    file_name = "{0}{1}".format(table["output-file"], table["output-file-ext"])
+
+    logging.info("    Writing file: '{0}'".format(file_name))
+    with open(file_name, "w") as file_handler:
+        file_handler.write(",".join(header) + "\n")
+        for test in tbl_sorted:
+            file_handler.write(",".join([str(item) for item in test]) + '\n')
+
+    txt_file_name = "{0}.txt".format(table["output-file"])
+    logging.info("    Writing file: '{0}'".format(txt_file_name))
+    convert_csv_to_pretty_txt(file_name, txt_file_name)
+
+
+def table_failed_tests_html(table, input_data):
+    """Generate the table(s) with algorithm: table_failed_tests_html
+    specified in the specification file.
+
+    :param table: Table to generate.
+    :param input_data: Data to process.
+    :type table: pandas.Series
+    :type input_data: InputData
+    """
+
+    logging.info("  Generating the table {0} ...".
+                 format(table.get("title", "")))
+
+    try:
+        with open(table["input-file"], 'rb') as csv_file:
+            csv_content = csv.reader(csv_file, delimiter=',', quotechar='"')
+            csv_lst = [item for item in csv_content]
+    except KeyError:
+        logging.warning("The input file is not defined.")
+        return
+    except csv.Error as err:
+        logging.warning("Not possible to process the file '{0}'.\n{1}".
+                        format(table["input-file"], err))
+        return
+
+    # Table:
+    failed_tests = ET.Element("table", attrib=dict(width="100%", border='0'))
+
+    # Table header:
+    tr = ET.SubElement(failed_tests, "tr", attrib=dict(bgcolor="#7eade7"))
+    for idx, item in enumerate(csv_lst[0]):
+        alignment = "left" if idx == 0 else "center"
+        th = ET.SubElement(tr, "th", attrib=dict(align=alignment))
+        th.text = item
+
+    # Rows:
+    colors = {"very-bad": ("#ffcccc", "#ff9999"),
+              "bad": ("#e9f1fb", "#d4e4f7")}
+    for r_idx, row in enumerate(csv_lst[1:]):
+        if int(row[1]) > 7:
+            color = "very-bad"
+        else:
+            color = "bad"
+        background = colors[color][r_idx % 2]
+        tr = ET.SubElement(failed_tests, "tr", attrib=dict(bgcolor=background))
+
+        # Columns:
+        for c_idx, item in enumerate(row):
+            alignment = "left" if c_idx == 0 else "center"
+            td = ET.SubElement(tr, "td", attrib=dict(align=alignment))
+            # Name:
+            if c_idx == 0:
+                url = _generate_url("../trending/", item)
+                ref = ET.SubElement(td, "a", attrib=dict(href=url))
+                ref.text = item
+            else:
+                td.text = item
+    try:
+        with open(table["output-file"], 'w') as html_file:
+            logging.info("    Writing file: '{0}'".format(table["output-file"]))
+            html_file.write(".. raw:: html\n\n\t")
+            html_file.write(ET.tostring(failed_tests))
+            html_file.write("\n\t<p><br><br></p>\n")
+    except KeyError:
+        logging.warning("The output file is not defined.")
+        return
index 0bb2b6c..cf13237 100644 (file)
@@ -464,6 +464,7 @@ class ExecutionChecker(ResultVisitor):
         test_result["doc"] = replace(doc_str, ' |br| [', '[', maxreplace=1)
         test_result["msg"] = test.message.replace('\n', ' |br| '). \
             replace('\r', '').replace('"', "'")
+        test_result["status"] = test.status
         if test.status == "PASS" and ("NDRPDRDISC" in tags or
                                       "TCP" in tags or
                                       "MRR" in tags):
@@ -507,6 +508,7 @@ class ExecutionChecker(ResultVisitor):
                 test_result["result"] = dict()
                 test_result["result"]["value"] = int(groups.group(2))
                 test_result["result"]["unit"] = groups.group(1)
+
             elif test_type in ("MRR", ):
                 groups = re.search(self.REGEX_MRR, test.message)
                 test_result["result"] = dict()
@@ -516,8 +518,6 @@ class ExecutionChecker(ResultVisitor):
                 test_result["result"]["throughput"] = int(
                     test_result["result"]["rx"] /
                     test_result["result"]["duration"])
-        else:
-            test_result["status"] = test.status
 
         self._test_ID = test.longname.lower()
         self._data["tests"][self._test_ID] = test_result
index 735fd21..43117cc 100644 (file)
@@ -17,7 +17,6 @@
 
 import logging
 import csv
-import prettytable
 import pandas as pd
 
 from string import replace
@@ -26,7 +25,8 @@ from numpy import nan, isnan
 from xml.etree import ElementTree as ET
 
 from errors import PresentationError
-from utils import mean, stdev, relative_change, classify_anomalies
+from utils import mean, stdev, relative_change, classify_anomalies, \
+    convert_csv_to_pretty_txt
 
 
 def generate_tables(spec, data):
@@ -506,18 +506,8 @@ def table_performance_comparison(table, input_data):
                      ]
 
     for i, txt_name in enumerate(tbl_names_txt):
-        txt_table = None
         logging.info("      Writing file: '{0}'".format(txt_name))
-        with open(tbl_names[i], 'rb') as csv_file:
-            csv_content = csv.reader(csv_file, delimiter=',', quotechar='"')
-            for row in csv_content:
-                if txt_table is None:
-                    txt_table = prettytable.PrettyTable(row)
-                else:
-                    txt_table.add_row(row)
-            txt_table.align["Test case"] = "l"
-        with open(txt_name, "w") as txt_file:
-            txt_file.write(str(txt_table))
+        convert_csv_to_pretty_txt(tbl_names[i], txt_name)
 
     # Selected tests in csv:
     input_file = "{0}-ndr-1t1c-full{1}".format(table["output-file"],
@@ -685,18 +675,8 @@ def table_performance_comparison_mrr(table, input_data):
                      ]
 
     for i, txt_name in enumerate(tbl_names_txt):
-        txt_table = None
         logging.info("      Writing file: '{0}'".format(txt_name))
-        with open(tbl_names[i], 'rb') as csv_file:
-            csv_content = csv.reader(csv_file, delimiter=',', quotechar='"')
-            for row in csv_content:
-                if txt_table is None:
-                    txt_table = prettytable.PrettyTable(row)
-                else:
-                    txt_table.add_row(row)
-            txt_table.align["Test case"] = "l"
-        with open(txt_name, "w") as txt_file:
-            txt_file.write(str(txt_table))
+        convert_csv_to_pretty_txt(tbl_names[i], txt_name)
 
 
 def table_performance_trending_dashboard(table, input_data):
@@ -810,18 +790,92 @@ def table_performance_trending_dashboard(table, input_data):
             file_handler.write(",".join([str(item) for item in test]) + '\n')
 
     txt_file_name = "{0}.txt".format(table["output-file"])
-    txt_table = None
     logging.info("    Writing file: '{0}'".format(txt_file_name))
-    with open(file_name, 'rb') as csv_file:
-        csv_content = csv.reader(csv_file, delimiter=',', quotechar='"')
-        for row in csv_content:
-            if txt_table is None:
-                txt_table = prettytable.PrettyTable(row)
-            else:
-                txt_table.add_row(row)
-        txt_table.align["Test case"] = "l"
-    with open(txt_file_name, "w") as txt_file:
-        txt_file.write(str(txt_table))
+    convert_csv_to_pretty_txt(file_name, txt_file_name)
+
+
+def _generate_url(base, test_name):
+    """Generate URL to a trending plot from the name of the test case.
+
+    :param base: The base part of URL common to all test cases.
+    :param test_name: The name of the test case.
+    :type base: str
+    :type test_name: str
+    :returns: The URL to the plot with the trending data for the given test
+        case.
+    :rtype str
+    """
+
+    url = base
+    file_name = ""
+    anchor = "#"
+    feature = ""
+
+    if "lbdpdk" in test_name or "lbvpp" in test_name:
+        file_name = "link_bonding.html"
+
+    elif "testpmd" in test_name or "l3fwd" in test_name:
+        file_name = "dpdk.html"
+
+    elif "memif" in test_name:
+        file_name = "container_memif.html"
+
+    elif "srv6" in test_name:
+        file_name = "srv6.html"
+
+    elif "vhost" in test_name:
+        if "l2xcbase" in test_name or "l2bdbasemaclrn" in test_name:
+            file_name = "vm_vhost_l2.html"
+        elif "ip4base" in test_name:
+            file_name = "vm_vhost_ip4.html"
+
+    elif "ipsec" in test_name:
+        file_name = "ipsec.html"
+
+    elif "ethip4lispip" in test_name or "ethip4vxlan" in test_name:
+        file_name = "ip4_tunnels.html"
+
+    elif "ip4base" in test_name or "ip4scale" in test_name:
+        file_name = "ip4.html"
+        if "iacl" in test_name or "snat" in test_name or "cop" in test_name:
+            feature = "-features"
+
+    elif "ip6base" in test_name or "ip6scale" in test_name:
+        file_name = "ip6.html"
+
+    elif "l2xcbase" in test_name or "l2xcscale" in test_name \
+            or "l2bdbasemaclrn" in test_name or "l2bdscale" in test_name \
+            or "l2dbbasemaclrn" in test_name or "l2dbscale" in test_name:
+        file_name = "l2.html"
+        if "iacl" in test_name:
+            feature = "-features"
+
+    if "x520" in test_name:
+        anchor += "x520-"
+    elif "x710" in test_name:
+        anchor += "x710-"
+    elif "xl710" in test_name:
+        anchor += "xl710-"
+
+    if "64b" in test_name:
+        anchor += "64b-"
+    elif "78b" in test_name:
+        anchor += "78b-"
+    elif "imix" in test_name:
+        anchor += "imix-"
+    elif "9000b" in test_name:
+        anchor += "9000b-"
+    elif "1518" in test_name:
+        anchor += "1518b-"
+
+    if "1t1c" in test_name:
+        anchor += "1t1c"
+    elif "2t2c" in test_name:
+        anchor += "2t2c"
+    elif "4t4c" in test_name:
+        anchor += "4t4c"
+
+    return url + file_name + anchor + feature
 
 
 def table_performance_trending_dashboard_html(table, input_data):
@@ -879,83 +933,12 @@ def table_performance_trending_dashboard_html(table, input_data):
             alignment = "left" if c_idx == 0 else "center"
             td = ET.SubElement(tr, "td", attrib=dict(align=alignment))
             # Name:
-            url = "../trending/"
-            file_name = ""
-            anchor = "#"
-            feature = ""
             if c_idx == 0:
-                if "lbdpdk" in item or "lbvpp" in item:
-                    file_name = "link_bonding.html"
-
-                elif "testpmd" in item or "l3fwd" in item:
-                    file_name = "dpdk.html"
-
-                elif "memif" in item:
-                    file_name = "container_memif.html"
-
-                elif "srv6" in item:
-                    file_name = "srv6.html"
-
-                elif "vhost" in item:
-                    if "l2xcbase" in item or "l2bdbasemaclrn" in item:
-                        file_name = "vm_vhost_l2.html"
-                    elif "ip4base" in item:
-                        file_name = "vm_vhost_ip4.html"
-
-                elif "ipsec" in item:
-                    file_name = "ipsec.html"
-
-                elif "ethip4lispip" in item or "ethip4vxlan" in item:
-                    file_name = "ip4_tunnels.html"
-
-                elif "ip4base" in item or "ip4scale" in item:
-                    file_name = "ip4.html"
-                    if "iacl" in item or "snat" in item or "cop" in item:
-                        feature = "-features"
-
-                elif "ip6base" in item or "ip6scale" in item:
-                    file_name = "ip6.html"
-
-                elif "l2xcbase" in item or "l2xcscale" in item \
-                        or "l2bdbasemaclrn" in item or "l2bdscale" in item \
-                        or "l2dbbasemaclrn" in item or "l2dbscale" in item:
-                    file_name = "l2.html"
-                    if "iacl" in item:
-                        feature = "-features"
-
-                if "x520" in item:
-                    anchor += "x520-"
-                elif "x710" in item:
-                    anchor += "x710-"
-                elif "xl710" in item:
-                    anchor += "xl710-"
-
-                if "64b" in item:
-                    anchor += "64b-"
-                elif "78b" in item:
-                    anchor += "78b-"
-                elif "imix" in item:
-                    anchor += "imix-"
-                elif "9000b" in item:
-                    anchor += "9000b-"
-                elif "1518" in item:
-                    anchor += "1518b-"
-
-                if "1t1c" in item:
-                    anchor += "1t1c"
-                elif "2t2c" in item:
-                    anchor += "2t2c"
-                elif "4t4c" in item:
-                    anchor += "4t4c"
-
-                url = url + file_name + anchor + feature
-
+                url = _generate_url("../trending/", item)
                 ref = ET.SubElement(td, "a", attrib=dict(href=url))
                 ref.text = item
-
             else:
                 td.text = item
-
     try:
         with open(table["output-file"], 'w') as html_file:
             logging.info("    Writing file: '{0}'".format(table["output-file"]))
@@ -965,3 +948,155 @@ def table_performance_trending_dashboard_html(table, input_data):
     except KeyError:
         logging.warning("The output file is not defined.")
         return
+
+
+def table_failed_tests(table, input_data):
+    """Generate the table(s) with algorithm: table_failed_tests
+    specified in the specification file.
+
+    :param table: Table to generate.
+    :param input_data: Data to process.
+    :type table: pandas.Series
+    :type input_data: InputData
+    """
+
+    logging.info("  Generating the table {0} ...".
+                 format(table.get("title", "")))
+
+    # Transform the data
+    logging.info("    Creating the data set for the {0} '{1}'.".
+                 format(table.get("type", ""), table.get("title", "")))
+    data = input_data.filter_data(table, continue_on_error=True)
+
+    # Prepare the header of the tables
+    header = ["Test Case",
+              "Fails [#]",
+              "Last Fail [Timestamp]",
+              "Last Fail [VPP Build]",
+              "Last Fail [CSIT Build]"]
+
+    # Generate the data for the table according to the model in the table
+    # specification
+    tbl_dict = dict()
+    for job, builds in table["data"].items():
+        for build in builds:
+            build = str(build)
+            for tst_name, tst_data in data[job][build].iteritems():
+                if tst_name.lower() in table["ignore-list"]:
+                    continue
+                if tbl_dict.get(tst_name, None) is None:
+                    name = "{0}-{1}".format(tst_data["parent"].split("-")[0],
+                                            "-".join(tst_data["name"].
+                                                     split("-")[1:]))
+                    tbl_dict[tst_name] = {"name": name,
+                                          "data": OrderedDict()}
+                try:
+                    tbl_dict[tst_name]["data"][build] = (
+                        tst_data["status"],
+                        input_data.metadata(job, build).get("generated", ""),
+                        input_data.metadata(job, build).get("version", ""),
+                        build)
+                except (TypeError, KeyError):
+                    pass  # No data in output.xml for this test
+
+    tbl_lst = list()
+    for tst_data in tbl_dict.values():
+        win_size = min(len(tst_data["data"]), table["window"])
+        fails_nr = 0
+        for val in tst_data["data"].values()[-win_size:]:
+            if val[0] == "FAIL":
+                fails_nr += 1
+                fails_last_date = val[1]
+                fails_last_vpp = val[2]
+                fails_last_csit = val[3]
+        if fails_nr:
+            tbl_lst.append([tst_data["name"],
+                            fails_nr,
+                            fails_last_date,
+                            fails_last_vpp,
+                            "mrr-daily-build-{0}".format(fails_last_csit)])
+
+    tbl_lst.sort(key=lambda rel: rel[2], reverse=True)
+    tbl_sorted = list()
+    for nrf in range(table["window"], -1, -1):
+        tbl_fails = [item for item in tbl_lst if item[1] == nrf]
+        tbl_sorted.extend(tbl_fails)
+    file_name = "{0}{1}".format(table["output-file"], table["output-file-ext"])
+
+    logging.info("    Writing file: '{0}'".format(file_name))
+    with open(file_name, "w") as file_handler:
+        file_handler.write(",".join(header) + "\n")
+        for test in tbl_sorted:
+            file_handler.write(",".join([str(item) for item in test]) + '\n')
+
+    txt_file_name = "{0}.txt".format(table["output-file"])
+    logging.info("    Writing file: '{0}'".format(txt_file_name))
+    convert_csv_to_pretty_txt(file_name, txt_file_name)
+
+
+def table_failed_tests_html(table, input_data):
+    """Generate the table(s) with algorithm: table_failed_tests_html
+    specified in the specification file.
+
+    :param table: Table to generate.
+    :param input_data: Data to process.
+    :type table: pandas.Series
+    :type input_data: InputData
+    """
+
+    logging.info("  Generating the table {0} ...".
+                 format(table.get("title", "")))
+
+    try:
+        with open(table["input-file"], 'rb') as csv_file:
+            csv_content = csv.reader(csv_file, delimiter=',', quotechar='"')
+            csv_lst = [item for item in csv_content]
+    except KeyError:
+        logging.warning("The input file is not defined.")
+        return
+    except csv.Error as err:
+        logging.warning("Not possible to process the file '{0}'.\n{1}".
+                        format(table["input-file"], err))
+        return
+
+    # Table:
+    failed_tests = ET.Element("table", attrib=dict(width="100%", border='0'))
+
+    # Table header:
+    tr = ET.SubElement(failed_tests, "tr", attrib=dict(bgcolor="#7eade7"))
+    for idx, item in enumerate(csv_lst[0]):
+        alignment = "left" if idx == 0 else "center"
+        th = ET.SubElement(tr, "th", attrib=dict(align=alignment))
+        th.text = item
+
+    # Rows:
+    colors = {"very-bad": ("#ffcccc", "#ff9999"),
+              "bad": ("#e9f1fb", "#d4e4f7")}
+    for r_idx, row in enumerate(csv_lst[1:]):
+        if int(row[1]) > 7:
+            color = "very-bad"
+        else:
+            color = "bad"
+        background = colors[color][r_idx % 2]
+        tr = ET.SubElement(failed_tests, "tr", attrib=dict(bgcolor=background))
+
+        # Columns:
+        for c_idx, item in enumerate(row):
+            alignment = "left" if c_idx == 0 else "center"
+            td = ET.SubElement(tr, "td", attrib=dict(align=alignment))
+            # Name:
+            if c_idx == 0:
+                url = _generate_url("../trending/", item)
+                ref = ET.SubElement(td, "a", attrib=dict(href=url))
+                ref.text = item
+            else:
+                td.text = item
+    try:
+        with open(table["output-file"], 'w') as html_file:
+            logging.info("    Writing file: '{0}'".format(table["output-file"]))
+            html_file.write(".. raw:: html\n\n\t")
+            html_file.write(ET.tostring(failed_tests))
+            html_file.write("\n\t<p><br><br></p>\n")
+    except KeyError:
+        logging.warning("The output file is not defined.")
+        return
index 0bb2b6c..cf13237 100644 (file)
@@ -464,6 +464,7 @@ class ExecutionChecker(ResultVisitor):
         test_result["doc"] = replace(doc_str, ' |br| [', '[', maxreplace=1)
         test_result["msg"] = test.message.replace('\n', ' |br| '). \
             replace('\r', '').replace('"', "'")
+        test_result["status"] = test.status
         if test.status == "PASS" and ("NDRPDRDISC" in tags or
                                       "TCP" in tags or
                                       "MRR" in tags):
@@ -507,6 +508,7 @@ class ExecutionChecker(ResultVisitor):
                 test_result["result"] = dict()
                 test_result["result"]["value"] = int(groups.group(2))
                 test_result["result"]["unit"] = groups.group(1)
+
             elif test_type in ("MRR", ):
                 groups = re.search(self.REGEX_MRR, test.message)
                 test_result["result"] = dict()
@@ -516,8 +518,6 @@ class ExecutionChecker(ResultVisitor):
                 test_result["result"]["throughput"] = int(
                     test_result["result"]["rx"] /
                     test_result["result"]["duration"])
-        else:
-            test_result["status"] = test.status
 
         self._test_ID = test.longname.lower()
         self._data["tests"][self._test_ID] = test_result
index 555cfae..2273eb6 100644 (file)
@@ -91,7 +91,7 @@
       height: 800
       yaxis:
         showticklabels: True
-        tickformat: ".3s"
+        tickformat: ".4s"
         title: "Throughput [pps]"
         hoverformat: ".4s"
         range: []
       height: 800
       yaxis:
         showticklabels: True
-        tickformat: ".3s"
+        tickformat: ".4s"
         title: "Throughput [pps]"
         hoverformat: ".4s"
         range: []
   - "parent"
   - "result"
   ignore-list:
+    # This test is "ndrdisc" test and was improperly tagged. It was fixed
+    # but it remains in the old output.xml files.
   - "tests.vpp.perf.l2.10ge2p1x520-eth-l2bdscale1mmaclrn-mrr.tc01-64b-1t1c-eth-l2bdscale1mmaclrn-ndrdisc"
   outlier-const: 1.5
   window: 14
   - "parent"
   - "result"
   ignore-list:
+    # This test is "ndrdisc" test and was improperly tagged. It was fixed
+    # but it remains in the old output.xml files.
   - "tests.vpp.perf.l2.10ge2p1x520-eth-l2bdscale1mmaclrn-mrr.tc05-64b-2t2c-eth-l2bdscale1mmaclrn-ndrdisc"
   outlier-const: 1.5
   window: 14
   - "parent"
   - "result"
   ignore-list:
+    # This test is "ndrdisc" test and was improperly tagged. It was fixed
+    # but it remains in the old output.xml files.
   - "tests.vpp.perf.l2.10ge2p1x520-eth-l2bdscale1mmaclrn-mrr.tc09-64b-4t4c-eth-l2bdscale1mmaclrn-ndrdisc"
   outlier-const: 1.5
   window: 14
   input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c.csv"
   output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c.rst"
 
+-
+  type: "table"
+  title: "Failed MRR tests"
+  algorithm: "table_failed_tests"
+  output-file-ext: ".csv"
+  output-file: "{DIR[STATIC,VPP]}/failed-tests"
+  data: "plot-performance-trending-all"
+  filter: "'MRR'"
+  parameters:
+  - "name"
+  - "parent"
+  - "status"
+  ignore-list:
+    # This test is "ndrdisc" test and was improperly tagged. It was fixed
+    # but it remains in the old output.xml files.
+  - "tests.vpp.perf.l2.10ge2p1x520-eth-l2bdscale1mmaclrn-mrr.tc01-64b-1t1c-eth-l2bdscale1mmaclrn-ndrdisc"
+  window: 14
+
+-
+  type: "table"
+  title: "HTML Failed MRR tests"
+  algorithm: "table_failed_tests_html"
+  input-file: "{DIR[STATIC,VPP]}/failed-tests.csv"
+  output-file: "{DIR[STATIC,VPP]}/failed-tests.rst"
+
 
 ################################################################################
 ###                                 C P T A                                  ###
index a4e2466..a2aa0dc 100644 (file)
@@ -17,8 +17,9 @@
 import multiprocessing
 import subprocess
 import numpy as np
-import pandas as pd
 import logging
+import csv
+import prettytable
 
 from os import walk, makedirs, environ
 from os.path import join, isdir
@@ -253,6 +254,29 @@ def classify_anomalies(data):
     return classification, avgs
 
 
+def convert_csv_to_pretty_txt(csv_file, txt_file):
+    """Convert the given csv table to pretty text table.
+
+    :param csv_file: The path to the input csv file.
+    :param txt_file: The path to the output pretty text file.
+    :type csv_file: str
+    :type txt_file: str
+    """
+
+    txt_table = None
+    with open(csv_file, 'rb') as csv_file:
+        csv_content = csv.reader(csv_file, delimiter=',', quotechar='"')
+        for row in csv_content:
+            if txt_table is None:
+                txt_table = prettytable.PrettyTable(row)
+            else:
+                txt_table.add_row(row)
+        txt_table.align["Test case"] = "l"
+    if txt_table:
+        with open(txt_file, "w") as txt_file:
+            txt_file.write(str(txt_table))
+
+
 class Worker(multiprocessing.Process):
     """Worker class used to process tasks in separate parallel processes.
     """
index 1901165..7ee0c0c 100644 (file)
@@ -91,7 +91,7 @@
       height: 800
       yaxis:
         showticklabels: True
-        tickformat: ".3s"
+        tickformat: ".4s"
         title: "Throughput [pps]"
         hoverformat: ".4s"
         range: []
       height: 800
       yaxis:
         showticklabels: True
-        tickformat: ".3s"
+        tickformat: ".4s"
         title: "Throughput [pps]"
         hoverformat: ".4s"
         range: []
   - "parent"
   - "result"
   ignore-list:
+    # This test is "ndrdisc" test and was improperly tagged. It was fixed
+    # but it remains in the old output.xml files.
   - "tests.vpp.perf.l2.10ge2p1x520-eth-l2bdscale1mmaclrn-mrr.tc01-64b-1t1c-eth-l2bdscale1mmaclrn-ndrdisc"
   outlier-const: 1.5
   window: 14
   - "parent"
   - "result"
   ignore-list:
+    # This test is "ndrdisc" test and was improperly tagged. It was fixed
+    # but it remains in the old output.xml files.
   - "tests.vpp.perf.l2.10ge2p1x520-eth-l2bdscale1mmaclrn-mrr.tc05-64b-2t2c-eth-l2bdscale1mmaclrn-ndrdisc"
   outlier-const: 1.5
   window: 14
   - "parent"
   - "result"
   ignore-list:
+    # This test is "ndrdisc" test and was improperly tagged. It was fixed
+    # but it remains in the old output.xml files.
   - "tests.vpp.perf.l2.10ge2p1x520-eth-l2bdscale1mmaclrn-mrr.tc09-64b-4t4c-eth-l2bdscale1mmaclrn-ndrdisc"
   outlier-const: 1.5
   window: 14
   input-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c.csv"
   output-file: "{DIR[STATIC,VPP]}/performance-trending-dashboard-4t4c.rst"
 
+-
+  type: "table"
+  title: "Failed MRR tests"
+  algorithm: "table_failed_tests"
+  output-file-ext: ".csv"
+  output-file: "{DIR[STATIC,VPP]}/failed-tests"
+  data: "plot-performance-trending-all"
+  filter: "'MRR'"
+  parameters:
+  - "name"
+  - "parent"
+  - "status"
+  ignore-list:
+    # This test is "ndrdisc" test and was improperly tagged. It was fixed
+    # but it remains in the old output.xml files.
+  - "tests.vpp.perf.l2.10ge2p1x520-eth-l2bdscale1mmaclrn-mrr.tc01-64b-1t1c-eth-l2bdscale1mmaclrn-ndrdisc"
+  window: 14
+
+-
+  type: "table"
+  title: "HTML Failed MRR tests"
+  algorithm: "table_failed_tests_html"
+  input-file: "{DIR[STATIC,VPP]}/failed-tests.csv"
+  output-file: "{DIR[STATIC,VPP]}/failed-tests.rst"
+
 
 ################################################################################
 ###                                 C P T A                                  ###
index 0a9d985..ba32932 100644 (file)
@@ -19,6 +19,8 @@ import subprocess
 import numpy as np
 import pandas as pd
 import logging
+import csv
+import prettytable
 
 from os import walk, makedirs, environ
 from os.path import join, isdir
@@ -316,6 +318,29 @@ def classify_anomalies(data, window):
     return classification
 
 
+def convert_csv_to_pretty_txt(csv_file, txt_file):
+    """Convert the given csv table to pretty text table.
+
+    :param csv_file: The path to the input csv file.
+    :param txt_file: The path to the output pretty text file.
+    :type csv_file: str
+    :type txt_file: str
+    """
+
+    txt_table = None
+    with open(csv_file, 'rb') as csv_file:
+        csv_content = csv.reader(csv_file, delimiter=',', quotechar='"')
+        for row in csv_content:
+            if txt_table is None:
+                txt_table = prettytable.PrettyTable(row)
+            else:
+                txt_table.add_row(row)
+        txt_table.align["Test case"] = "l"
+    if txt_table:
+        with open(txt_file, "w") as txt_file:
+            txt_file.write(str(txt_table))
+
+
 class Worker(multiprocessing.Process):
     """Worker class used to process tasks in separate parallel processes.
     """

©2016 FD.io a Linux Foundation Collaborative Project. All Rights Reserved.
Linux Foundation is a registered trademark of The Linux Foundation. Linux is a registered trademark of Linus Torvalds.
Please see our privacy policy and terms of use.