X-Git-Url: https://gerrit.fd.io/r/gitweb?p=csit.git;a=blobdiff_plain;f=resources%2Ftools%2Fdash%2Fapp%2Fpal%2Ftrending%2Fgraphs.py;h=150b7056ba7d680e0992b96d4c802a920299b38f;hp=8cb96ea3b5a12740dfb4055feb1da7d51ba68ae0;hb=06d3f7331f9f10d99baa334b1808dfdc9c6fc8be;hpb=e972e67afac3ab3eb785668d01d3bdf1833eade9 diff --git a/resources/tools/dash/app/pal/trending/graphs.py b/resources/tools/dash/app/pal/trending/graphs.py index 8cb96ea3b5..150b7056ba 100644 --- a/resources/tools/dash/app/pal/trending/graphs.py +++ b/resources/tools/dash/app/pal/trending/graphs.py @@ -14,67 +14,121 @@ """ """ - import logging import plotly.graph_objects as go import pandas as pd -import re + +import hdrh.histogram +import hdrh.codec from datetime import datetime from numpy import isnan -from dash import no_update from ..jumpavg import classify -_COLORS = ( - u"#1A1110", - u"#DA2647", - u"#214FC6", - u"#01786F", - u"#BD8260", - u"#FFD12A", - u"#A6E7FF", - u"#738276", - u"#C95A49", - u"#FC5A8D", - u"#CEC8EF", - u"#391285", - u"#6F2DA8", - u"#FF878D", - u"#45A27D", - u"#FFD0B9", - u"#FD5240", - u"#DB91EF", - u"#44D7A8", - u"#4F86F7", - u"#84DE02", - u"#FFCFF1", - u"#614051" -) +_NORM_FREQUENCY = 2.0 # [GHz] +_FREQURENCY = { # [GHz] + "2n-aws": 1.000, + "2n-dnv": 2.000, + "2n-clx": 2.300, + "2n-icx": 2.600, + "2n-skx": 2.500, + "2n-tx2": 2.500, + "2n-zn2": 2.900, + "3n-alt": 3.000, + "3n-aws": 1.000, + "3n-dnv": 2.000, + "3n-icx": 2.600, + "3n-skx": 2.500, + "3n-tsh": 2.200 +} + _ANOMALY_COLOR = { - u"regression": 0.0, - u"normal": 0.5, - u"progression": 1.0 + "regression": 0.0, + "normal": 0.5, + "progression": 1.0 } -_COLORSCALE = [ - [0.00, u"red"], - [0.33, u"red"], - [0.33, u"white"], - [0.66, u"white"], - [0.66, u"green"], - [1.00, u"green"] +_COLORSCALE_TPUT = [ + [0.00, "red"], + [0.33, "red"], + [0.33, "white"], + [0.66, "white"], + [0.66, "green"], + [1.00, "green"] ] +_TICK_TEXT_TPUT = ["Regression", "Normal", "Progression"] +_COLORSCALE_LAT = [ + [0.00, "green"], + [0.33, "green"], + [0.33, "white"], + [0.66, "white"], + [0.66, "red"], + [1.00, "red"] +] +_TICK_TEXT_LAT = ["Progression", "Normal", "Regression"] _VALUE = { "mrr": "result_receive_rate_rate_avg", "ndr": "result_ndr_lower_rate_value", - "pdr": "result_pdr_lower_rate_value" + "pdr": "result_pdr_lower_rate_value", + "pdr-lat": "result_latency_forward_pdr_50_avg" } _UNIT = { "mrr": "result_receive_rate_rate_unit", "ndr": "result_ndr_lower_rate_unit", - "pdr": "result_pdr_lower_rate_unit" + "pdr": "result_pdr_lower_rate_unit", + "pdr-lat": "result_latency_forward_pdr_50_unit" } +_LAT_HDRH = ( # Do not change the order + "result_latency_forward_pdr_0_hdrh", + "result_latency_reverse_pdr_0_hdrh", + "result_latency_forward_pdr_10_hdrh", + "result_latency_reverse_pdr_10_hdrh", + "result_latency_forward_pdr_50_hdrh", + "result_latency_reverse_pdr_50_hdrh", + "result_latency_forward_pdr_90_hdrh", + "result_latency_reverse_pdr_90_hdrh", +) +# This value depends on latency stream rate (9001 pps) and duration (5s). +# Keep it slightly higher to ensure rounding errors to not remove tick mark. +PERCENTILE_MAX = 99.999501 + +_GRAPH_LAT_HDRH_DESC = { + "result_latency_forward_pdr_0_hdrh": "No-load.", + "result_latency_reverse_pdr_0_hdrh": "No-load.", + "result_latency_forward_pdr_10_hdrh": "Low-load, 10% PDR.", + "result_latency_reverse_pdr_10_hdrh": "Low-load, 10% PDR.", + "result_latency_forward_pdr_50_hdrh": "Mid-load, 50% PDR.", + "result_latency_reverse_pdr_50_hdrh": "Mid-load, 50% PDR.", + "result_latency_forward_pdr_90_hdrh": "High-load, 90% PDR.", + "result_latency_reverse_pdr_90_hdrh": "High-load, 90% PDR." +} + + +def _get_color(idx: int) -> str: + """ + """ + _COLORS = ( + "#1A1110", "#DA2647", "#214FC6", "#01786F", "#BD8260", "#FFD12A", + "#A6E7FF", "#738276", "#C95A49", "#FC5A8D", "#CEC8EF", "#391285", + "#6F2DA8", "#FF878D", "#45A27D", "#FFD0B9", "#FD5240", "#DB91EF", + "#44D7A8", "#4F86F7", "#84DE02", "#FFCFF1", "#614051" + ) + return _COLORS[idx % len(_COLORS)] + + +def _get_hdrh_latencies(row: pd.Series, name: str) -> dict: + """ + """ + + latencies = {"name": name} + for key in _LAT_HDRH: + try: + latencies[key] = row[key] + except KeyError: + return None + + return latencies def _classify_anomalies(data): @@ -104,7 +158,7 @@ def _classify_anomalies(data): stdv = 0.0 for sample in data.values(): if isnan(sample): - classification.append(u"outlier") + classification.append("outlier") avgs.append(sample) stdevs.append(sample) continue @@ -120,191 +174,325 @@ def _classify_anomalies(data): stdevs.append(stdv) values_left -= 1 continue - classification.append(u"normal") + classification.append("normal") avgs.append(avg) stdevs.append(stdv) values_left -= 1 return classification, avgs, stdevs -def trending_tput(data: pd.DataFrame, sel:dict, layout: dict, start: datetime, - end: datetime): +def select_trending_data(data: pd.DataFrame, itm:dict) -> pd.DataFrame: """ """ - if not sel: - return no_update, no_update + phy = itm["phy"].split("-") + if len(phy) == 4: + topo, arch, nic, drv = phy + if drv == "dpdk": + drv = "" + else: + drv += "-" + drv = drv.replace("_", "-") + else: + return None + + core = str() if itm["dut"] == "trex" else f"{itm['core']}" + ttype = "ndrpdr" if itm["testtype"] in ("ndr", "pdr") else itm["testtype"] + dut_v100 = "none" if itm["dut"] == "trex" else itm["dut"] + dut_v101 = itm["dut"] + + df = data.loc[( + ( + ( + (data["version"] == "1.0.0") & + (data["dut_type"].str.lower() == dut_v100) + ) | + ( + (data["version"] == "1.0.1") & + (data["dut_type"].str.lower() == dut_v101) + ) + ) & + (data["test_type"] == ttype) & + (data["passed"] == True) + )] + df = df[df.job.str.endswith(f"{topo}-{arch}")] + df = df[df.test_id.str.contains( + f"^.*[.|-]{nic}.*{itm['framesize']}-{core}-{drv}{itm['test']}-{ttype}$", + regex=True + )].sort_values(by="start_time", ignore_index=True) + + return df + + +def _generate_trending_traces(ttype: str, name: str, df: pd.DataFrame, + start: datetime, end: datetime, color: str, norm_factor: float) -> list: + """ + """ + + df = df.dropna(subset=[_VALUE[ttype], ]) + if df.empty: + return list() + df = df.loc[((df["start_time"] >= start) & (df["start_time"] <= end))] + if df.empty: + return list() - def _generate_traces(ttype: str, name: str, df: pd.DataFrame, - start: datetime, end: datetime, color: str): + x_axis = df["start_time"].tolist() + y_data = [itm * norm_factor for itm in df[_VALUE[ttype]].tolist()] - df = df.dropna(subset=[_VALUE[ttype], ]) - if df.empty: - return list() + anomalies, trend_avg, trend_stdev = _classify_anomalies( + {k: v for k, v in zip(x_axis, y_data)} + ) - x_axis = [d for d in df["start_time"] if d >= start and d <= end] + hover = list() + customdata = list() + for _, row in df.iterrows(): + d_type = "trex" if row["dut_type"] == "none" else row["dut_type"] + hover_itm = ( + f"date: {row['start_time'].strftime('%Y-%m-%d %H:%M:%S')}
" + f" [{row[_UNIT[ttype]]}]: {row[_VALUE[ttype]]:,.0f}
" + f"" + f"{d_type}-ref: {row['dut_version']}
" + f"csit-ref: {row['job']}/{row['build']}
" + f"hosts: {', '.join(row['hosts'])}" + ) + if ttype == "mrr": + stdev = ( + f"stdev [{row['result_receive_rate_rate_unit']}]: " + f"{row['result_receive_rate_rate_stdev']:,.0f}
" + ) + else: + stdev = "" + hover_itm = hover_itm.replace( + "", "latency" if ttype == "pdr-lat" else "average" + ).replace("", stdev) + hover.append(hover_itm) + if ttype == "pdr-lat": + customdata.append(_get_hdrh_latencies(row, name)) - anomalies, trend_avg, trend_stdev = _classify_anomalies( - {k: v for k, v in zip(x_axis, df[_VALUE[ttype]])} + hover_trend = list() + for avg, stdev, (_, row) in zip(trend_avg, trend_stdev, df.iterrows()): + d_type = "trex" if row["dut_type"] == "none" else row["dut_type"] + hover_itm = ( + f"date: {row['start_time'].strftime('%Y-%m-%d %H:%M:%S')}
" + f"trend [pps]: {avg:,.0f}
" + f"stdev [pps]: {stdev:,.0f}
" + f"{d_type}-ref: {row['dut_version']}
" + f"csit-ref: {row['job']}/{row['build']}
" + f"hosts: {', '.join(row['hosts'])}" ) + if ttype == "pdr-lat": + hover_itm = hover_itm.replace("[pps]", "[us]") + hover_trend.append(hover_itm) + traces = [ + go.Scatter( # Samples + x=x_axis, + y=y_data, + name=name, + mode="markers", + marker={ + "size": 5, + "color": color, + "symbol": "circle", + }, + text=hover, + hoverinfo="text+name", + showlegend=True, + legendgroup=name, + customdata=customdata + ), + go.Scatter( # Trend line + x=x_axis, + y=trend_avg, + name=name, + mode="lines", + line={ + "shape": "linear", + "width": 1, + "color": color, + }, + text=hover_trend, + hoverinfo="text+name", + showlegend=False, + legendgroup=name, + ) + ] + + if anomalies: + anomaly_x = list() + anomaly_y = list() + anomaly_color = list() hover = list() - for _, row in df.iterrows(): - hover_itm = ( - f"date: " - f"{row['start_time'].strftime('%d-%m-%Y %H:%M:%S')}
" - f"average [{row[_UNIT[ttype]]}]: " - f"{row[_VALUE[ttype]]}
" - f"{row['dut_type']}-ref: {row['dut_version']}
" - f"csit-ref: {row['job']}/{row['build']}" - ) - if ttype == "mrr": - stdev = ( - f"stdev [{row['result_receive_rate_rate_unit']}]: " - f"{row['result_receive_rate_rate_stdev']}
" + for idx, anomaly in enumerate(anomalies): + if anomaly in ("regression", "progression"): + anomaly_x.append(x_axis[idx]) + anomaly_y.append(trend_avg[idx]) + anomaly_color.append(_ANOMALY_COLOR[anomaly]) + hover_itm = ( + f"date: {x_axis[idx].strftime('%Y-%m-%d %H:%M:%S')}
" + f"trend [pps]: {trend_avg[idx]:,.0f}
" + f"classification: {anomaly}" ) - else: - stdev = "" - hover_itm = hover_itm.replace("", stdev) - hover.append(hover_itm) - - hover_trend = list() - for avg, stdev in zip(trend_avg, trend_stdev): - hover_trend.append( - f"trend [pps]: {avg}
" - f"stdev [pps]: {stdev}" - ) - - traces = [ - go.Scatter( # Samples - x=x_axis, - y=df[_VALUE[ttype]], - name=name, + if ttype == "pdr-lat": + hover_itm = hover_itm.replace("[pps]", "[us]") + hover.append(hover_itm) + anomaly_color.extend([0.0, 0.5, 1.0]) + traces.append( + go.Scatter( + x=anomaly_x, + y=anomaly_y, mode="markers", - marker={ - u"size": 5, - u"color": color, - u"symbol": u"circle", - }, text=hover, - hoverinfo=u"text+name", - showlegend=True, - legendgroup=name, - ), - go.Scatter( # Trend line - x=x_axis, - y=trend_avg, - name=name, - mode="lines", - line={ - u"shape": u"linear", - u"width": 1, - u"color": color, - }, - text=hover_trend, - hoverinfo=u"text+name", + hoverinfo="text+name", showlegend=False, legendgroup=name, - ) - ] - - if anomalies: - anomaly_x = list() - anomaly_y = list() - anomaly_color = list() - ticktext = [u"Regression", u"Normal", u"Progression"] - for idx, anomaly in enumerate(anomalies): - if anomaly in (u"regression", u"progression"): - anomaly_x.append(x_axis[idx]) - anomaly_y.append(trend_avg[idx]) - anomaly_color.append(_ANOMALY_COLOR[anomaly]) - anomaly_color.append([0.0, 0.5, 1.0]) - traces.append( - go.Scatter( - x=anomaly_x, - y=anomaly_y, - mode=u"markers", - hoverinfo=u"none", - showlegend=False, - legendgroup=name, - name=f"{name}-anomalies", - marker={ - u"size": 15, - u"symbol": u"circle-open", - u"color": anomaly_color, - u"colorscale": _COLORSCALE, - u"showscale": True, - u"line": { - u"width": 2 - }, - u"colorbar": { - u"y": 0.5, - u"len": 0.8, - u"title": u"Circles Marking Data Classification", - u"titleside": u"right", - u"titlefont": { - u"size": 14 - }, - u"tickmode": u"array", - u"tickvals": [0.167, 0.500, 0.833], - u"ticktext": ticktext, - u"ticks": u"", - u"ticklen": 0, - u"tickangle": -90, - u"thickness": 10 - } + name=name, + marker={ + "size": 15, + "symbol": "circle-open", + "color": anomaly_color, + "colorscale": _COLORSCALE_LAT \ + if ttype == "pdr-lat" else _COLORSCALE_TPUT, + "showscale": True, + "line": { + "width": 2 + }, + "colorbar": { + "y": 0.5, + "len": 0.8, + "title": "Circles Marking Data Classification", + "titleside": "right", + "tickmode": "array", + "tickvals": [0.167, 0.500, 0.833], + "ticktext": _TICK_TEXT_LAT \ + if ttype == "pdr-lat" else _TICK_TEXT_TPUT, + "ticks": "", + "ticklen": 0, + "tickangle": -90, + "thickness": 10 } - ) + } ) + ) + + return traces + - return traces +def graph_trending(data: pd.DataFrame, sel:dict, layout: dict, + start: datetime, end: datetime, normalize: bool) -> tuple: + """ + """ - # Generate graph: - fig = go.Figure() + if not sel: + return None, None + + fig_tput = None + fig_lat = None for idx, itm in enumerate(sel): - phy = itm["phy"].split("-") - if len(phy) == 4: - topo, arch, nic, drv = phy - if drv in ("dpdk", "ixgbe"): - drv = "" - else: - drv += "-" - drv = drv.replace("_", "-") - else: + + df = select_trending_data(data, itm) + if df is None or df.empty: continue - cadence = \ - "weekly" if (arch == "aws" or itm["testtype"] != "mrr") else "daily" - sel_topo_arch = ( - f"csit-vpp-perf-" - f"{itm['testtype'] if itm['testtype'] == 'mrr' else 'ndrpdr'}-" - f"{cadence}-master-{topo}-{arch}" - ) - df_sel = data.loc[(data["job"] == sel_topo_arch)] - regex = ( - f"^.*{nic}.*\.{itm['framesize']}-{itm['core']}-{drv}{itm['test']}-" - f"{'mrr' if itm['testtype'] == 'mrr' else 'ndrpdr'}$" + + name = "-".join((itm["dut"], itm["phy"], itm["framesize"], itm["core"], + itm["test"], itm["testtype"], )) + if normalize: + phy = itm["phy"].split("-") + topo_arch = f"{phy[0]}-{phy[1]}" if len(phy) == 4 else str() + norm_factor = (_NORM_FREQUENCY / _FREQURENCY[topo_arch]) \ + if topo_arch else 1.0 + else: + norm_factor = 1.0 + traces = _generate_trending_traces( + itm["testtype"], name, df, start, end, _get_color(idx), norm_factor ) - df = df_sel.loc[ - df_sel["test_id"].apply( - lambda x: True if re.search(regex, x) else False + if traces: + if not fig_tput: + fig_tput = go.Figure() + fig_tput.add_traces(traces) + + if itm["testtype"] == "pdr": + traces = _generate_trending_traces( + "pdr-lat", name, df, start, end, _get_color(idx), norm_factor ) - ].sort_values(by="start_time", ignore_index=True) - name = ( - f"{itm['phy']}-{itm['framesize']}-{itm['core']}-" - f"{itm['test']}-{itm['testtype']}" - ) - for trace in _generate_traces(itm['testtype'], name, df, start, end, - _COLORS[idx % len(_COLORS)]): - fig.add_trace(trace) + if traces: + if not fig_lat: + fig_lat = go.Figure() + fig_lat.add_traces(traces) + + if fig_tput: + fig_tput.update_layout(layout.get("plot-trending-tput", dict())) + if fig_lat: + fig_lat.update_layout(layout.get("plot-trending-lat", dict())) + + return fig_tput, fig_lat + + +def graph_hdrh_latency(data: dict, layout: dict) -> go.Figure: + """ + """ - style={ - "vertical-align": "top", - "display": "inline-block", - "width": "80%", - "padding": "5px" - } + fig = None - layout = layout.get("plot-trending", dict()) - fig.update_layout(layout) + traces = list() + for idx, (lat_name, lat_hdrh) in enumerate(data.items()): + try: + decoded = hdrh.histogram.HdrHistogram.decode(lat_hdrh) + except (hdrh.codec.HdrLengthException, TypeError) as err: + continue + previous_x = 0.0 + prev_perc = 0.0 + xaxis = list() + yaxis = list() + hovertext = list() + for item in decoded.get_recorded_iterator(): + # The real value is "percentile". + # For 100%, we cut that down to "x_perc" to avoid + # infinity. + percentile = item.percentile_level_iterated_to + x_perc = min(percentile, PERCENTILE_MAX) + xaxis.append(previous_x) + yaxis.append(item.value_iterated_to) + hovertext.append( + f"{_GRAPH_LAT_HDRH_DESC[lat_name]}
" + f"Direction: {('W-E', 'E-W')[idx % 2]}
" + f"Percentile: {prev_perc:.5f}-{percentile:.5f}%
" + f"Latency: {item.value_iterated_to}uSec" + ) + next_x = 100.0 / (100.0 - x_perc) + xaxis.append(next_x) + yaxis.append(item.value_iterated_to) + hovertext.append( + f"{_GRAPH_LAT_HDRH_DESC[lat_name]}
" + f"Direction: {('W-E', 'E-W')[idx % 2]}
" + f"Percentile: {prev_perc:.5f}-{percentile:.5f}%
" + f"Latency: {item.value_iterated_to}uSec" + ) + previous_x = next_x + prev_perc = percentile + + traces.append( + go.Scatter( + x=xaxis, + y=yaxis, + name=_GRAPH_LAT_HDRH_DESC[lat_name], + mode="lines", + legendgroup=_GRAPH_LAT_HDRH_DESC[lat_name], + showlegend=bool(idx % 2), + line=dict( + color=_get_color(int(idx/2)), + dash="solid", + width=1 if idx % 2 else 2 + ), + hovertext=hovertext, + hoverinfo="text" + ) + ) + if traces: + fig = go.Figure() + fig.add_traces(traces) + layout_hdrh = layout.get("plot-hdrh-latency", None) + if lat_hdrh: + fig.update_layout(layout_hdrh) - return fig, style + return fig