X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=resources%2Ftools%2Fpresentation%2Fnew%2Fgenerator_files.py;fp=resources%2Ftools%2Fpresentation%2Fnew%2Fgenerator_files.py;h=f8428adf8ccdc33d1eab33ff085ad81f88d8628d;hb=beeb2acb9ac153eaa54983bea46a76d596168965;hp=0000000000000000000000000000000000000000;hpb=3dcef45002a1b82c4503ec590d680950930fa193;p=csit.git diff --git a/resources/tools/presentation/new/generator_files.py b/resources/tools/presentation/new/generator_files.py new file mode 100644 index 0000000000..f8428adf8c --- /dev/null +++ b/resources/tools/presentation/new/generator_files.py @@ -0,0 +1,177 @@ +# Copyright (c) 2017 Cisco and/or its affiliates. +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at: +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +"""Algorithms to generate files. +""" + + +import logging + +from utils import get_files, get_rst_title_char + +RST_INCLUDE_TABLE = ("\n.. only:: html\n\n" + " .. csv-table::\n" + " :header-rows: 1\n" + " :widths: auto\n" + " :align: center\n" + " :file: {file_html}\n" + "\n.. only:: latex\n\n" + "\n .. raw:: latex\n\n" + " \csvautolongtable{{{file_latex}}}\n\n") + + +def generate_files(spec, data): + """Generate all files specified in the specification file. + + :param spec: Specification read from the specification file. + :param data: Data to process. + :type spec: Specification + :type data: InputData + """ + + logging.info("Generating the files ...") + for file_spec in spec.files: + try: + eval(file_spec["algorithm"])(file_spec, data) + except NameError as err: + logging.error("Probably algorithm '{alg}' is not defined: {err}". + format(alg=file_spec["algorithm"], err=repr(err))) + logging.info("Done.") + + +def _tests_in_suite(suite_name, tests): + """Check if the suite includes tests. + + :param suite_name: Name of the suite to be checked. + :param tests: Set of tests + :type suite_name: str + :type tests: pandas.Series + :returns: True if the suite includes tests. + :rtype: bool + """ + + for key in tests.keys(): + if suite_name == tests[key]["parent"]: + return True + return False + + +def file_test_results(file_spec, input_data): + """Generate the file(s) with algorithm: file_test_results specified in the + specification file. + + :param file_spec: File to generate. + :param input_data: Data to process. + :type file_spec: pandas.Series + :type input_data: InputData + """ + + file_name = "{0}{1}".format(file_spec["output-file"], + file_spec["output-file-ext"]) + rst_header = file_spec["file-header"] + + logging.info(" Generating the file {0} ...".format(file_name)) + + table_lst = get_files(file_spec["dir-tables"], ".csv", full_path=True) + if len(table_lst) == 0: + logging.error(" No tables to include in '{0}'. Skipping.". + format(file_spec["dir-tables"])) + return None + + job = file_spec["data"].keys()[0] + build = str(file_spec["data"][job][0]) + + logging.info(" Writing file '{0}'".format(file_name)) + + suites = input_data.suites(job, build)[file_spec["data-start-level"]:] + suites.sort_index(inplace=True) + + with open(file_name, "w") as file_handler: + file_handler.write(rst_header) + for suite_longname, suite in suites.iteritems(): + suite_name = suite["name"] + file_handler.write("\n{0}\n{1}\n".format( + suite_name, get_rst_title_char( + suite["level"] - file_spec["data-start-level"] - 1) * + len(suite_name))) + file_handler.write("\n{0}\n".format( + suite["doc"].replace('|br|', '\n\n -'))) + if _tests_in_suite(suite_name, input_data.tests(job, build)): + for tbl_file in table_lst: + if suite_name in tbl_file: + file_handler.write( + RST_INCLUDE_TABLE.format( + file_latex=tbl_file, + file_html=tbl_file.split("/")[-1])) + + logging.info(" Done.") + + +def file_merged_test_results(file_spec, input_data): + """Generate the file(s) with algorithm: file_merged_test_results specified + in the specification file. + + :param file_spec: File to generate. + :param input_data: Data to process. + :type file_spec: pandas.Series + :type input_data: InputData + """ + + file_name = "{0}{1}".format(file_spec["output-file"], + file_spec["output-file-ext"]) + rst_header = file_spec["file-header"] + + logging.info(" Generating the file {0} ...".format(file_name)) + + table_lst = get_files(file_spec["dir-tables"], ".csv", full_path=True) + if len(table_lst) == 0: + logging.error(" No tables to include in '{0}'. Skipping.". + format(file_spec["dir-tables"])) + return None + + logging.info(" Writing file '{0}'".format(file_name)) + + logging.info(" Creating the data set for the {0} '{1}'.". + format(file_spec.get("type", ""), file_spec.get("title", ""))) + tests = input_data.filter_data(file_spec) + tests = input_data.merge_data(tests) + + logging.info(" Creating the data set for the {0} '{1}'.". + format(file_spec.get("type", ""), file_spec.get("title", ""))) + suites = input_data.filter_data(file_spec, data_set="suites") + suites = input_data.merge_data(suites) + suites.sort_index(inplace=True) + + with open(file_name, "w") as file_handler: + file_handler.write(rst_header) + for suite_longname, suite in suites.iteritems(): + if "ndrchk" in suite_longname or "pdrchk" in suite_longname: + continue + if len(suite_longname.split(".")) <= file_spec["data-start-level"]: + continue + suite_name = suite["name"] + file_handler.write("\n{0}\n{1}\n".format( + suite_name, get_rst_title_char( + suite["level"] - file_spec["data-start-level"] - 1) * + len(suite_name))) + file_handler.write("\n{0}\n".format( + suite["doc"].replace('|br|', '\n\n -'))) + if _tests_in_suite(suite_name, tests): + for tbl_file in table_lst: + if suite_name in tbl_file: + file_handler.write( + RST_INCLUDE_TABLE.format( + file_latex=tbl_file, + file_html=tbl_file.split("/")[-1])) + + logging.info(" Done.")