| # test result tool - report text based test results |
| # |
| # Copyright (c) 2019, Intel Corporation. |
| # Copyright (c) 2019, Linux Foundation |
| # |
| # SPDX-License-Identifier: GPL-2.0-only |
| # |
| |
| import os |
| import glob |
| import json |
| import resulttool.resultutils as resultutils |
| from oeqa.utils.git import GitRepo |
| import oeqa.utils.gitarchive as gitarchive |
| |
| |
| class ResultsTextReport(object): |
| def __init__(self): |
| self.ptests = {} |
| self.ltptests = {} |
| self.ltpposixtests = {} |
| self.result_types = {'passed': ['PASSED', 'passed', 'PASS', 'XFAIL'], |
| 'failed': ['FAILED', 'failed', 'FAIL', 'ERROR', 'error', 'UNKNOWN', 'XPASS'], |
| 'skipped': ['SKIPPED', 'skipped', 'UNSUPPORTED', 'UNTESTED', 'UNRESOLVED']} |
| |
| |
| def handle_ptest_result(self, k, status, result, machine): |
| if machine not in self.ptests: |
| self.ptests[machine] = {} |
| |
| if k == 'ptestresult.sections': |
| # Ensure tests without any test results still show up on the report |
| for suite in result['ptestresult.sections']: |
| if suite not in self.ptests[machine]: |
| self.ptests[machine][suite] = { |
| 'passed': 0, 'failed': 0, 'skipped': 0, 'duration' : '-', |
| 'failed_testcases': [], "testcases": set(), |
| } |
| if 'duration' in result['ptestresult.sections'][suite]: |
| self.ptests[machine][suite]['duration'] = result['ptestresult.sections'][suite]['duration'] |
| if 'timeout' in result['ptestresult.sections'][suite]: |
| self.ptests[machine][suite]['duration'] += " T" |
| return True |
| |
| # process test result |
| try: |
| _, suite, test = k.split(".", 2) |
| except ValueError: |
| return True |
| |
| # Handle 'glib-2.0' |
| if 'ptestresult.sections' in result and suite not in result['ptestresult.sections']: |
| try: |
| _, suite, suite1, test = k.split(".", 3) |
| if suite + "." + suite1 in result['ptestresult.sections']: |
| suite = suite + "." + suite1 |
| except ValueError: |
| pass |
| |
| if suite not in self.ptests[machine]: |
| self.ptests[machine][suite] = { |
| 'passed': 0, 'failed': 0, 'skipped': 0, 'duration' : '-', |
| 'failed_testcases': [], "testcases": set(), |
| } |
| |
| # do not process duplicate results |
| if test in self.ptests[machine][suite]["testcases"]: |
| print("Warning duplicate ptest result '{}.{}' for {}".format(suite, test, machine)) |
| return False |
| |
| for tk in self.result_types: |
| if status in self.result_types[tk]: |
| self.ptests[machine][suite][tk] += 1 |
| self.ptests[machine][suite]["testcases"].add(test) |
| return True |
| |
| def handle_ltptest_result(self, k, status, result, machine): |
| if machine not in self.ltptests: |
| self.ltptests[machine] = {} |
| |
| if k == 'ltpresult.sections': |
| # Ensure tests without any test results still show up on the report |
| for suite in result['ltpresult.sections']: |
| if suite not in self.ltptests[machine]: |
| self.ltptests[machine][suite] = {'passed': 0, 'failed': 0, 'skipped': 0, 'duration' : '-', 'failed_testcases': []} |
| if 'duration' in result['ltpresult.sections'][suite]: |
| self.ltptests[machine][suite]['duration'] = result['ltpresult.sections'][suite]['duration'] |
| if 'timeout' in result['ltpresult.sections'][suite]: |
| self.ltptests[machine][suite]['duration'] += " T" |
| return |
| try: |
| _, suite, test = k.split(".", 2) |
| except ValueError: |
| return |
| # Handle 'glib-2.0' |
| if 'ltpresult.sections' in result and suite not in result['ltpresult.sections']: |
| try: |
| _, suite, suite1, test = k.split(".", 3) |
| print("split2: %s %s %s" % (suite, suite1, test)) |
| if suite + "." + suite1 in result['ltpresult.sections']: |
| suite = suite + "." + suite1 |
| except ValueError: |
| pass |
| if suite not in self.ltptests[machine]: |
| self.ltptests[machine][suite] = {'passed': 0, 'failed': 0, 'skipped': 0, 'duration' : '-', 'failed_testcases': []} |
| for tk in self.result_types: |
| if status in self.result_types[tk]: |
| self.ltptests[machine][suite][tk] += 1 |
| |
| def handle_ltpposixtest_result(self, k, status, result, machine): |
| if machine not in self.ltpposixtests: |
| self.ltpposixtests[machine] = {} |
| |
| if k == 'ltpposixresult.sections': |
| # Ensure tests without any test results still show up on the report |
| for suite in result['ltpposixresult.sections']: |
| if suite not in self.ltpposixtests[machine]: |
| self.ltpposixtests[machine][suite] = {'passed': 0, 'failed': 0, 'skipped': 0, 'duration' : '-', 'failed_testcases': []} |
| if 'duration' in result['ltpposixresult.sections'][suite]: |
| self.ltpposixtests[machine][suite]['duration'] = result['ltpposixresult.sections'][suite]['duration'] |
| return |
| try: |
| _, suite, test = k.split(".", 2) |
| except ValueError: |
| return |
| # Handle 'glib-2.0' |
| if 'ltpposixresult.sections' in result and suite not in result['ltpposixresult.sections']: |
| try: |
| _, suite, suite1, test = k.split(".", 3) |
| if suite + "." + suite1 in result['ltpposixresult.sections']: |
| suite = suite + "." + suite1 |
| except ValueError: |
| pass |
| if suite not in self.ltpposixtests[machine]: |
| self.ltpposixtests[machine][suite] = {'passed': 0, 'failed': 0, 'skipped': 0, 'duration' : '-', 'failed_testcases': []} |
| for tk in self.result_types: |
| if status in self.result_types[tk]: |
| self.ltpposixtests[machine][suite][tk] += 1 |
| |
| def get_aggregated_test_result(self, logger, testresult, machine): |
| test_count_report = {'passed': 0, 'failed': 0, 'skipped': 0, 'failed_testcases': []} |
| result = testresult.get('result', []) |
| for k in result: |
| test_status = result[k].get('status', []) |
| if k.startswith("ptestresult."): |
| if not self.handle_ptest_result(k, test_status, result, machine): |
| continue |
| elif k.startswith("ltpresult."): |
| self.handle_ltptest_result(k, test_status, result, machine) |
| elif k.startswith("ltpposixresult."): |
| self.handle_ltpposixtest_result(k, test_status, result, machine) |
| |
| # process result if it was not skipped by a handler |
| for tk in self.result_types: |
| if test_status in self.result_types[tk]: |
| test_count_report[tk] += 1 |
| if test_status in self.result_types['failed']: |
| test_count_report['failed_testcases'].append(k) |
| return test_count_report |
| |
| def print_test_report(self, template_file_name, test_count_reports): |
| from jinja2 import Environment, FileSystemLoader |
| script_path = os.path.dirname(os.path.realpath(__file__)) |
| file_loader = FileSystemLoader(script_path + '/template') |
| env = Environment(loader=file_loader, trim_blocks=True) |
| template = env.get_template(template_file_name) |
| havefailed = False |
| reportvalues = [] |
| machines = [] |
| cols = ['passed', 'failed', 'skipped'] |
| maxlen = {'passed' : 0, 'failed' : 0, 'skipped' : 0, 'result_id': 0, 'testseries' : 0, 'ptest' : 0 ,'ltptest': 0, 'ltpposixtest': 0} |
| for line in test_count_reports: |
| total_tested = line['passed'] + line['failed'] + line['skipped'] |
| vals = {} |
| vals['result_id'] = line['result_id'] |
| vals['testseries'] = line['testseries'] |
| vals['sort'] = line['testseries'] + "_" + line['result_id'] |
| vals['failed_testcases'] = line['failed_testcases'] |
| for k in cols: |
| vals[k] = "%d (%s%%)" % (line[k], format(line[k] / total_tested * 100, '.0f')) |
| for k in maxlen: |
| if k in vals and len(vals[k]) > maxlen[k]: |
| maxlen[k] = len(vals[k]) |
| reportvalues.append(vals) |
| if line['failed_testcases']: |
| havefailed = True |
| if line['machine'] not in machines: |
| machines.append(line['machine']) |
| reporttotalvalues = {} |
| for k in cols: |
| reporttotalvalues[k] = '%s' % sum([line[k] for line in test_count_reports]) |
| reporttotalvalues['count'] = '%s' % len(test_count_reports) |
| for (machine, report) in self.ptests.items(): |
| for ptest in self.ptests[machine]: |
| if len(ptest) > maxlen['ptest']: |
| maxlen['ptest'] = len(ptest) |
| for (machine, report) in self.ltptests.items(): |
| for ltptest in self.ltptests[machine]: |
| if len(ltptest) > maxlen['ltptest']: |
| maxlen['ltptest'] = len(ltptest) |
| for (machine, report) in self.ltpposixtests.items(): |
| for ltpposixtest in self.ltpposixtests[machine]: |
| if len(ltpposixtest) > maxlen['ltpposixtest']: |
| maxlen['ltpposixtest'] = len(ltpposixtest) |
| output = template.render(reportvalues=reportvalues, |
| reporttotalvalues=reporttotalvalues, |
| havefailed=havefailed, |
| machines=machines, |
| ptests=self.ptests, |
| ltptests=self.ltptests, |
| ltpposixtests=self.ltpposixtests, |
| maxlen=maxlen) |
| print(output) |
| |
| def view_test_report(self, logger, source_dir, branch, commit, tag, use_regression_map, raw_test, selected_test_case_only): |
| def print_selected_testcase_result(testresults, selected_test_case_only): |
| for testsuite in testresults: |
| for resultid in testresults[testsuite]: |
| result = testresults[testsuite][resultid]['result'] |
| test_case_result = result.get(selected_test_case_only, {}) |
| if test_case_result.get('status'): |
| print('Found selected test case result for %s from %s' % (selected_test_case_only, |
| resultid)) |
| print(test_case_result['status']) |
| else: |
| print('Could not find selected test case result for %s from %s' % (selected_test_case_only, |
| resultid)) |
| if test_case_result.get('log'): |
| print(test_case_result['log']) |
| test_count_reports = [] |
| configmap = resultutils.store_map |
| if use_regression_map: |
| configmap = resultutils.regression_map |
| if commit: |
| if tag: |
| logger.warning("Ignoring --tag as --commit was specified") |
| tag_name = "{branch}/{commit_number}-g{commit}/{tag_number}" |
| repo = GitRepo(source_dir) |
| revs = gitarchive.get_test_revs(logger, repo, tag_name, branch=branch) |
| rev_index = gitarchive.rev_find(revs, 'commit', commit) |
| testresults = resultutils.git_get_result(repo, revs[rev_index][2], configmap=configmap) |
| elif tag: |
| repo = GitRepo(source_dir) |
| testresults = resultutils.git_get_result(repo, [tag], configmap=configmap) |
| else: |
| testresults = resultutils.load_resultsdata(source_dir, configmap=configmap) |
| if raw_test: |
| raw_results = {} |
| for testsuite in testresults: |
| result = testresults[testsuite].get(raw_test, {}) |
| if result: |
| raw_results[testsuite] = {raw_test: result} |
| if raw_results: |
| if selected_test_case_only: |
| print_selected_testcase_result(raw_results, selected_test_case_only) |
| else: |
| print(json.dumps(raw_results, sort_keys=True, indent=4)) |
| else: |
| print('Could not find raw test result for %s' % raw_test) |
| return 0 |
| if selected_test_case_only: |
| print_selected_testcase_result(testresults, selected_test_case_only) |
| return 0 |
| for testsuite in testresults: |
| for resultid in testresults[testsuite]: |
| skip = False |
| result = testresults[testsuite][resultid] |
| machine = result['configuration']['MACHINE'] |
| |
| # Check to see if there is already results for these kinds of tests for the machine |
| for key in result['result'].keys(): |
| testtype = str(key).split('.')[0] |
| if ((machine in self.ltptests and testtype == "ltpiresult" and self.ltptests[machine]) or |
| (machine in self.ltpposixtests and testtype == "ltpposixresult" and self.ltpposixtests[machine])): |
| print("Already have test results for %s on %s, skipping %s" %(str(key).split('.')[0], machine, resultid)) |
| skip = True |
| break |
| if skip: |
| break |
| |
| test_count_report = self.get_aggregated_test_result(logger, result, machine) |
| test_count_report['machine'] = machine |
| test_count_report['testseries'] = result['configuration']['TESTSERIES'] |
| test_count_report['result_id'] = resultid |
| test_count_reports.append(test_count_report) |
| self.print_test_report('test_report_full_text.txt', test_count_reports) |
| |
| def report(args, logger): |
| report = ResultsTextReport() |
| report.view_test_report(logger, args.source_dir, args.branch, args.commit, args.tag, args.use_regression_map, |
| args.raw_test_only, args.selected_test_case_only) |
| return 0 |
| |
| def register_commands(subparsers): |
| """Register subcommands from this plugin""" |
| parser_build = subparsers.add_parser('report', help='summarise test results', |
| description='print a text-based summary of the test results', |
| group='analysis') |
| parser_build.set_defaults(func=report) |
| parser_build.add_argument('source_dir', |
| help='source file/directory/URL that contain the test result files to summarise') |
| parser_build.add_argument('--branch', '-B', default='master', help="Branch to find commit in") |
| parser_build.add_argument('--commit', help="Revision to report") |
| parser_build.add_argument('-t', '--tag', default='', |
| help='source_dir is a git repository, report on the tag specified from that repository') |
| parser_build.add_argument('-m', '--use_regression_map', action='store_true', |
| help='instead of the default "store_map", use the "regression_map" for report') |
| parser_build.add_argument('-r', '--raw_test_only', default='', |
| help='output raw test result only for the user provided test result id') |
| parser_build.add_argument('-s', '--selected_test_case_only', default='', |
| help='output selected test case result for the user provided test case id, if both test ' |
| 'result id and test case id are provided then output the selected test case result ' |
| 'from the provided test result id') |