| #!/usr/bin/python3 | 
 | # | 
 | # Build performance test script | 
 | # | 
 | # Copyright (c) 2016, Intel Corporation. | 
 | # | 
 | # This program is free software; you can redistribute it and/or modify it | 
 | # under the terms and conditions of the GNU General Public License, | 
 | # version 2, as published by the Free Software Foundation. | 
 | # | 
 | # This program is distributed in the hope it will be useful, but WITHOUT | 
 | # ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | 
 | # FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for | 
 | # more details. | 
 | # | 
 | """Build performance test script""" | 
 | import argparse | 
 | import errno | 
 | import fcntl | 
 | import json | 
 | import logging | 
 | import os | 
 | import re | 
 | import shutil | 
 | import sys | 
 | from datetime import datetime | 
 |  | 
 | sys.path.insert(0, os.path.dirname(os.path.realpath(__file__)) + '/lib') | 
 | import scriptpath | 
 | scriptpath.add_oe_lib_path() | 
 | scriptpath.add_bitbake_lib_path() | 
 | import oeqa.buildperf | 
 | from oeqa.buildperf import (BuildPerfTestLoader, BuildPerfTestResult, | 
 |                             BuildPerfTestRunner, KernelDropCaches) | 
 | from oeqa.utils.commands import runCmd | 
 | from oeqa.utils.metadata import metadata_from_bb, write_metadata_file | 
 |  | 
 |  | 
 | # Set-up logging | 
 | LOG_FORMAT = '[%(asctime)s] %(levelname)s: %(message)s' | 
 | logging.basicConfig(level=logging.INFO, format=LOG_FORMAT, | 
 |                     datefmt='%Y-%m-%d %H:%M:%S') | 
 | log = logging.getLogger() | 
 |  | 
 |  | 
 | def acquire_lock(lock_f): | 
 |     """Acquire flock on file""" | 
 |     log.debug("Acquiring lock %s", os.path.abspath(lock_f.name)) | 
 |     try: | 
 |         fcntl.flock(lock_f, fcntl.LOCK_EX | fcntl.LOCK_NB) | 
 |     except IOError as err: | 
 |         if err.errno == errno.EAGAIN: | 
 |             return False | 
 |         raise | 
 |     log.debug("Lock acquired") | 
 |     return True | 
 |  | 
 |  | 
 | def pre_run_sanity_check(): | 
 |     """Sanity check of build environment""" | 
 |     build_dir = os.environ.get("BUILDDIR") | 
 |     if not build_dir: | 
 |         log.error("BUILDDIR not set. Please run the build environmnent setup " | 
 |                   "script.") | 
 |         return False | 
 |     if os.getcwd() != build_dir: | 
 |         log.error("Please run this script under BUILDDIR (%s)", build_dir) | 
 |         return False | 
 |  | 
 |     ret = runCmd('which bitbake', ignore_status=True) | 
 |     if ret.status: | 
 |         log.error("bitbake command not found") | 
 |         return False | 
 |     return True | 
 |  | 
 | def setup_file_logging(log_file): | 
 |     """Setup loggin to file""" | 
 |     log_dir = os.path.dirname(log_file) | 
 |     if not os.path.exists(log_dir): | 
 |         os.makedirs(log_dir) | 
 |     formatter = logging.Formatter(LOG_FORMAT) | 
 |     handler = logging.FileHandler(log_file) | 
 |     handler.setFormatter(formatter) | 
 |     log.addHandler(handler) | 
 |  | 
 |  | 
 | def archive_build_conf(out_dir): | 
 |     """Archive build/conf to test results""" | 
 |     src_dir = os.path.join(os.environ['BUILDDIR'], 'conf') | 
 |     tgt_dir = os.path.join(out_dir, 'build', 'conf') | 
 |     os.makedirs(os.path.dirname(tgt_dir)) | 
 |     shutil.copytree(src_dir, tgt_dir) | 
 |  | 
 |  | 
 | def update_globalres_file(result_obj, filename, metadata): | 
 |     """Write results to globalres csv file""" | 
 |     # Map test names to time and size columns in globalres | 
 |     # The tuples represent index and length of times and sizes | 
 |     # respectively | 
 |     gr_map = {'test1': ((0, 1), (8, 1)), | 
 |               'test12': ((1, 1), (None, None)), | 
 |               'test13': ((2, 1), (9, 1)), | 
 |               'test2': ((3, 1), (None, None)), | 
 |               'test3': ((4, 3), (None, None)), | 
 |               'test4': ((7, 1), (10, 2))} | 
 |  | 
 |     values = ['0'] * 12 | 
 |     for status, test, _ in result_obj.all_results(): | 
 |         if status in ['ERROR', 'SKIPPED']: | 
 |             continue | 
 |         (t_ind, t_len), (s_ind, s_len) = gr_map[test.name] | 
 |         if t_ind is not None: | 
 |             values[t_ind:t_ind + t_len] = test.times | 
 |         if s_ind is not None: | 
 |             values[s_ind:s_ind + s_len] = test.sizes | 
 |  | 
 |     log.debug("Writing globalres log to %s", filename) | 
 |     rev_info = metadata['layers']['meta'] | 
 |     with open(filename, 'a') as fobj: | 
 |         fobj.write('{},{}:{},{},'.format(metadata['hostname'], | 
 |                                          rev_info['branch'], | 
 |                                          rev_info['commit'], | 
 |                                          rev_info['commit'])) | 
 |         fobj.write(','.join(values) + '\n') | 
 |  | 
 |  | 
 | def parse_args(argv): | 
 |     """Parse command line arguments""" | 
 |     parser = argparse.ArgumentParser( | 
 |         formatter_class=argparse.ArgumentDefaultsHelpFormatter) | 
 |  | 
 |     parser.add_argument('-D', '--debug', action='store_true', | 
 |                         help='Enable debug level logging') | 
 |     parser.add_argument('--globalres-file', | 
 |                         type=os.path.abspath, | 
 |                         help="Append results to 'globalres' csv file") | 
 |     parser.add_argument('--lock-file', default='./oe-build-perf.lock', | 
 |                         metavar='FILENAME', type=os.path.abspath, | 
 |                         help="Lock file to use") | 
 |     parser.add_argument('-o', '--out-dir', default='results-{date}', | 
 |                         type=os.path.abspath, | 
 |                         help="Output directory for test results") | 
 |     parser.add_argument('-x', '--xml', action='store_true', | 
 |                         help='Enable JUnit xml output') | 
 |     parser.add_argument('--log-file', | 
 |                         default='{out_dir}/oe-build-perf-test.log', | 
 |                         help="Log file of this script") | 
 |     parser.add_argument('--run-tests', nargs='+', metavar='TEST', | 
 |                         help="List of tests to run") | 
 |  | 
 |     return parser.parse_args(argv) | 
 |  | 
 |  | 
 | def main(argv=None): | 
 |     """Script entry point""" | 
 |     args = parse_args(argv) | 
 |  | 
 |     # Set-up log file | 
 |     out_dir = args.out_dir.format(date=datetime.now().strftime('%Y%m%d%H%M%S')) | 
 |     setup_file_logging(args.log_file.format(out_dir=out_dir)) | 
 |  | 
 |     if args.debug: | 
 |         log.setLevel(logging.DEBUG) | 
 |  | 
 |     lock_f = open(args.lock_file, 'w') | 
 |     if not acquire_lock(lock_f): | 
 |         log.error("Another instance of this script is running, exiting...") | 
 |         return 1 | 
 |  | 
 |     if not pre_run_sanity_check(): | 
 |         return 1 | 
 |  | 
 |     # Check our capability to drop caches and ask pass if needed | 
 |     KernelDropCaches.check() | 
 |  | 
 |     # Load build perf tests | 
 |     loader = BuildPerfTestLoader() | 
 |     if args.run_tests: | 
 |         suite = loader.loadTestsFromNames(args.run_tests, oeqa.buildperf) | 
 |     else: | 
 |         suite = loader.loadTestsFromModule(oeqa.buildperf) | 
 |  | 
 |     # Save test metadata | 
 |     metadata = metadata_from_bb() | 
 |     log.info("Testing Git revision branch:commit %s:%s (%s)", | 
 |              metadata['layers']['meta']['branch'], | 
 |              metadata['layers']['meta']['commit'], | 
 |              metadata['layers']['meta']['commit_count']) | 
 |     if args.xml: | 
 |         write_metadata_file(os.path.join(out_dir, 'metadata.xml'), metadata) | 
 |     else: | 
 |         with open(os.path.join(out_dir, 'metadata.json'), 'w') as fobj: | 
 |             json.dump(metadata, fobj, indent=2) | 
 |     archive_build_conf(out_dir) | 
 |  | 
 |     runner = BuildPerfTestRunner(out_dir, verbosity=2) | 
 |  | 
 |     # Suppress logger output to stderr so that the output from unittest | 
 |     # is not mixed with occasional logger output | 
 |     log.handlers[0].setLevel(logging.CRITICAL) | 
 |  | 
 |     # Run actual tests | 
 |     result = runner.run(suite) | 
 |  | 
 |     # Restore logger output to stderr | 
 |     log.handlers[0].setLevel(log.level) | 
 |  | 
 |     if args.xml: | 
 |         result.write_results_xml() | 
 |     else: | 
 |         result.write_results_json() | 
 |     result.write_buildstats_json() | 
 |     if args.globalres_file: | 
 |         update_globalres_file(result, args.globalres_file, metadata) | 
 |     if result.wasSuccessful(): | 
 |         return 0 | 
 |  | 
 |     return 2 | 
 |  | 
 |  | 
 | if __name__ == '__main__': | 
 |     sys.exit(main()) | 
 |  |