summaryrefslogtreecommitdiffstats
path: root/import-layers/yocto-poky/scripts/oe-build-perf-test
diff options
context:
space:
mode:
Diffstat (limited to 'import-layers/yocto-poky/scripts/oe-build-perf-test')
-rwxr-xr-ximport-layers/yocto-poky/scripts/oe-build-perf-test223
1 files changed, 0 insertions, 223 deletions
diff --git a/import-layers/yocto-poky/scripts/oe-build-perf-test b/import-layers/yocto-poky/scripts/oe-build-perf-test
deleted file mode 100755
index 669470fa9..000000000
--- a/import-layers/yocto-poky/scripts/oe-build-perf-test
+++ /dev/null
@@ -1,223 +0,0 @@
-#!/usr/bin/python3
-#
-# Build performance test script
-#
-# Copyright (c) 2016, Intel Corporation.
-#
-# This program is free software; you can redistribute it and/or modify it
-# under the terms and conditions of the GNU General Public License,
-# version 2, as published by the Free Software Foundation.
-#
-# This program is distributed in the hope it will be useful, but WITHOUT
-# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
-# FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
-# more details.
-#
-"""Build performance test script"""
-import argparse
-import errno
-import fcntl
-import json
-import logging
-import os
-import re
-import shutil
-import sys
-from datetime import datetime
-
-sys.path.insert(0, os.path.dirname(os.path.realpath(__file__)) + '/lib')
-import scriptpath
-scriptpath.add_oe_lib_path()
-scriptpath.add_bitbake_lib_path()
-import oeqa.buildperf
-from oeqa.buildperf import (BuildPerfTestLoader, BuildPerfTestResult,
- BuildPerfTestRunner, KernelDropCaches)
-from oeqa.utils.commands import runCmd
-from oeqa.utils.metadata import metadata_from_bb, write_metadata_file
-
-
-# Set-up logging
-LOG_FORMAT = '[%(asctime)s] %(levelname)s: %(message)s'
-logging.basicConfig(level=logging.INFO, format=LOG_FORMAT,
- datefmt='%Y-%m-%d %H:%M:%S')
-log = logging.getLogger()
-
-
-def acquire_lock(lock_f):
- """Acquire flock on file"""
- log.debug("Acquiring lock %s", os.path.abspath(lock_f.name))
- try:
- fcntl.flock(lock_f, fcntl.LOCK_EX | fcntl.LOCK_NB)
- except IOError as err:
- if err.errno == errno.EAGAIN:
- return False
- raise
- log.debug("Lock acquired")
- return True
-
-
-def pre_run_sanity_check():
- """Sanity check of build environment"""
- build_dir = os.environ.get("BUILDDIR")
- if not build_dir:
- log.error("BUILDDIR not set. Please run the build environmnent setup "
- "script.")
- return False
- if os.getcwd() != build_dir:
- log.error("Please run this script under BUILDDIR (%s)", build_dir)
- return False
-
- ret = runCmd('which bitbake', ignore_status=True)
- if ret.status:
- log.error("bitbake command not found")
- return False
- return True
-
-def setup_file_logging(log_file):
- """Setup loggin to file"""
- log_dir = os.path.dirname(log_file)
- if not os.path.exists(log_dir):
- os.makedirs(log_dir)
- formatter = logging.Formatter(LOG_FORMAT)
- handler = logging.FileHandler(log_file)
- handler.setFormatter(formatter)
- log.addHandler(handler)
-
-
-def archive_build_conf(out_dir):
- """Archive build/conf to test results"""
- src_dir = os.path.join(os.environ['BUILDDIR'], 'conf')
- tgt_dir = os.path.join(out_dir, 'build', 'conf')
- os.makedirs(os.path.dirname(tgt_dir))
- shutil.copytree(src_dir, tgt_dir)
-
-
-def update_globalres_file(result_obj, filename, metadata):
- """Write results to globalres csv file"""
- # Map test names to time and size columns in globalres
- # The tuples represent index and length of times and sizes
- # respectively
- gr_map = {'test1': ((0, 1), (8, 1)),
- 'test12': ((1, 1), (None, None)),
- 'test13': ((2, 1), (9, 1)),
- 'test2': ((3, 1), (None, None)),
- 'test3': ((4, 3), (None, None)),
- 'test4': ((7, 1), (10, 2))}
-
- values = ['0'] * 12
- for status, test, _ in result_obj.all_results():
- if status in ['ERROR', 'SKIPPED']:
- continue
- (t_ind, t_len), (s_ind, s_len) = gr_map[test.name]
- if t_ind is not None:
- values[t_ind:t_ind + t_len] = test.times
- if s_ind is not None:
- values[s_ind:s_ind + s_len] = test.sizes
-
- log.debug("Writing globalres log to %s", filename)
- rev_info = metadata['layers']['meta']
- with open(filename, 'a') as fobj:
- fobj.write('{},{}:{},{},'.format(metadata['hostname'],
- rev_info['branch'],
- rev_info['commit'],
- rev_info['commit']))
- fobj.write(','.join(values) + '\n')
-
-
-def parse_args(argv):
- """Parse command line arguments"""
- parser = argparse.ArgumentParser(
- formatter_class=argparse.ArgumentDefaultsHelpFormatter)
-
- parser.add_argument('-D', '--debug', action='store_true',
- help='Enable debug level logging')
- parser.add_argument('--globalres-file',
- type=os.path.abspath,
- help="Append results to 'globalres' csv file")
- parser.add_argument('--lock-file', default='./oe-build-perf.lock',
- metavar='FILENAME', type=os.path.abspath,
- help="Lock file to use")
- parser.add_argument('-o', '--out-dir', default='results-{date}',
- type=os.path.abspath,
- help="Output directory for test results")
- parser.add_argument('-x', '--xml', action='store_true',
- help='Enable JUnit xml output')
- parser.add_argument('--log-file',
- default='{out_dir}/oe-build-perf-test.log',
- help="Log file of this script")
- parser.add_argument('--run-tests', nargs='+', metavar='TEST',
- help="List of tests to run")
-
- return parser.parse_args(argv)
-
-
-def main(argv=None):
- """Script entry point"""
- args = parse_args(argv)
-
- # Set-up log file
- out_dir = args.out_dir.format(date=datetime.now().strftime('%Y%m%d%H%M%S'))
- setup_file_logging(args.log_file.format(out_dir=out_dir))
-
- if args.debug:
- log.setLevel(logging.DEBUG)
-
- lock_f = open(args.lock_file, 'w')
- if not acquire_lock(lock_f):
- log.error("Another instance of this script is running, exiting...")
- return 1
-
- if not pre_run_sanity_check():
- return 1
-
- # Check our capability to drop caches and ask pass if needed
- KernelDropCaches.check()
-
- # Load build perf tests
- loader = BuildPerfTestLoader()
- if args.run_tests:
- suite = loader.loadTestsFromNames(args.run_tests, oeqa.buildperf)
- else:
- suite = loader.loadTestsFromModule(oeqa.buildperf)
-
- # Save test metadata
- metadata = metadata_from_bb()
- log.info("Testing Git revision branch:commit %s:%s (%s)",
- metadata['layers']['meta']['branch'],
- metadata['layers']['meta']['commit'],
- metadata['layers']['meta']['commit_count'])
- if args.xml:
- write_metadata_file(os.path.join(out_dir, 'metadata.xml'), metadata)
- else:
- with open(os.path.join(out_dir, 'metadata.json'), 'w') as fobj:
- json.dump(metadata, fobj, indent=2)
- archive_build_conf(out_dir)
-
- runner = BuildPerfTestRunner(out_dir, verbosity=2)
-
- # Suppress logger output to stderr so that the output from unittest
- # is not mixed with occasional logger output
- log.handlers[0].setLevel(logging.CRITICAL)
-
- # Run actual tests
- result = runner.run(suite)
-
- # Restore logger output to stderr
- log.handlers[0].setLevel(log.level)
-
- if args.xml:
- result.write_results_xml()
- else:
- result.write_results_json()
- result.write_buildstats_json()
- if args.globalres_file:
- update_globalres_file(result, args.globalres_file, metadata)
- if result.wasSuccessful():
- return 0
-
- return 2
-
-
-if __name__ == '__main__':
- sys.exit(main())
-
OpenPOWER on IntegriCloud