summaryrefslogtreecommitdiffstats
path: root/llvm/utils/update_analyze_test_checks.py
diff options
context:
space:
mode:
authorSimon Pilgrim <llvm-dev@redking.me.uk>2018-04-06 12:36:27 +0000
committerSimon Pilgrim <llvm-dev@redking.me.uk>2018-04-06 12:36:27 +0000
commit5334a2c5715b772b67eeb33165e5a3654a81f011 (patch)
treea215fee88132741309d68eee03f28fec911ed022 /llvm/utils/update_analyze_test_checks.py
parent09eeb3a8b90affc07f31a84b6c766ec803fbfb31 (diff)
downloadbcm5719-llvm-5334a2c5715b772b67eeb33165e5a3654a81f011.tar.gz
bcm5719-llvm-5334a2c5715b772b67eeb33165e5a3654a81f011.zip
[UpdateTestChecks] Add update_analyze_test_checks.py for cost model analysis generation
The script allows the auto-generation of checks for cost model tests to speed up their creation and help improve coverage, which will help a lot with PR36550. If the need arises we can add support for other analyze passes as well, but the cost models was the one I needed to get done - at the moment it just warns that any other analysis mode is unsupported. I've regenerated a couple of x86 test files to show the effect. Differential Revision: https://reviews.llvm.org/D45272 llvm-svn: 329390
Diffstat (limited to 'llvm/utils/update_analyze_test_checks.py')
-rw-r--r--llvm/utils/update_analyze_test_checks.py191
1 files changed, 191 insertions, 0 deletions
diff --git a/llvm/utils/update_analyze_test_checks.py b/llvm/utils/update_analyze_test_checks.py
new file mode 100644
index 00000000000..b9175ae7327
--- /dev/null
+++ b/llvm/utils/update_analyze_test_checks.py
@@ -0,0 +1,191 @@
+#!/usr/bin/env python2.7
+
+"""A script to generate FileCheck statements for 'opt' analysis tests.
+
+This script is a utility to update LLVM opt analysis test cases with new
+FileCheck patterns. It can either update all of the tests in the file or
+a single test function.
+
+Example usage:
+$ update_analyze_test_checks.py --opt=../bin/opt test/foo.ll
+
+Workflow:
+1. Make a compiler patch that requires updating some number of FileCheck lines
+ in regression test files.
+2. Save the patch and revert it from your local work area.
+3. Update the RUN-lines in the affected regression tests to look canonical.
+ Example: "; RUN: opt < %s -analyze -cost-model -S | FileCheck %s"
+4. Refresh the FileCheck lines for either the entire file or select functions by
+ running this script.
+5. Commit the fresh baseline of checks.
+6. Apply your patch from step 1 and rebuild your local binaries.
+7. Re-run this script on affected regression tests.
+8. Check the diffs to ensure the script has done something reasonable.
+9. Submit a patch including the regression test diffs for review.
+
+A common pattern is to have the script insert complete checking of every
+instruction. Then, edit it down to only check the relevant instructions.
+The script is designed to make adding checks to a test case fast, it is *not*
+designed to be authoratitive about what constitutes a good test!
+"""
+
+import argparse
+import itertools
+import os # Used to advertise this file's name ("autogenerated_note").
+import string
+import subprocess
+import sys
+import tempfile
+import re
+
+from UpdateTestChecks import common
+
+ADVERT = '; NOTE: Assertions have been autogenerated by '
+
+# RegEx: this is where the magic happens.
+
+IR_FUNCTION_RE = re.compile('^\s*define\s+(?:internal\s+)?[^@]*@([\w-]+)\s*\(')
+
+
+
+
+
+def main():
+ from argparse import RawTextHelpFormatter
+ parser = argparse.ArgumentParser(description=__doc__, formatter_class=RawTextHelpFormatter)
+ parser.add_argument('-v', '--verbose', action='store_true',
+ help='Show verbose output')
+ parser.add_argument('--opt-binary', default='opt',
+ help='The opt binary used to generate the test case')
+ parser.add_argument(
+ '--function', help='The function in the test file to update')
+ parser.add_argument('tests', nargs='+')
+ args = parser.parse_args()
+
+ autogenerated_note = (ADVERT + 'utils/' + os.path.basename(__file__))
+
+ opt_basename = os.path.basename(args.opt_binary)
+ if (opt_basename != "opt"):
+ print >>sys.stderr, 'ERROR: Unexpected opt name: ' + opt_basename
+ sys.exit(1)
+
+ for test in args.tests:
+ if args.verbose:
+ print >>sys.stderr, 'Scanning for RUN lines in test file: %s' % (test,)
+ with open(test) as f:
+ input_lines = [l.rstrip() for l in f]
+
+ raw_lines = [m.group(1)
+ for m in [common.RUN_LINE_RE.match(l) for l in input_lines] if m]
+ run_lines = [raw_lines[0]] if len(raw_lines) > 0 else []
+ for l in raw_lines[1:]:
+ if run_lines[-1].endswith("\\"):
+ run_lines[-1] = run_lines[-1].rstrip("\\") + " " + l
+ else:
+ run_lines.append(l)
+
+ if args.verbose:
+ print >>sys.stderr, 'Found %d RUN lines:' % (len(run_lines),)
+ for l in run_lines:
+ print >>sys.stderr, ' RUN: ' + l
+
+ prefix_list = []
+ for l in run_lines:
+ (tool_cmd, filecheck_cmd) = tuple([cmd.strip() for cmd in l.split('|', 1)])
+
+ if not tool_cmd.startswith(opt_basename + ' '):
+ print >>sys.stderr, 'WARNING: Skipping non-%s RUN line: %s' % (opt_basename, l)
+ continue
+
+ if not filecheck_cmd.startswith('FileCheck '):
+ print >>sys.stderr, 'WARNING: Skipping non-FileChecked RUN line: ' + l
+ continue
+
+ tool_cmd_args = tool_cmd[len(opt_basename):].strip()
+ tool_cmd_args = tool_cmd_args.replace('< %s', '').replace('%s', '').strip()
+
+ check_prefixes = [item for m in common.CHECK_PREFIX_RE.finditer(filecheck_cmd)
+ for item in m.group(1).split(',')]
+ if not check_prefixes:
+ check_prefixes = ['CHECK']
+
+ # FIXME: We should use multiple check prefixes to common check lines. For
+ # now, we just ignore all but the last.
+ prefix_list.append((check_prefixes, tool_cmd_args))
+
+ func_dict = {}
+ for prefixes, _ in prefix_list:
+ for prefix in prefixes:
+ func_dict.update({prefix: dict()})
+ for prefixes, opt_args in prefix_list:
+ if args.verbose:
+ print >>sys.stderr, 'Extracted opt cmd: ' + opt_basename + ' ' + opt_args
+ print >>sys.stderr, 'Extracted FileCheck prefixes: ' + str(prefixes)
+
+ raw_tool_outputs = common.invoke_tool(args.opt_binary, opt_args, test)
+
+ # Split analysis outputs by "Printing analysis " declarations.
+ for raw_tool_output in re.split(r'Printing analysis ', raw_tool_outputs):
+ common.build_function_body_dictionary(
+ common.ANALYZE_FUNCTION_RE, common.scrub_body, [],
+ raw_tool_output, prefixes, func_dict, args.verbose)
+
+ is_in_function = False
+ is_in_function_start = False
+ prefix_set = set([prefix for prefixes, _ in prefix_list for prefix in prefixes])
+ if args.verbose:
+ print >>sys.stderr, 'Rewriting FileCheck prefixes: %s' % (prefix_set,)
+ output_lines = []
+ output_lines.append(autogenerated_note)
+
+ for input_line in input_lines:
+ if is_in_function_start:
+ if input_line == '':
+ continue
+ if input_line.lstrip().startswith(';'):
+ m = common.CHECK_RE.match(input_line)
+ if not m or m.group(1) not in prefix_set:
+ output_lines.append(input_line)
+ continue
+
+ # Print out the various check lines here.
+ common.add_analyze_checks(output_lines, ';', prefix_list, func_dict, func_name)
+ is_in_function_start = False
+
+ if is_in_function:
+ if common.should_add_line_to_output(input_line, prefix_set):
+ # This input line of the function body will go as-is into the output.
+ # Except make leading whitespace uniform: 2 spaces.
+ input_line = common.SCRUB_LEADING_WHITESPACE_RE.sub(r' ', input_line)
+ output_lines.append(input_line)
+ else:
+ continue
+ if input_line.strip() == '}':
+ is_in_function = False
+ continue
+
+ # Discard any previous script advertising.
+ if input_line.startswith(ADVERT):
+ continue
+
+ # If it's outside a function, it just gets copied to the output.
+ output_lines.append(input_line)
+
+ m = IR_FUNCTION_RE.match(input_line)
+ if not m:
+ continue
+ func_name = m.group(1)
+ if args.function is not None and func_name != args.function:
+ # When filtering on a specific function, skip all others.
+ continue
+ is_in_function = is_in_function_start = True
+
+ if args.verbose:
+ print>>sys.stderr, 'Writing %d lines to %s...' % (len(output_lines), test)
+
+ with open(test, 'wb') as f:
+ f.writelines([l + '\n' for l in output_lines])
+
+
+if __name__ == '__main__':
+ main()
OpenPOWER on IntegriCloud