summaryrefslogtreecommitdiffstats
path: root/lldb
diff options
context:
space:
mode:
authorJohnny Chen <johnny.chen@apple.com>2011-10-22 00:57:05 +0000
committerJohnny Chen <johnny.chen@apple.com>2011-10-22 00:57:05 +0000
commitb8da426285eafb6b55ef29ec2f8223e0004332db (patch)
treef950c68757ad81642f40211f26ebc7a6e7ced294 /lldb
parentf385f12e178861f219e02784475e39cc73b07325 (diff)
downloadbcm5719-llvm-b8da426285eafb6b55ef29ec2f8223e0004332db.tar.gz
bcm5719-llvm-b8da426285eafb6b55ef29ec2f8223e0004332db.zip
Add bench.py as a driver script to run some benchmarks on lldb.
Add benchmarks for expression evaluations (TestExpressionCmd.py) and disassembly (TestDoAttachThenDisassembly.py). An example: [17:45:55] johnny:/Volumes/data/lldb/svn/trunk/test $ ./bench.py 2>&1 | grep -P '^lldb.*benchmark:' lldb startup delay (create fresh target) benchmark: Avg: 0.104274 (Laps: 30, Total Elapsed Time: 3.128214) lldb startup delay (set first breakpoint) benchmark: Avg: 0.102216 (Laps: 30, Total Elapsed Time: 3.066470) lldb frame variable benchmark: Avg: 1.649162 (Laps: 20, Total Elapsed Time: 32.983245) lldb stepping benchmark: Avg: 0.104409 (Laps: 50, Total Elapsed Time: 5.220461) lldb expr cmd benchmark: Avg: 0.206774 (Laps: 25, Total Elapsed Time: 5.169350) lldb disassembly benchmark: Avg: 0.089086 (Laps: 10, Total Elapsed Time: 0.890859) llvm-svn: 142708
Diffstat (limited to 'lldb')
-rwxr-xr-xlldb/test/bench.py48
-rw-r--r--lldb/test/benchmarks/disassembly/TestDoAttachThenDisassembly.py66
-rw-r--r--lldb/test/benchmarks/expression/TestExpressionCmd.py76
-rw-r--r--lldb/test/benchmarks/startup/TestStartupDelays.py2
-rwxr-xr-xlldb/test/dotest.py9
5 files changed, 196 insertions, 5 deletions
diff --git a/lldb/test/bench.py b/lldb/test/bench.py
new file mode 100755
index 00000000000..634fb182c2c
--- /dev/null
+++ b/lldb/test/bench.py
@@ -0,0 +1,48 @@
+#!/usr/bin/env python
+
+"""
+A simple bench runner which delegates to the ./dotest.py test driver to run the
+benchmarks defined in the list named 'benches'.
+
+You need to hand edit 'benches' to modify/change the command lines passed to the
+test driver.
+
+Use the following to get only the benchmark results in your terminal output:
+
+ ./bench.py 2>&1 | grep -P '^lldb.*benchmark:'
+"""
+
+import os, sys
+import re
+
+# dotest.py invocation with no '-e exe-path' uses lldb as the inferior program,
+# unless there is a mentioning of custom executable program.
+benches = [
+ # Measure startup delays creating a target and setting a breakpoint at main.
+ './dotest.py -v +b -n -p TestStartupDelays.py',
+
+ # Measure 'frame variable' response after stopping at Driver::MainLoop().
+ './dotest.py -v +b -x "-F Driver::MainLoop()" -n -p TestFrameVariableResponse.py',
+
+ # Measure stepping speed after stopping at Driver::MainLoop().
+ './dotest.py -v +b -x "-F Driver::MainLoop()" -n -p TestSteppingSpeed.py',
+
+ # Measure expression cmd response with a simple custom executable program.
+ './dotest.py +b -n -p TestExpressionCmd.py',
+
+ # Attach to a spawned lldb process then run disassembly benchmarks.
+ './dotest.py -v +b -n -p TestDoAttachThenDisassembly.py'
+]
+
+def main():
+ """Read the items from 'benches' and run the command line one by one."""
+ print "Starting bench runner...."
+
+ for command in benches:
+ print "Running %s" % (command)
+ os.system(command)
+
+ print "Bench runner done."
+
+if __name__ == '__main__':
+ main()
diff --git a/lldb/test/benchmarks/disassembly/TestDoAttachThenDisassembly.py b/lldb/test/benchmarks/disassembly/TestDoAttachThenDisassembly.py
new file mode 100644
index 00000000000..90cf7a87cae
--- /dev/null
+++ b/lldb/test/benchmarks/disassembly/TestDoAttachThenDisassembly.py
@@ -0,0 +1,66 @@
+"""Test lldb's disassemblt speed."""
+
+import os, sys
+import unittest2
+import lldb
+import pexpect
+from lldbbench import *
+
+class AttachThenDisassemblyBench(BenchBase):
+
+ mydir = os.path.join("benchmarks", "disassembly")
+
+ def setUp(self):
+ BenchBase.setUp(self)
+
+ @benchmarks_test
+ def test_attach_then_disassembly(self):
+ """Attach to a spawned lldb process then run disassembly benchmarks."""
+ print
+ self.run_lldb_attach_then_disassembly(10)
+ print "lldb disassembly benchmark:", self.stopwatch
+
+ def run_lldb_attach_then_disassembly(self, count):
+ target = self.dbg.CreateTarget(self.lldbHere)
+
+ # Spawn a new process and don't display the stdout if not in TraceOn() mode.
+ import subprocess
+ popen = subprocess.Popen([self.lldbHere, self.lldbOption],
+ stdout = open(os.devnull, 'w') if not self.TraceOn() else None)
+ if self.TraceOn():
+ print "pid of spawned process: %d" % popen.pid
+
+ # Attach to the launched lldb process.
+ listener = lldb.SBListener("my.attach.listener")
+ error = lldb.SBError()
+ process = target.AttachToProcessWithID(listener, popen.pid, error)
+
+ # Set thread0 as the selected thread, followed by the 'MainLoop' frame
+ # as the selected frame. Then do disassembly on the function.
+ thread0 = process.GetThreadAtIndex(0)
+ process.SetSelectedThread(thread0)
+ i = 0
+ found = False
+ for f in thread0:
+ #print "frame#%d %s" % (i, f.GetFunctionName())
+ if "MainLoop" in f.GetFunctionName():
+ found = True
+ thread0.SetSelectedFrame(i)
+ if self.TraceOn():
+ print "Found frame#%d for function 'MainLoop'" % i
+ break
+ i += 1
+
+ # Reset the stopwatch now.
+ self.stopwatch.reset()
+ for i in range(count):
+ with self.stopwatch:
+ # Disassemble the function.
+ self.runCmd("disassemble -f")
+
+
+if __name__ == '__main__':
+ import atexit
+ lldb.SBDebugger.Initialize()
+ atexit.register(lambda: lldb.SBDebugger.Terminate())
+ unittest2.main()
diff --git a/lldb/test/benchmarks/expression/TestExpressionCmd.py b/lldb/test/benchmarks/expression/TestExpressionCmd.py
new file mode 100644
index 00000000000..4993ee00282
--- /dev/null
+++ b/lldb/test/benchmarks/expression/TestExpressionCmd.py
@@ -0,0 +1,76 @@
+"""Test lldb's expression evaluations and collect statistics."""
+
+import os, sys
+import unittest2
+import lldb
+import pexpect
+from lldbbench import *
+
+class ExpressionEvaluationCase(BenchBase):
+
+ mydir = os.path.join("benchmarks", "expression")
+
+ def setUp(self):
+ BenchBase.setUp(self)
+ self.source = 'main.cpp'
+ self.line_to_break = line_number(self.source, '// Set breakpoint here.')
+ self.count = lldb.bmIterationCount
+ if self.count <= 0:
+ self.count = 25
+
+ @benchmarks_test
+ def test_expr_cmd(self):
+ """Test lldb's expression commands and collect statistics."""
+ self.buildDefault()
+ self.exe_name = 'a.out'
+
+ print
+ self.run_lldb_repeated_exprs(self.exe_name, self.count)
+ print "lldb expr cmd benchmark:", self.stopwatch
+
+ def run_lldb_repeated_exprs(self, exe_name, count):
+ exe = os.path.join(os.getcwd(), exe_name)
+
+ # Set self.child_prompt, which is "(lldb) ".
+ self.child_prompt = '(lldb) '
+ prompt = self.child_prompt
+
+ # Reset the stopwatch now.
+ self.stopwatch.reset()
+ for i in range(count):
+ # So that the child gets torn down after the test.
+ self.child = pexpect.spawn('%s %s %s' % (self.lldbExec, self.lldbOption, exe))
+ child = self.child
+
+ # Turn on logging for what the child sends back.
+ if self.TraceOn():
+ child.logfile_read = sys.stdout
+
+ child.expect_exact(prompt)
+ child.sendline('breakpoint set -f %s -l %d' % (self.source, self.line_to_break))
+ child.expect_exact(prompt)
+ child.sendline('run')
+ child.expect_exact(prompt)
+ expr_cmd1 = 'expr ptr[j]->point.x'
+ expr_cmd2 = 'expr ptr[j]->point.y'
+
+ with self.stopwatch:
+ child.sendline(expr_cmd1)
+ child.expect_exact(prompt)
+ child.sendline(expr_cmd2)
+ child.expect_exact(prompt)
+
+ child.sendline('quit')
+ try:
+ self.child.expect(pexpect.EOF)
+ except:
+ pass
+
+ self.child = None
+
+
+if __name__ == '__main__':
+ import atexit
+ lldb.SBDebugger.Initialize()
+ atexit.register(lambda: lldb.SBDebugger.Terminate())
+ unittest2.main()
diff --git a/lldb/test/benchmarks/startup/TestStartupDelays.py b/lldb/test/benchmarks/startup/TestStartupDelays.py
index f062354ac6c..1a15a917cdd 100644
--- a/lldb/test/benchmarks/startup/TestStartupDelays.py
+++ b/lldb/test/benchmarks/startup/TestStartupDelays.py
@@ -26,7 +26,7 @@ class StartupDelaysBench(BenchBase):
self.count = lldb.bmIterationCount
if self.count <= 0:
- self.count = 15
+ self.count = 30
@benchmarks_test
def test_startup_delay(self):
diff --git a/lldb/test/dotest.py b/lldb/test/dotest.py
index 7c3ba7af708..f33aa6f00b9 100755
--- a/lldb/test/dotest.py
+++ b/lldb/test/dotest.py
@@ -1062,10 +1062,11 @@ for ia in range(len(archs) if iterArchs else 1):
#print "sys.stdout name is", sys.stdout.name
# First, write out the number of collected test cases.
- sys.stderr.write(separator + "\n")
- sys.stderr.write("Collected %d test%s\n\n"
- % (suite.countTestCases(),
- suite.countTestCases() != 1 and "s" or ""))
+ if not noHeaders:
+ sys.stderr.write(separator + "\n")
+ sys.stderr.write("Collected %d test%s\n\n"
+ % (suite.countTestCases(),
+ suite.countTestCases() != 1 and "s" or ""))
class LLDBTestResult(unittest2.TextTestResult):
"""
OpenPOWER on IntegriCloud