| """Test evaluating expressions repeatedly comparing lldb against gdb.""" |
| |
| import sys |
| import lldb |
| from lldbsuite.test.lldbbench import BenchBase |
| from lldbsuite.test.decorators import * |
| from lldbsuite.test.lldbtest import * |
| from lldbsuite.test import configuration |
| from lldbsuite.test import lldbutil |
| |
| |
| class RepeatedExprsCase(BenchBase): |
| def setUp(self): |
| BenchBase.setUp(self) |
| self.source = "main.cpp" |
| self.line_to_break = line_number(self.source, "// Set breakpoint here.") |
| self.lldb_avg = None |
| self.gdb_avg = None |
| self.count = 100 |
| |
| @benchmarks_test |
| @expectedFailureAll( |
| oslist=["windows"], |
| bugnumber="llvm.org/pr22274: need a pexpect replacement for windows", |
| ) |
| def test_compare_lldb_to_gdb(self): |
| """Test repeated expressions with lldb vs. gdb.""" |
| self.build() |
| self.exe_name = "a.out" |
| |
| print() |
| self.run_lldb_repeated_exprs(self.exe_name, self.count) |
| print("lldb benchmark:", self.stopwatch) |
| self.run_gdb_repeated_exprs(self.exe_name, self.count) |
| print("gdb benchmark:", self.stopwatch) |
| print("lldb_avg/gdb_avg: %f" % (self.lldb_avg / self.gdb_avg)) |
| |
| def run_lldb_repeated_exprs(self, exe_name, count): |
| import pexpect |
| |
| exe = self.getBuildArtifact(exe_name) |
| |
| # Set self.child_prompt, which is "(lldb) ". |
| self.child_prompt = "(lldb) " |
| prompt = self.child_prompt |
| |
| # So that the child gets torn down after the test. |
| self.child = pexpect.spawn( |
| "%s %s %s" % (lldbtest_config.lldbExec, self.lldbOption, exe) |
| ) |
| child = self.child |
| |
| # Turn on logging for what the child sends back. |
| if self.TraceOn(): |
| child.logfile_read = sys.stdout |
| |
| child.expect_exact(prompt) |
| child.sendline("breakpoint set -f %s -l %d" % (self.source, self.line_to_break)) |
| child.expect_exact(prompt) |
| child.sendline("run") |
| child.expect_exact(prompt) |
| expr_cmd1 = "expr ptr[j]->point.x" |
| expr_cmd2 = "expr ptr[j]->point.y" |
| |
| # Reset the stopwatch now. |
| self.stopwatch.reset() |
| for i in range(count): |
| with self.stopwatch: |
| child.sendline(expr_cmd1) |
| child.expect_exact(prompt) |
| child.sendline(expr_cmd2) |
| child.expect_exact(prompt) |
| child.sendline("process continue") |
| child.expect_exact(prompt) |
| |
| child.sendline("quit") |
| try: |
| self.child.expect(pexpect.EOF) |
| except: |
| pass |
| |
| self.lldb_avg = self.stopwatch.avg() |
| if self.TraceOn(): |
| print("lldb expression benchmark:", str(self.stopwatch)) |
| self.child = None |
| |
| def run_gdb_repeated_exprs(self, exe_name, count): |
| import pexpect |
| |
| exe = self.getBuildArtifact(exe_name) |
| |
| # Set self.child_prompt, which is "(gdb) ". |
| self.child_prompt = "(gdb) " |
| prompt = self.child_prompt |
| |
| # So that the child gets torn down after the test. |
| self.child = pexpect.spawn("gdb --nx %s" % exe) |
| child = self.child |
| |
| # Turn on logging for what the child sends back. |
| if self.TraceOn(): |
| child.logfile_read = sys.stdout |
| |
| child.expect_exact(prompt) |
| child.sendline("break %s:%d" % (self.source, self.line_to_break)) |
| child.expect_exact(prompt) |
| child.sendline("run") |
| child.expect_exact(prompt) |
| expr_cmd1 = "print ptr[j]->point.x" |
| expr_cmd2 = "print ptr[j]->point.y" |
| |
| # Reset the stopwatch now. |
| self.stopwatch.reset() |
| for i in range(count): |
| with self.stopwatch: |
| child.sendline(expr_cmd1) |
| child.expect_exact(prompt) |
| child.sendline(expr_cmd2) |
| child.expect_exact(prompt) |
| child.sendline("continue") |
| child.expect_exact(prompt) |
| |
| child.sendline("quit") |
| child.expect_exact("The program is running. Exit anyway?") |
| child.sendline("y") |
| try: |
| self.child.expect(pexpect.EOF) |
| except: |
| pass |
| |
| self.gdb_avg = self.stopwatch.avg() |
| if self.TraceOn(): |
| print("gdb expression benchmark:", str(self.stopwatch)) |
| self.child = None |